Documentation

Welcome to bugAgent

We're thrilled to have you here! Dive in to learn how to give your AI agents their own bug reporting enrichment platform — along with a host of features and tools designed for the agentic era of software development.

bugAgent's mission is to 10x the velocity of testers by removing administration friction and letting them do what they do best — test in flow. Whether you're building autonomous coding assistants, integrating QA into CI/CD pipelines, or looking for a smarter way to manage bugs — you're in the right place.

Get Started

Jump into the area that fits your needs:

Why bugAgent? Our Philosophy

Context-driven testing. Measurable quality. Continuous improvement.

The Most Context-Aware Bug Reporting Platform

Every feature in bugAgent is designed to capture, preserve, and use context. Traditional bug trackers lose critical information at every handoff — between the tester who found the issue, the report that describes it, and the developer who fixes it. bugAgent eliminates those gaps by maintaining deep context across every touchpoint:

AI Assistant

The AI Assistant knows all your bug reports, notes, automations, and schedules for the current project. It operates with full awareness of your testing history and backlog. Trigger web and mobile performance tests, list runs, and review results through conversation.

Quality Scoring

Every report is scored across 10 context-driven dimensions — reproduction steps, expected vs actual, environment, evidence, root cause, impact, context & history, heuristics & oracles, clarity, and actionability.

Claude Analysis

Root cause analysis uses the full context of each report — including custom project instructions, tech stack details, and related history — to provide targeted fix suggestions.

FAB SDK

The SDK automatically captures page context: DOM state, network requests, console logs with stack traces, WCAG accessibility violations, performance metrics, and user actions.

Session Replay

DOM mutation recording captures the full testing context — every click, scroll, input, and page navigation — so developers see exactly what happened.

Cross-Platform Sync

Notes tie directly to projects and reports for full traceability. The Kanban board syncs bi-directionally with Jira, keeping context intact across platforms.

Quality of Testing = Quality of Product

bugAgent doesn't just track bugs — it actively measures and improves the quality of testing itself. The core thesis is simple: higher quality testing produces better bug reports, which leads to faster fixes and higher product quality.

Quality Score (1–10)

Every report is automatically scored on 10 context-driven dimensions. This measures the quality of the testing, not just the severity of the bug. Teams see exactly which dimensions need improvement — missing reproduction steps, vague expected behavior, no environment data — and can act on it.

Quality Score Trend

The Analytics dashboard tracks average quality scores over time, so teams can see whether their testing practices are improving week over week. Leadership gets a clear, data-driven view of testing effectiveness.

Quality Testing Health Score (0–100)

A single metric for testing effectiveness that combines report quality, resolution rate, automation pass rate, and severity distribution. Visible at the top of the Analytics dashboard with a gauge display. This gives leadership one number to answer: "How good is our testing?"

AI Formatting & Templates

AI-powered report formatting, structured templates (Bug Template, Checklist, Outline), and auto-classification help testers produce higher-quality reports with less effort.

Voice-to-Text & Annotations

Voice input via Whisper transcription and the built-in markup tool lower the barrier to thorough reporting. When it's easy to add detail, reports get better.

Coverage Mind Map

The automation coverage mind map shows where Playwright scripts are strong or weak, helping teams identify testing gaps and prioritize new automation.

The Continuous Improvement Loop

These capabilities create a virtuous cycle that raises product quality over time:

Measure testing quality → Identify gaps → Improve practices → Re-measure → Ship better software

Every report scored, every quality trend tracked, and every coverage gap identified feeds back into better testing practices. Teams using bugAgent don't just find more bugs — they get measurably better at finding them.

🎯
bugAgent is the only platform that treats testing quality as a first-class metric. The Quality Score, Health Score, and Analytics dashboard together give teams the tools to continuously improve how they test — not just what they find.

Introduction

Give AI agents a bug platform.

Software is increasingly written, tested, and maintained by AI agents. These agents need infrastructure — not just for writing code, but for the entire quality lifecycle. bugAgent is the platform that gives agents (and the humans who work alongside them) a shared language for bugs: one that's structured, enriched, and ready for action.

We built bugAgent because we believe the next generation of software teams will be hybrid — humans and agents collaborating in real-time. For that to work, bugs can't live in a spreadsheet or a GUI-first tracker. They need an API-first, protocol-native home where every participant (agent, human, CI pipeline, or IDE) can read, write, and act on them instantly.

What is bugAgent?

bugAgent is an API platform for giving AI agents their own bug reporting and product quality infrastructure. It goes beyond tracking — bugAgent is a report enrichment platform that classifies, contextualizes, and routes bugs, feature requests, enhancements, and more automatically.

With bugAgent, agents can:

  • Manage reports autonomously — create, classify, query, update, and resolve bug reports, feature requests, enhancements, and more without human intervention. Each report includes a built-in timer you can start, stop, and resume to track exact testing time down to the second. Click the time display to manually adjust. The timer is available on both the creation form and the report detail page
  • Enrich reports with AI — every report is auto-classified across 19 types — bugs (UI, performance, crash, security, logic, data, network, accessibility, compatibility), feature requests, enhancements, technical debt, documentation, DevOps, UX improvements, and integrations — with a confidence score
  • Deploy human testers Coming Soon — request on-demand exploratory or Playwright automation testing from human QA teams
  • Scan code and UI Coming Soon — run multi-agent swarms to check for abnormalities, vulnerabilities, and broken or outdated dependencies
  • Sync to external tools — push enriched reports to Jira with mapped fields, priority, and labels. Playwright scripts auto-sync to GitHub repos via the GitHub integration
  • AI Assistant — your QA command center — a context-aware assistant that goes far beyond report creation. Create and update bug reports, change status, severity, and type, add comments, create testing notes in 5 formats (Markdown, Plain Text, Checklist, Outline, Bug Template), list and search reports, notes, automations, and schedules, and send feedback — all through natural conversation. Use voice input powered by Whisper transcription, attach files, and let the AI analyze session replays to auto-draft reports. Start a new chat anytime with the New Chat button. Available in both the dashboard and the in-app FAB popup (no login required for FAB)
  • bugAgent Tools (SDK) — the SDK captures the last 60 seconds of clicks, navigation, DOM mutations, errors, and network failures. Includes console log capture with full stack traces (last 50 entries), a network waterfall of failed and slow API requests, real-time performance metrics (page load, FCP, FPS, memory, long tasks), optional screen + voice recording via the FAB toolbar, DOM snapshot replay, and a browser markup/annotation tool for drawing circles, arrows, and freehand annotations on screenshots before submitting. Available on Pro and Team plans
  • Quality Score — measure your testing — every report is scored 1–10 to measure the quality of the testing, not just the severity of the bug. The score evaluates 10 context-driven dimensions: reproduction steps, expected vs actual, environment details, evidence, root cause analysis, impact assessment, context & history, heuristics & oracles, clarity & structure, and actionability. Teams track their average score trending over time in Analytics, and the Quality Testing Health Score (0–100) gives leadership a single metric for testing effectiveness. Visual badges (red 1–3, amber 4–6, green 7–10) help teams identify weak reports before development starts. Available via dashboard, API, and MCP
  • Detect duplicates automatically — when a new report is created, text similarity analysis compares titles and descriptions against existing reports in the same project, flagging potential duplicates with match scores and direct links
  • Analytics Dashboard — a full-screen suite with 12+ interactive charts: Bug Reports Over Time (stacked bars by severity), Quality Score Trend (line chart), Severity/Status/Type breakdowns (donuts + bars), Top Bug Reporters leaderboard, Automation Health (pass/fail), Time Tracking by day and category, Notes Created sparkline, and Resolution Time. Includes a Quality Testing Health Score gauge (0–100) with crown icon on the best-performing area. Filter by time range (7/14/30/90 days) and project. Pro and Team plans. Also available via the REST API (GET /analytics) and MCP get_stats tool
  • No-code web automation — record browser actions with zero code — AI generates production-ready Playwright scripts automatically. Run tests on virtual viewports (30+ emulated devices) or live real browsers via BrowserStack (Chrome, Firefox, Safari, Edge on Windows and macOS). End-to-end testing that builds, manages, and scales with your product. Toggle between Virtual and Live modes on each run or schedule. AI optimization powered by Sonnet 4. Scripts track up to 10 versions with one-click undo. Line-by-line pass/fail highlighting. Pro and Team plans include an automation coverage mind map on the dashboard
  • Notes — capture testing observations in 5 formats (Markdown, Plain Text, Bug Template, Checklist, Outline). Includes voice-to-text dictation via OpenAI Whisper, a built-in timer you can start, stop, and resume to track exact testing time down to the second — click the time display to manually edit. Tracked time transfers automatically when converting a note to a bug report. Also includes file attachments up to 10 MB, private or shared visibility, auto-save, and keyword search with project, author, and date filters. Available on all plans
  • Time Tracking — log daily QA effort with categories (testing, bug-triage, automation, review, meeting). Create, update, and delete time entries per project. Filter by period (today, week, month), category, and team member. View daily, weekly, and monthly summaries in the analytics dashboard. Available via dashboard, API, and MCP. Team and Enterprise plans
  • Test Cases — a complete test management system inspired by TestRail. Create test cases with detailed steps (action + expected result) and drag-and-drop step reordering. Group them into test suites, create and assign test runs to team members with notifications, execute by marking each case pass/fail/blocked/skip with auto-save, instantly file bug reports from failures with full test context — bug reports link back to the test case and test run for full traceability. Re-run only failed cases, view reports with pass rate and date filtering, color-coded cards, and CSV export. Use Rapid Mode for fast test case creation. Available via dashboard, REST API, and MCP
  • Geo-Snap — enter any URL and capture screenshots from 50+ countries simultaneously. Verify localized content (currency, language, regional pricing), test geo-redirects, monitor competitors across markets, and check compliance with local regulations like cookie banners and privacy notices. Free plan: 1 country at a time with 10 saved screenshots. Pro/Team: up to 5 countries simultaneously with unlimited saved screenshots. Available via dashboard, REST API, and MCP create_geo_snap tool
  • Mobile Testing — upload APK (Android) or IPA (iOS) apps and run automated tests on real devices. Write Appium scripts for complex interactions or Maestro YAML for simple flows. Run on 3,000+ real devices via BrowserStack and Maestro Cloud. Every run produces video recordings, screenshots, device logs, and performance metrics. Failures auto-create bug reports with all artifacts. Schedule runs on multiple devices with Slack/email notifications. Device matrix testing across OS versions. Free: 1 run/day, Pro: unlimited, Team: unlimited. Available via dashboard, REST API, and MCP tools
  • Performance testing — Page quality audits (Performance, Accessibility, Best Practices, SEO) combined with load testing using virtual users (RPS, latency percentiles p50/p90/p95/p99). Mobile app performance testing on real Android/iOS devices via BrowserStack App Profiling — track CPU, memory, FPS, battery drain, and launch time with a Web/Mobile toggle. Core Web Vitals tracking, auto-bug creation when thresholds are breached (web scores or mobile metrics) with screenshot, recommendations, diagnostics, and quality scoring. CSV report export and archive/unarchive runs. Pro/Team/Enterprise plans
  • Security scanning — OWASP ZAP + Nuclei web DAST, MobSF mobile binary analysis, and Semgrep code SAST. Findings categorized by severity (Critical, High, Medium, Low, Info) with CVSS scores, CWE IDs, and remediation guidance. Auto-creates security bug reports for Critical/High findings. Pro/Team/Enterprise plans
  • Kanban Board — visualize bug reports across 8 status columns (New, Awaiting Triage, Confirmed, In Progress, Resolved, Retesting, Closed, Reopened) with drag-and-drop status management. Dragging a card updates the status in real time and syncs bi-directionally to Jira instantly. Toggle between list and kanban views with a persistent preference. Each card displays severity, type, description preview, and timestamps
  • Team Booster — scale your QA team instantly with pre-configured tester accounts. Fill in team size (1–10), location, duration, technical level, and budget — accounts are created in seconds and testers appear in your Team Management page with full platform access. Available on Pro and Team plans. You will not be charged until approval has been given. Also available programmatically via the REST API (POST /team-booster) and the MCP scale_team tool

This allows agents to assume their own identity and communicate via the universal protocol of bugs — with services, people, and other agents across any IDE or platform.

Feature Guide

AI Assistant — Example Questions

The AI Assistant is available in the dashboard (bottom-right chat panel) and the in-app FAB popup. It understands your organization's context — projects, reports, team members, automations, and settings. Here are examples of what you can ask:

Bug Reports

  • "Create a bug report for the login page — the password field doesn't accept special characters"
  • "Show me all critical bugs from the last week"
  • "Change the status of the checkout bug to in-progress"
  • "Add a comment to the API timeout report saying we reproduced it on staging"
  • "What's the severity breakdown of open bugs?"
  • "Update the login bug severity to critical"
  • "Find duplicate reports about the search feature"
  • "Create a feature request for dark mode support"

Web Automation (Playwright)

  • "List my automations"
  • "Run the homepage smoke test on desktop"
  • "Run the checkout flow on iPhone 15"
  • "Schedule the regression suite to run every weekday at 9am"
  • "What device options are available for automation?"
  • "Regenerate the script for the login test"
  • "Show me my automation schedules"

Mobile Automation

  • "List my mobile automations"
  • "Show recent mobile test runs"
  • "What devices can I test my Android app on?"
  • "How do I upload an APK for testing?"
  • "What script types are supported for mobile testing?"

Performance & Load Testing

  • "List my performance tests"
  • "Run the homepage performance test"
  • "Show me recent performance runs"
  • "What's my average performance score?"
  • "How many performance tests do I have left this month?"
  • "What Core Web Vitals does bugAgent track?"
  • "Explain what LCP, FID, and CLS mean"
  • "What are good thresholds for a mobile performance test?"

Mobile Performance Testing

  • "What metrics does mobile performance testing capture?"
  • "What's a good CPU threshold for an Android app?"
  • "How do I set up a mobile performance test?"
  • "Can I test both Android and iOS performance?"
  • "What devices are available for mobile performance profiling?"

Security Scanning

  • "List my security scans"
  • "Run the production security scan"
  • "Show me recent security scan results"
  • "How many security scans do I have left this month?"
  • "What does OWASP Top 10 cover?"
  • "Explain the difference between DAST and SAST"

Test Cases

  • "Create a test case for the user registration flow"
  • "List all critical test cases"
  • "Create a test suite called Smoke Tests"
  • "Start a test run from the Regression suite"
  • "Show me recent test runs"
  • "Export the Sprint 12 test run report as CSV"

Notes & Time Tracking

  • "Create a testing note in checklist format"
  • "Log 2 hours of exploratory testing"
  • "Add 45 minutes of bug triage time"
  • "Show my notes from this week"
  • "What time tracking categories are available?"

Team & Organization

  • "Show me the team members"
  • "Who's on my team?"
  • "Change my timezone to Eastern"

General QA & Testing

  • "What's the difference between regression and smoke testing?"
  • "Help me write better bug reports"
  • "What should I include in a performance test plan?"
  • "Explain boundary value analysis"
  • "What are the best practices for mobile app testing?"
Bug Reports

Bug reports are the core of bugAgent. Create reports manually, via the AI Assistant, through the MCP server, or automatically from test failures. Each report includes:

  • 19 report types — functional, UI/UX, performance, security, crash, data integrity, accessibility, compatibility, network, logic, feature request, enhancement, technical debt, documentation, DevOps, UX improvement, and integration
  • 4 severity levels — critical, high, medium, low
  • 8 status columns — New, Awaiting Triage, Confirmed, In Progress, Resolved, Retesting, Closed, Reopened
  • Quality Score — AI-scored 1-10 across 10 dimensions (reproduction, expected vs actual, environment, evidence, root cause, impact, context, heuristics, clarity, actionability)
  • Built-in timer — start/stop/resume to track testing time per report
  • Media attachments — screenshots, videos, PDFs, logs (up to 10 MB each)
  • Jira sync — bi-directional sync with mapped fields, priority, labels, and attachments
  • Claude analysis — send any report to Claude for root cause analysis with probable cause, suggested fix, verification steps, and risk assessment (requires Claude connection in Settings → Integrations)
  • Impact Score — 0-100 composite of severity, frequency, and affected users
Performance & Load Testing

Web performance runs page quality audits scoring Performance, Accessibility, Best Practices, and SEO (0-100). Tracks Core Web Vitals: LCP, FID, CLS, FCP, Speed Index, TTFB, TTI, and Total Blocking Time. Includes actionable opportunities and diagnostics with estimated savings.

Load testing simulates concurrent virtual users with configurable duration and ramp-up. Measures requests per second, error rate, and latency percentiles (p50, p90, p95, p99). Results stream to Grafana Cloud for dashboard viewing.

Mobile performance profiles real Android/iOS apps on BrowserStack devices. Metrics include CPU (avg/max %), memory (avg/peak MB), FPS (avg/min), battery drain (%), app launch time (ms), network data (KB), and disk I/O (KB). Select from Google Pixel 7/8, Samsung Galaxy S23/S24, iPhone 14/15 series.

Auto-bug creation — configure thresholds (e.g. performance score < 50, CPU > 80%, p95 > 3000ms) and bug reports are created automatically when breached, with full metrics, recommendations, screenshot, and quality scoring.

Plan limits — Pro: 1 test/month, Team: 2/month, Enterprise: 10/month. Each device in a multi-device mobile test counts as 1 credit. Performance+ ($299/mo): unlimited tests.

Security Scanning

Web DAST scans websites for OWASP Top 10 vulnerabilities using OWASP ZAP and Nuclei. Checks for XSS, SQL injection, CVEs, security misconfigurations, broken authentication, and more. Each finding includes a CVSS score, CWE ID, affected URL, evidence, and remediation guidance.

Mobile binary analysis uses MobSF to scan APK (Android) and IPA (iOS) binaries for insecure data storage, hardcoded secrets, weak cryptography, and manifest/entitlement issues.

Code SAST runs Semgrep rules against source repositories to detect injection flaws, insecure deserialization, hardcoded secrets, and code-level security issues. Includes dependency audit (npm audit, pip-audit) and license scanning for GPL/AGPL/copyleft compliance.

Auto-bug creation — Critical and High severity findings automatically create security-type bug reports with full vulnerability details, evidence, and remediation steps.

Plan limits — Pro: 2 scans/month, Team: 5/month, Enterprise: 20/month. Security+ ($499/mo): unlimited scans.

Scheduled Tasks & Automation

Schedule any web automation to run automatically using cron expressions. Configure device/viewport, timezone, and failure notifications (email, Slack, or both). View upcoming and past runs on the calendar-style Scheduled page.

  • Cron scheduling — flexible scheduling: daily, weekdays, specific days, hourly, or custom cron expressions
  • Device selection — 30+ virtual viewport devices or BrowserStack live browsers
  • Failure notifications — email, Slack, or both when a scheduled run fails
  • Auto-bug creation — failed automation runs automatically create bug reports with video, screenshots, and error logs
  • Concurrency — up to 5 parallel runs on the runner service
Notes

Capture testing observations and session notes during exploratory testing or structured test execution.

  • 5 formats — Markdown, Plain Text, Bug Template, Checklist, Outline
  • Voice dictation — speak your notes via OpenAI Whisper transcription
  • Built-in timer — track time spent on each note, transfers to bug reports on conversion
  • Visibility — private (only you) or shared (visible to team)
  • File attachments — up to 10 MB per attachment
  • Search & filter — keyword search, project, author, and date filters
Time Tracking

Log daily QA effort with categories for analytics and team visibility. Team and Enterprise plans.

  • Categories — Testing, Exploratory Testing, Test Case Writing, Test Execution, Regression Testing, Bug Reporting, Bug Verification, Test Planning, Environment Setup, WCAG/Accessibility Testing, Development, Code Review, Bug Fixing, Architecture/Design, Documentation, DevOps/CI-CD, Debugging, Research/Spike, Deployment, Meeting, Special Project, Communication, Admin, Planning
  • Filters — by period (today, week, month), project, category, and team member
  • Analytics — daily, weekly, and monthly summaries in the dashboard
  • AI Assistant — log time via conversation: "log 2 hours of testing"
bugAgent Tools (SDK / FAB)

A floating action button (FAB) that installs on any website with one line of code. Five tools in one:

  • Automate — captures DOM mutations, user clicks, navigation, form interactions, console errors (last 50 with stack traces), and failed network requests. Records the last 60 seconds of activity for session replay
  • Recorder — records screen and voice with one click. Video is attached to bug reports automatically
  • Annotate — marks up screenshots with a 7-color drawing tool (circles, arrows, freehand) before submitting
  • WCAG Analyzer — runs axe-core accessibility checks and groups results by severity (critical, serious, moderate, minor) with WCAG rule IDs and affected elements
  • Rewind — replays the last 5 user actions with a ghost cursor animation to show exactly what happened

Installation:

<script src="https://app.bugagent.com/sdk/bugagent-sdk.js" data-key="YOUR_API_KEY" data-position="bottom-right"></script>

Get your API key from Settings → Developers. The FAB works without login — anyone on your site can report bugs. Pro and Team plans.

Integrations

Connect bugAgent to your existing tools from Settings → Integrations.

  • Jira — bi-directional sync with OAuth 2.0. Push reports to Jira with mapped severity/priority, labels, and attachments. Pull updates back. Auto-sync polls periodically. One connection shared by all team members. Force sync button on any report
  • Claude AI — connect your Anthropic API key to enable "Send to Claude" on bug reports. Claude analyzes the report and returns root cause analysis, suggested fix, verification steps, and risk assessment. Also powers the AI Assistant's scoring and analysis
  • GitHub — connect a GitHub repo to auto-push Playwright automation scripts. Scripts sync on save/update. Branch and commit management included
  • Slack — receive failure notifications from scheduled automation runs. Configure per-schedule: email, Slack, or both
Analytics Dashboard

12+ interactive charts: Bug Reports Over Time, Quality Score Trend, Severity/Status/Type breakdowns, Top Reporters leaderboard, Automation Health, Time Tracking, Performance Score Trend, and Quality Testing Health Score. Filter by time range (7/14/30/90 days) and project. Performance analytics include web audit averages, mobile device metrics, and score trend bars. Pro/Team/Enterprise plans.

Security & 2FA

Two-factor authentication via authenticator apps (TOTP), rate limiting on all endpoints, input validation, SQL injection protection via parameterized queries, file upload sanitization (type + size checks), HTTPS-only, session management with secure cookies, and role-based access control (Owner > Admin > Manager > Contributor > Tester).

Mobile Testing

Upload APK (Android) or IPA (iOS) apps and run automated tests on 3,000+ real devices via BrowserStack. Write Appium scripts or Maestro YAML. Record actions with Appetize simulator. Every run produces video recordings, screenshots, device logs, and performance metrics. Failed runs auto-create bug reports. Schedule runs with Slack/email notifications. Pro/Team/Enterprise plans.

Geo-Snap

Enter any URL and capture screenshots from 50+ countries simultaneously. Verify localized content, test geo-redirects, monitor competitors, and check compliance with local regulations. Free: 1 country, 10 saved screenshots. Pro/Team: up to 5 countries simultaneously, unlimited saved screenshots.

Test Cases

Create test cases with detailed steps (action + expected result), group into suites, assign test runs to team members with notifications. Execute by marking each case pass/fail/blocked/skip with auto-save. File bug reports from failures with full test context. Re-run only failed cases. CSV export. Rapid Mode for fast creation. Available via dashboard, REST API, and MCP.

🏗️
We handle the infrastructure — bug classification, storage, enrichment, Jira sync, team management, and analytics — so you can focus on building reliable software.

What can autonomous agents do on this platform?

bugAgent gives autonomous agents the same access as the AI Assistant — every feature is available through the REST API and Model Context Protocol. Agents can file bugs, read reports, trigger tests, analyze results, update statuses, and orchestrate entire QA workflows without human intervention. Here are real-world autonomous cycles that agents can run on the platform:

🔄
The Self-Healing Bug Cycle

A dev agent reads the latest bug reports from bugAgent, identifies a fixable issue, patches the code, pushes the fix to a branch, then updates the bug report status to "resolved" with a link to the PR. A human reviews the fix and closes the report — or reopens it with feedback for the agent to try again.

🧪
Test → Fix → Verify Loop

An agent creates test cases for a feature, schedules a Playwright automation to run nightly, and monitors the results. When a test fails, the agent files a bug report, a second agent picks up the bug and applies a fix, then the first agent re-runs the automation to verify. The cycle repeats until all tests pass — with humans reviewing each fix before merge.

📊
Performance Guardian

An agent runs web performance tests on every deploy, checks that the performance score stays above 80 and LCP stays under 2.5 seconds. It runs mobile performance tests weekly to monitor CPU, memory, and FPS. If any metric degrades, the agent files a performance bug with the full report, recommendations, and a link to the trend data in the analytics dashboard.

🚀
CI/CD Pipeline Integration

A CI/CD agent triggers a full test suite after every merge to main: Playwright web automations, mobile tests on real devices, and a performance audit. It collects all results, files bug reports for any failures with video and error logs, updates the deployment ticket, and blocks the release if critical issues are found. The entire pipeline talks to bugAgent through the API — no dashboard needed.

🤖
Autonomous QA Agent

A fully autonomous QA agent reviews the analytics dashboard daily. It checks the Quality Testing Health Score, identifies the weakest areas, creates targeted test cases, runs them, files bugs for failures, and logs time entries for the testing effort. It sends a summary note to the team: "Ran 15 regression tests. 2 failures found. Bug reports filed. Estimated fix effort: 3 hours."

📱
Mobile App Release Cycle

Before every app release, an agent uploads the new APK/IPA to bugAgent, runs mobile performance profiling on 4 target devices, compares CPU and memory against the previous build, and files regression bugs if any metric worsens by more than 10%. It also runs mobile automation tests and generates a release readiness report with pass/fail summary, performance comparison, and outstanding bugs.

🔍
Code Review + Bug Detection

A code review agent analyzes PRs for potential issues (XSS, SQL injection, performance regressions, accessibility violations). For each issue found, it files a classified bug report with severity, affected file, line number, and a recommended fix. The dev agent reads the report, applies the fix, and updates the bug status. bugAgent tracks the full lifecycle from detection to resolution.

⚙️
Continuous Quality Monitoring

An agent runs on a schedule: every hour it checks the latest bug reports, performance runs, and automation results. It computes a health score, compares against yesterday's baseline, and sends a Slack notification if quality drops. Weekly, it generates a comprehensive quality report as a bugAgent note, tagging areas that need attention. The team lead reviews it every Monday — no manual data gathering needed.

💡
Every action above is available through the REST API and MCP server. Agents authenticate with API keys, read and write structured JSON, and orchestrate workflows programmatically. The platform is the operating system for autonomous software quality.

AI Assistant

The AI Assistant is bugAgent's command center for testers, developers, and QA leads. Instead of navigating menus and filling forms, you talk to the assistant in natural language and it takes action on your behalf — creating reports, managing automations, logging time, searching your backlog, and more.

Where to Find It

  • Dashboard chat panel — a persistent chat panel in the bottom-right corner of the dashboard. Always available while you work. Click "New Chat" to start a fresh conversation at any time.
  • FAB popup — the floating action button (FAB) on sites running the bugAgent SDK includes an AI Assistant popup. No login required for reporters — anyone on your site can ask the assistant to help file a bug.

Context Awareness

The assistant operates with deep awareness of your organization. It knows your projects, bug reports, team members, automations, schedules, notes, time entries, and integration settings. When you ask "show me critical bugs from last week," it queries your actual data. When you say "run the checkout test on iPhone 15," it finds the right automation and triggers it on the right device.

Input Methods

  • Text — type naturally. The assistant understands intent, not just keywords.
  • Voice — click the microphone icon for Whisper-powered transcription. Speak your bug report, note, or question and the assistant converts it to text.
  • File attachments — drag or click to attach screenshots, videos, logs, or documents. The assistant can analyze session replays and auto-draft bug reports from the captured data.

What It Can Do

The assistant supports the full breadth of bugAgent operations:

  • Bug reports — create, update, search, change status/severity/type, add comments. Supports all 19 report types and 8 statuses. Example: "Create a critical bug for the checkout page — the total doesn't include tax."
  • Web automation — list automations, trigger runs on any of the 30+ device viewports, check run results, regenerate scripts, and manage schedules. Example: "Run the login smoke test on Galaxy S23."
  • Mobile testing — list mobile automations, show recent runs, explain device options and script types. Example: "Show recent mobile test runs."
  • Performance testing — list performance tests, trigger runs, review results, check averages, explain Core Web Vitals and thresholds. Example: "Run the homepage performance test."
  • Security scanning — list security scans, trigger scans, review findings, check usage. Example: "Run the production security scan."
  • Test cases — create test cases and suites, start runs, show results. Example: "Create a test case for the password reset flow."
  • Notes — create notes in all 5 formats (Markdown, Plain Text, Bug Template, Checklist, Outline). Example: "Create a checklist note for the sprint 12 regression."
  • Time tracking — log hours by category. Example: "Log 2 hours of exploratory testing."
  • Jira sync — push reports to Jira, check sync status. Example: "Sync the checkout bug to Jira."
  • Team — list team members, check roles. Example: "Who's on my team?"

Teaching Mode

The assistant also functions as a QA mentor. Ask it about testing methodologies, best practices, or how to improve your workflow:

  • "What's the difference between regression and smoke testing?"
  • "Help me write better bug reports."
  • "Explain boundary value analysis."
  • "What should I include in a mobile test plan?"

Bug Reports

Bug reports are the foundation of bugAgent. Every report is a structured, enriched record that captures not just what went wrong, but the full context needed to fix it. Reports are created manually, by AI agents, through the SDK, from automation failures, from performance threshold breaches, or from test case failures — and every path produces the same rich, actionable output.

Report Types (19)

Every report is auto-classified into one of 19 types with a confidence score: Functional, UI/UX, Performance, Security, Crash, Data Integrity, Accessibility, Compatibility, Network, Logic, Feature Request, Enhancement, Technical Debt, Documentation, DevOps, UX Improvement, Integration, and more. No manual tagging required — just describe the issue and the AI handles classification.

Severity & Status

  • 4 severity levels — Critical, High, Medium, Low. Mapped to Jira priority on sync.
  • 8 statuses — New, Awaiting Triage, Confirmed, In Progress, Resolved, Retesting, Closed, Reopened. Drag cards on the Kanban board to change status instantly.

Quality Score (1–10)

Every report is automatically scored across 10 context-driven dimensions to measure the quality of the testing, not just the bug. The dimensions are: reproduction steps, expected vs actual behavior, environment details, evidence, root cause analysis, impact assessment, context and history, heuristics and oracles, clarity and structure, and actionability. Each dimension is scored 0.0–1.0 and combined into a final score. Visual badges (red 1–3, amber 4–6, green 7–10) appear on report cards and the Kanban board. Teams track their average score over time in Analytics.

Impact Score (0–100)

A composite metric combining severity, frequency, and affected users. Helps teams prioritize which bugs to fix first based on real-world impact rather than gut feeling.

Built-in Timer

Every report includes a timer you can start, stop, and resume to track exact testing time down to the second. Click the time display to manually adjust. The timer is available on both the creation form and the report detail page, so you always know how long a bug took to find and document.

Media Attachments

Attach screenshots, videos, PDFs, logs, and other files up to 10 MB each. Attachments are stored securely and displayed inline on the report detail page. When reports sync to Jira, attachments are pushed alongside.

Jira Sync

Push any report to Jira with one click. Fields are mapped automatically: severity to priority, type to labels, description to the Jira description field. Sync is bi-directional — changes in Jira flow back to bugAgent. Auto-sync polls periodically, and a force sync button is available on any report.

Claude Analysis

Send any report to Claude for root cause analysis. Claude returns a probable cause, affected areas, a suggested fix with code-level specifics, verification steps, and a risk assessment. Enable auto-push per project to analyze every new report automatically, or push manually from the dashboard, API, or MCP.

Duplicate Detection

When a new report is created, text similarity analysis compares titles and descriptions against existing reports in the same project. Potential duplicates are flagged with match scores and direct links, keeping backlogs clean and preventing duplicate work.

Kanban Board

Visualize reports across 8 status columns with drag-and-drop. Dragging a card updates the status in real time and syncs to Jira instantly. Toggle between list and kanban views with a persistent preference. Each card shows severity, type, description preview, and timestamps.

Auto-Created Reports

Reports are created automatically from multiple sources: Playwright automation failures (with video and error logs), performance test threshold breaches (with metrics and recommendations), mobile test failures (with device logs and screenshots), and test case execution failures (with full test context). Every auto-created report links back to its source for full traceability.

Access

Available via the dashboard, REST API (full CRUD), MCP server (create, list, get, update tools), the SDK/FAB, and the AI Assistant.

Performance & Load Testing

bugAgent provides comprehensive performance testing across web and mobile platforms. Run page quality audits, simulate concurrent load, and profile real mobile devices — all from the same dashboard. When performance degrades, bug reports are created automatically with full metrics and fix recommendations.

Web Page Quality Audits

Run audits that score your pages across four categories (0–100 each): Performance, Accessibility, Best Practices, and SEO. Each audit captures Core Web Vitals — LCP (Largest Contentful Paint), FID (First Input Delay), CLS (Cumulative Layout Shift), FCP (First Contentful Paint), Speed Index, TTFB (Time to First Byte), TTI (Time to Interactive), and Total Blocking Time. Results include actionable opportunities with estimated savings and detailed diagnostics.

Load Testing

Simulate concurrent virtual users hitting your endpoints with configurable duration and ramp-up. Measure requests per second (RPS), error rate, and latency percentiles at p50, p90, p95, and p99. Results stream to Grafana Cloud for real-time dashboard viewing. Identify bottlenecks before your users do.

Mobile Performance (BrowserStack App Profiling)

Profile real Android and iOS apps on physical devices via BrowserStack. Metrics captured include CPU usage (average and max %), memory consumption (average and peak MB), frames per second (average and minimum), battery drain (%), app launch time (ms), network data transferred (KB), and disk I/O (KB). Select from Google Pixel 7/8, Samsung Galaxy S23/S24, iPhone 14/15 series, and more. Toggle between Web and Mobile modes on the dashboard.

Auto-Bug Creation

Configure thresholds (for example, performance score below 50, CPU above 80%, p95 latency above 3000ms) and bug reports are created automatically when breached. Each auto-created report includes the full metrics snapshot, screenshot, recommendations from the audit, diagnostics, and a quality score. This turns performance monitoring into a proactive QA pipeline.

Core Web Vitals Tracking

Track Core Web Vitals over time to identify trends and regressions. The Performance Score Trend chart on the Analytics dashboard shows how your scores change across audits, helping you correlate deploys with performance shifts.

CSV Export & Archive

Export any performance run to CSV for reporting or offline analysis. Archive completed runs to keep your list clean, and view archived runs anytime with the status filter.

Performance Trends

The dedicated Trends page (Performance → Trends) charts all metrics over time across completed runs. Select any test config and a date range (7d, 30d, or 90d).

Web trends include a multi-line Lighthouse score chart (Performance, Accessibility, Best Practices, SEO), Core Web Vitals charts with threshold lines (LCP 2500ms, FCP 1800ms, CLS 0.1, TTFB 800ms), and load test metric charts (p95, RPS, error rate).

Mobile trends chart CPU average (%), memory peak (MB), FPS average, and launch time (ms). Use trends to track whether optimizations actually improve metrics over time.

The AI Assistant can answer questions about your trends: "How has my performance score changed?", "Is my LCP getting better?", "Are there any CWV regressions?", "What's the p95 trend for my load tests?"

Plan Limits

Pro: 1 performance test per month (includes trends). Team: 2 per month. Enterprise: 10 per month. Each device in a multi-device mobile test counts as 1 credit. Trends page is available on all paid plans at no extra cost. Performance+ add-on ($299/mo): unlimited tests for any paid plan.

Performance Analytics

The Analytics dashboard includes a Performance section showing web audit score averages, mobile device metrics, and performance score trend bars. Filter by time range and project to track improvements.

Security Scanning

bugAgent provides comprehensive vulnerability scanning across web applications, mobile binaries, and source code. Scans use industry-standard tools and produce findings categorized by severity with actionable remediation guidance. Findings above your configured threshold auto-create security bug reports with quality scores.

Web DAST (Quick Scanner + Nuclei)

Dynamic Application Security Testing scans live websites for OWASP Top 10 vulnerabilities. The Quick Scanner (~30 seconds) checks security headers, HTTPS enforcement, cookie security, exposed files (.env, .git), information disclosure, and server version leaks. Nuclei runs 4,000+ community vulnerability templates for CVE detection, misconfigurations, default credentials, and known exploits.

Scan Depth

Choose from three scan depths: Quick (~30 seconds) runs the fast header/exposure checks only. Regular (~5 minutes) adds focused Nuclei templates for common misconfigurations and vulnerabilities. Deep (up to 30 minutes) runs the full 4,000+ Nuclei template library for comprehensive OWASP Top 10 coverage.

Authenticated Scanning

Provide a login URL, username, and password to scan pages behind authentication. bugAgent uses Playwright to log into your app, captures session cookies, and passes them to all scan tools so they test authenticated pages rather than just the login screen. The scan report shows an "Authenticated" badge confirming the login succeeded.

Domain Verification

Before scanning a domain, verify ownership via DNS TXT record. Add a _bugagent-verify.yourdomain.com TXT record with the provided token. This prevents scanning domains you do not own. Verification is required for all web scans.

Mobile Binary Analysis (MobSF)

Upload APK (Android) or IPA (iOS) binaries for static analysis via MobSF. Analysis covers manifest/entitlement misconfigurations, hardcoded secrets and API keys, weak cryptographic implementations, dangerous permissions, certificate issues, tracker detection, and binary-level vulnerabilities. Each finding includes severity, CWE reference, OWASP Mobile mapping, and remediation guidance.

OWASP ZAP (Active Scanning)

When available, OWASP ZAP runs a baseline scan via Docker alongside Quick Scanner and Nuclei. ZAP spiders the target, performs passive analysis, and detects reflected XSS, CSRF, session management issues, and other dynamic vulnerabilities that passive tools cannot find. Session cookies from authenticated scanning are passed to ZAP automatically.

Code SAST (Semgrep)

Connect a public or private GitHub, GitLab, or Bitbucket repository. bugAgent clones the repo (shallow, depth=1), runs Semgrep with community rules across 30+ languages, plus dependency audit and license scanning. Findings include file path, line number, code snippet, CWE reference, and remediation guidance.

Three depth levels: Quick (OWASP Top 10 + secrets detection), Regular (auto + OWASP + secrets + CWE Top 25 + dependency audit), Deep (all of above + security-audit + supply-chain + JWT + crypto + command injection + license scanning).

Dependency audit detects vulnerable npm and pip packages with CVE references and fix versions. License scanning flags GPL, AGPL, and other copyleft licenses that may require compliance review for commercial use.

Dependency Vulnerability Checks

Code scans (Regular and Deep) automatically detect vulnerable dependencies in your project. bugAgent runs npm audit for Node.js and pip-audit for Python projects, auto-detecting package managers from lock files (package-lock.json, requirements.txt, yarn.lock, pnpm-lock.yaml, Cargo.lock, go.sum, Gemfile.lock, composer.lock).

Each vulnerable package is reported with its CVE identifier, affected version range, recommended fix version, and severity. Findings appear in the scan report with the dependency-audit tool badge and are included in CSV exports and auto-bug creation.

License Compliance Scanning

Deep code scans check the full dependency tree for license compliance risks. bugAgent flags copyleft licenses that may require source code disclosure or restrict commercial use:

  • High risk — GPL-2.0, GPL-3.0, AGPL-3.0 (strong copyleft, may require releasing your source code)
  • Medium risk — LGPL, MPL, EPL, EUPL (weak copyleft, may have linking exceptions)
  • Unknown — UNLICENSED or missing license metadata (review recommended)
  • Low risk — MIT, Apache-2.0, BSD, ISC (permissive, not flagged)

License findings appear in the scan report with the license-scan tool badge. Ask the AI Assistant about your license risks: "Are there any GPL dependencies in my project?" or "Is my project safe for commercial use?"

Auto-Bug Creation

Configure a severity threshold (Critical, High, Medium, or Low) to auto-create security-type bug reports for qualifying findings — including code vulnerabilities, dependency CVEs, and license compliance issues. Each bug report includes the details, remediation steps, a link back to the full scan report, and an AI quality score.

Security Score

Each scan produces a 0–100 security score. Points are deducted per finding: Critical −25, High −10, Medium −5, Low −2. The score appears on the scan report page with a color-coded gauge (green/amber/red) and severity breakdown cards.

Plan Limits

Pro: 2 security scans per month. Team: 5 per month. Enterprise: 20 per month. Security+ add-on ($499/mo): unlimited scans for any paid plan.

Scheduled Tasks

Scheduling turns bugAgent automations into a continuous testing pipeline. Set any web automation to run automatically on a cron schedule, choose the device viewport, configure notifications, and let bugAgent do the rest. Failures create bug reports automatically, so regressions are caught before users report them.

Cron Scheduling

Use standard 5-field cron expressions for maximum flexibility: run daily, weekdays only, specific days of the week, hourly, or any custom pattern. The schedule page shows a calendar-style view of upcoming and past runs.

Device Selection

Choose from 30+ virtual viewport devices (desktops, laptops, iPhones, Android phones, tablets) or BrowserStack live browsers (Chrome, Firefox, Safari, Edge on Windows and macOS). The device is locked in when you create the schedule, so every run tests the same viewport consistently.

Failure Notifications

Configure notifications per schedule: email, Slack, or both. When a scheduled run fails, your team is alerted immediately with the error details and a link to the run results. No more finding out about regressions from users.

Auto-Bug Creation

Failed automation runs automatically create comprehensive bug reports including video evidence, screenshots, error logs, and line-by-line pass/fail highlighting from the script. The bug report links back to the automation and the specific run for full traceability.

Calendar View

The Scheduled page displays all schedules in a calendar-style layout showing upcoming runs, past results (pass/fail), and schedule details. Filter by automation or status to focus on what matters.

Concurrency

Up to 5 parallel runs on the runner service. This means multiple schedules can fire at the same time without queuing, so your overnight regression suite finishes faster.

Use Cases

  • Nightly regression — schedule your full test suite to run every night at 2 AM. Review results in the morning standup.
  • Post-deploy smoke test — trigger a quick smoke test after every deployment to catch breaking changes immediately.
  • Cross-device validation — schedule the same automation on multiple devices (desktop, iPhone, Galaxy) to ensure responsive layouts work everywhere.
  • Monitoring — run a critical path test every hour to detect outages before your monitoring tools do.

Notes

Notes are bugAgent's scratch pad for testers. Capture observations during exploratory testing sessions, plan upcoming test cycles, or document findings that aren't ready to become bug reports yet. Notes support five formats, voice dictation, a built-in timer, file attachments, and private or shared visibility.

Five Formats

  • Markdown — full Markdown support with headers, lists, code blocks, links, and emphasis. Best for structured documentation and detailed session notes.
  • Plain Text — simple unformatted text. Fast to write, easy to read. Good for quick observations.
  • Bug Template — a pre-structured template with sections for summary, steps to reproduce, expected behavior, actual behavior, and environment. Ideal for notes that might become bug reports.
  • Checklist — checkbox items you can check off as you go. Perfect for test execution checklists, pre-release verification lists, and deployment checklists.
  • Outline — hierarchical bullet points for organizing complex information. Great for test planning, feature breakdowns, and meeting notes.

Voice Dictation

Click the microphone icon to dictate your notes using OpenAI Whisper transcription. Speak naturally and the text appears in your note. Especially useful during hands-on testing when you can't type.

Built-in Timer

Every note includes a timer you can start, stop, and resume. Click the time display to manually edit the duration. When you convert a note to a bug report, the tracked time transfers automatically — so the effort you spent investigating is captured on the report.

Visibility

Notes can be private (visible only to you) or shared (visible to your entire team). Use private notes for personal observations during exploratory testing, and shared notes for test plans and session summaries the team needs to see.

File Attachments

Attach files up to 10 MB each — screenshots, logs, documents, or any file that provides context for your observations.

Search & Filter

Find notes quickly with keyword search, and filter by project, author, and date. Notes are also searchable through the AI Assistant.

Use Cases

  • Exploratory testing sessions — use Markdown or Plain Text to capture observations as you explore. Start the timer, test, and record what you find. Convert promising findings to bug reports.
  • Standup notes — create a shared Checklist note before standup with items to discuss. Check them off during the meeting.
  • Test planning — use Outline format to break down upcoming test cycles by feature, priority, and assignee.
  • Bug investigation — use Bug Template format to document your investigation steps before filing a formal report. The timer tracks how long the investigation took.

Time Tracking

Time Tracking lets QA teams log daily effort by category, giving leads and managers visibility into how testing time is spent. Available on Team and Enterprise plans, with entries accessible via the dashboard, REST API, and MCP.

24 Categories

Every time entry is tagged with one of 24 categories that cover the full QA lifecycle:

  • Testing, Exploratory Testing, Test Case Writing, Test Execution, Regression Testing
  • Bug Reporting, Bug Verification, Test Planning, Environment Setup
  • WCAG/Accessibility Testing
  • Development, Code Review, Bug Fixing, Architecture/Design
  • Documentation, DevOps/CI-CD, Debugging, Research/Spike, Deployment
  • Meeting, Special Project, Communication, Admin, Planning

Filters & Views

Filter entries by period (today, week, month, all), project, category, and team member. Sort by newest, oldest, most time, or least time. The Analytics dashboard includes daily, weekly, and monthly time tracking summaries with bar charts broken down by category.

AI Assistant Integration

Log time through natural conversation: "Log 2 hours of exploratory testing" or "Add 45 minutes of bug triage time." The assistant creates the entry with the right category and duration. Ask "What time tracking categories are available?" to see the full list.

Use Case: QA Team Leads

QA team leads use time tracking to understand where testing effort is concentrated. If 60% of time goes to bug triage and only 10% to exploratory testing, that signals a process problem. The Analytics integration surfaces these patterns automatically, helping leads rebalance effort toward higher-value activities like test planning and automation development.

bugAgent Tools (SDK)

The bugAgent SDK is a floating action button (FAB) that installs on any website with a single line of code. It gives testers, developers, and end users five powerful tools for capturing bugs with rich context — no login required for reporters. The AI analyzes the captured data and auto-drafts structured bug reports.

Installation

<script src="https://app.bugagent.com/sdk/bugagent-sdk.js" data-key="YOUR_API_KEY" data-position="bottom-right"></script>

Get your API key from Settings → Developers in the dashboard. The FAB appears as a small button on your site. Click it to expand the toolbar with all five tools. Available on Pro and Team plans.

Five Tools

  • Automate — captures the last 60 seconds of user activity: DOM mutations, clicks, navigation, form interactions, console errors (last 50 entries with full stack traces), and failed network requests. This context is attached to bug reports automatically, giving developers a complete picture of what happened before the bug was reported.
  • Recorder — records screen and voice with one click. The video is attached to the bug report automatically. The recording runs in a separate popup window (you may need to allow popups from app.bugagent.com in your browser settings).
  • Annotate — marks up screenshots with a 7-color drawing tool. Draw circles, arrows, and freehand annotations directly on the page before submitting. Annotated screenshots are attached to the report.
  • WCAG Analyzer — runs axe-core accessibility checks and groups results by severity (critical, serious, moderate, minor) with WCAG rule IDs and affected elements. Use this to quickly identify accessibility violations on any page.
  • Rewind — replays the last 5 user actions with a ghost cursor animation. Developers can see exactly what the user did leading up to the bug, without asking them to describe it.

How the AI Auto-Drafts Reports

When a user clicks "Submit" on the FAB, the SDK bundles all captured data — session replay, console logs, network waterfall, performance metrics, screenshots, annotations, and recordings — and sends it to bugAgent. The AI analyzes the bundle and drafts a structured bug report with a title, description, type classification, severity assessment, and environment details. The user can review and edit before final submission.

Use Case: Staging Site Testing

Install the SDK on your staging environment. Testers navigate the site normally. When they find a bug, they click the FAB, annotate the screenshot, optionally record their screen and voice, and hit submit. The AI creates a detailed bug report with full context. No forms to fill out. No login required. The tester stays in flow and the developer gets everything they need to reproduce and fix the issue.

No Login Required

The FAB works without authentication for reporters. Anyone on your site can report bugs. This makes it ideal for external testers, beta users, and stakeholders who don't have bugAgent accounts. Reports are created in your project automatically.

Integrations

bugAgent connects to the tools your team already uses. All integrations are configured from Settings → Integrations in the dashboard. Once connected, they operate automatically in the background — syncing data, sending notifications, and enriching your QA workflow.

Jira

The Jira integration provides bi-directional sync between bugAgent and Jira via OAuth 2.0. Push any report to Jira with one click, and updates flow back automatically.

  • OAuth 2.0 connection — connect once and the entire team shares the connection. No individual API keys needed.
  • Field mapping — bugAgent severity maps to Jira priority, types map to labels, descriptions sync with formatting preserved, and attachments are pushed alongside.
  • Bi-directional sync — changes in Jira (status, comments, priority) flow back to bugAgent. Auto-sync polls periodically. A force sync button is available on any report.
  • Per-report and bulk sync — push individual reports or configure auto-sync to push every new report as it's created.
  • Merge conflicts — when both sides have changed, use the merge tool to choose which values to keep (bugAgent or Jira) per field.

Claude AI

Connect your Anthropic API key to enable AI-powered analysis across bugAgent.

  • Root cause analysis — send any report to Claude for analysis. Returns probable cause, affected areas, suggested fix with code-level specifics, verification steps, and risk assessment.
  • Quality scoring — Claude powers the Quality Score dimensions that evaluate every report.
  • AI Assistant — the dashboard assistant uses Claude to understand context and take actions on your behalf.
  • Per-project configuration — enable auto-push to analyze every new report automatically. Add custom instructions (tech stack, architecture, focus areas) for more relevant analysis.

GitHub

Connect a GitHub repository to sync Playwright automation scripts. Scripts are pushed to GitHub on save and update, with branch and commit management included. This keeps your test scripts version-controlled alongside your application code.

Slack

Receive failure notifications from scheduled automation runs directly in Slack. Configure per schedule: email, Slack, or both. When a scheduled test fails, your team is alerted in the configured Slack channel with error details and a link to the run results.

The Problem

Traditional bug management systems were built for humans clicking through web interfaces. They assume someone will manually fill out a form, pick a category from a dropdown, assign a priority, and tag the right team. This model breaks down in the age of AI-assisted development:

🚫
GUI-first design

Most trackers are built around dashboards and forms. Agents can't easily navigate web UIs — they need APIs and protocols.

📋
Rigid, manual classification

Bugs require humans to pick types, priorities, and labels. This creates inconsistency and adds friction for both agents and developers.

🔄
Context switching

Developers bounce between their IDE, a browser-based tracker, and chat tools. Every switch breaks focus and costs time.

🛑
No agent identity

Traditional platforms don't support the concept of an AI agent as a first-class participant. Agents can't own bugs, track their own work, or communicate with other agents.

The Agentic Development Cycle

Software development is shifting from a human-only process to a collaborative cycle between humans and agents. In this new paradigm, AI agents write code, review pull requests, run tests, and monitor production — all alongside human developers.

A QA AI Agent fits naturally into this cycle as the quality layer — continuously scanning for issues, filing enriched bug reports, and coordinating fixes. But for this to work, the QA agent needs a platform that speaks its language: structured data over APIs, machine-readable classifications, and protocol-native communication.

bugAgent is that platform. It sits at the intersection where agents, humans, and tools converge — providing a universal protocol for quality that works across any IDE, CI system, or communication platform.

The Solution

bugAgent is an API-first bug enrichment platform designed for agents. Every feature, from report creation to Jira sync, is available through structured APIs and the Model Context Protocol — not locked behind a web UI.

API-First Infrastructure

What does API-first mean for a bug management platform?

🌐 REST API

Every operation available in the dashboard is also available via the REST API. Create reports, manage projects, generate API keys, sync to Jira — all programmatically.

🔌 Model Context Protocol

Native MCP support means any MCP-compatible client (Claude Desktop, Cursor, Windsurf, Claude Code) can use bugAgent as a tool — no integration code needed.

🔑 API Key Authentication

Agents authenticate with API keys, not browser sessions or OAuth flows. Generate keys from the dashboard, set them as environment variables, and your agent is ready.

📦 Structured Data

Every response is JSON. Every classification includes a type, severity, and confidence score. No scraping, no parsing — just clean, predictable data structures.

⚙️ Webhooks & Events Coming Soon

Subscribe to bug lifecycle events — new reports, status changes, Jira syncs — and react in real-time from your own services.

📊 Programmatic Analytics

12+ interactive charts — bug trends, quality scores, severity/status/type breakdowns, automation health, time tracking, and a Quality Testing Health Score gauge (0–100). Query via GET /analytics API or MCP get_stats tool. See full details →

⚡ Team Booster

Scale your QA team on demand. Provision pre-configured tester accounts by specifying team size, location, technical level, and budget. Testers are created instantly and appear in Team Management. Pro and Team plans only.

AI-Native Features

bugAgent doesn't just store bugs — it understands them:

Auto-Classification

Describe any issue in natural language. bugAgent classifies it across 19 types — from bugs (UI, performance, crash, security, logic, data, network, accessibility, compatibility) to feature requests, enhancements, technical debt, documentation, DevOps, UX improvements, and integrations — with a confidence score. No dropdowns, no manual tagging.

Enrichment Pipeline

Raw bug reports are transformed into structured, actionable records. Missing fields are inferred, severity is assessed, and context is added automatically.

Agent Identity

Agents are first-class citizens. They have their own accounts, API keys, and usage tracking. An agent's bug reports are attributed to it, creating a clear audit trail.

Duplicate Detection

Every new report is automatically compared against existing reports in the same project using text similarity on both titles and descriptions. Potential duplicates are flagged with match scores and linked directly on the bug detail page, keeping backlogs clean.

Quality Score — Testing Quality Measurement

Every report is scored 1–10 to measure the quality of the testing, not just the bug. The score evaluates 10 context-driven dimensions — reproduction steps, expected vs actual behavior, environment details, evidence, root cause analysis, impact assessment, context & history, heuristics & oracles, clarity & structure, and actionability — each scored 0.0–1.0 and combined into a final score. Teams track their average quality score trending over time in Analytics, and the Quality Testing Health Score (0–100) gives leadership a single metric for testing effectiveness. Visual badges (red 1–3, amber 4–6, green 7–10) appear on report cards, the kanban board, and API responses.

Protocol-Native

MCP support means agents don't need custom integration code. They communicate with bugAgent the same way they communicate with any other tool in their ecosystem.

Claude Integration

Self-healing bug analysis powered by Anthropic.

Connect your Anthropic API key to push bug reports to Claude for root cause analysis and fix suggestions. Claude analyzes each report and returns actionable insights including likely causes, affected areas, and recommended fixes. Available on Pro and Team plans.

How to Connect

Navigate to Settings → Integrations in the dashboard. In the Claude by Anthropic section, enter your Anthropic API key (starts with sk-ant-) and select your preferred model (Claude Sonnet 4, Claude Opus 4, or Claude Haiku 3.5). Click Connect to save.

Per-Project Configuration

Once connected, configure Claude on a per-project basis:

  • Auto-push bugs to Claude — toggle this on to automatically send every new bug report to Claude for analysis as soon as it's created
  • Custom instructions — add project-specific context such as your tech stack, architecture patterns, or areas of focus (up to 5,000 characters). Claude uses these instructions to provide more relevant analysis

Manual Analysis

From any bug report detail page, click Send to Claude to push the report for analysis on demand. If an analysis already exists, click Re-analyze to get a fresh evaluation. The analysis appears inline on the report page with formatted root cause details and fix suggestions.

API and MCP Access

Claude analysis is available through multiple interfaces:

  • REST API — call POST /api/claude/push with a report_id to trigger analysis programmatically. The response includes the full analysis text and timestamp. See the API reference for details.
  • MCP tool — use the push_to_claude tool from any MCP client (Claude Desktop, Cursor, VS Code, Claude Code). Pass a report_id and receive the analysis directly in your AI workflow. See the MCP documentation.
  • Dashboard — click Send to Claude on any bug report detail page for on-demand analysis with a visual interface.

The Self-Healing Cycle

bugAgent as the orchestration layer for autonomous software quality.

bugAgent connects QA, development, and AI into one continuous improvement cycle. When all components are active, software effectively heals itself — with humans reviewing and guiding every step.

1. Record — Capture Issues From Every Source

Issues enter the system through multiple channels, each producing structured, enriched bug reports:

  • FAB tools (SDK) — the floating action button captures DOM recordings, console logs, network waterfalls, performance metrics, and screenshots directly from the browser. The AI auto-drafts bug reports from session data.
  • Playwright automation — scheduled and on-demand test scripts detect regressions and failures automatically. Failed runs generate bug reports with video evidence and error details.
  • AI assistant — the dashboard and in-app AI assistant lets team members create detailed bug reports through natural conversation, including voice input and file attachments. Also supports triggering web and mobile performance tests, listing runs, and reviewing results.
  • MCP agents — AI agents running in Claude Desktop, Cursor, VS Code, or Claude Code file bug reports directly through the Model Context Protocol. CI/CD agents can report issues as part of the pipeline.
  • Email — send bugs via email and they are automatically classified, filed, and synced to your project tracker.

2. Diagnose — Claude Analyzes Root Cause

Once a bug report exists, Claude can analyze it to determine the probable root cause and suggest a fix:

  • Auto-push — enable per-project to automatically send every new bug report to Claude the moment it is created
  • Manual push — click Send to Claude from the dashboard, or use push_to_claude via MCP, or call POST /api/claude/push via REST API
  • Structured analysis — Claude returns probable cause, affected areas, a suggested fix with code-level specifics, verification steps, and a risk assessment
  • Custom instructions — per-project context (tech stack, architecture, focus areas) makes analysis more accurate and relevant

3. Automate — Verify Fixes Continuously

Playwright automation scripts provide the verification backbone:

  • Scheduled runs — tests run on a cron schedule to catch regressions before users do
  • On-demand runs — trigger tests via the dashboard, API, or MCP after applying a fix
  • 30+ device targets — run tests on desktops, laptops, iPhones, Android phones, and tablets. Select the device when triggering manual or scheduled runs
  • AI optimization — send any script to Sonnet 4 for a 12-point optimization that fixes selectors, wait strategies, assertions, error handling, auth patterns, mobile compatibility, and strict mode issues
  • Version history & undo — every script tracks up to 10 previous versions. One-click undo reverts to the prior version. Versions are saved before manual edits, AI optimization, and script regeneration
  • Line-by-line results — after a run completes, the script highlights passed lines in green and failed lines in red so you can see exactly which step broke
  • CI/CD integration — run tests as part of your deployment pipeline and auto-file bugs on failures
  • Video evidence — every run produces video results so teams can see exactly what happened

4. Heal — AI Agents Close the Loop

Connected AI systems can read Claude's analysis and act on it autonomously:

  • Read analysis via MCP — agents call get_bug_report to retrieve the Claude analysis including root cause and suggested fix
  • Apply fixes — coding agents (Claude Code, Cursor, custom agents) read the suggested fix and apply code changes directly
  • Trigger re-verification — after applying a fix, agents call run_automation to trigger the relevant Playwright tests
  • Update status — agents update the bug report status and add comments documenting what was changed and verified

5. Human in the Loop — Review, Approve, Guide

Humans remain essential at every stage of the cycle:

  • Review — teams review Claude's analysis and AI-applied fixes before merging to production
  • Approve — pull request reviews, test result sign-offs, and deployment approvals keep humans in control
  • Guide — custom instructions, project configuration, and priority settings direct how the system behaves
  • Override — any automated action can be paused, reversed, or manually corrected at any time

The Full Picture

When all components are connected, bugAgent becomes the orchestration layer for a full-circle autonomous QA pipeline:

Record → Diagnose → Automate → Heal → Record …

Each cycle produces better software. Bugs are caught earlier, fixes are more targeted, regressions are detected faster, and the AI learns from project-specific context. Teams spend less time on triage and more time building features — while bugAgent orchestrates the quality lifecycle around them.

The self-healing cycle works best with Claude integration (Pro or Team plan), Playwright automation, and at least one MCP-connected AI agent. Your Anthropic API key is stored securely and only used to send bug report data for analysis. You can disconnect at any time from Settings → Integrations.

bugAgent Skills

Integrations that connect external tools directly into bugAgent's context engine.

Skills are first-class integrations that plug external services into bugAgent. Each Skill enriches the context engine with data from tools your team already uses — syncing repositories, pulling analysis, bridging issue trackers, and more. Once enabled, Skills operate automatically in the background, keeping your QA data connected and up to date.

Available Skills

GitHub

Sync repositories to power Playwright automation scripts. Connect a repo, map it to a project, and bugAgent pulls test files and keeps scripts in sync with your codebase.

Claude

Push bug reports to Claude for root cause analysis. Returns probable cause, affected areas, a suggested fix with code-level specifics, and a risk assessment. Available on Pro and Team plans.

Jira

Bi-directional bug sync between bugAgent and Jira. Reports, comments, attachments, and severity flow both ways with auto-sync and manual merge controls.

How to Enable Skills

Navigate to Settings → Integrations in the dashboard. Each Skill has its own connection flow — OAuth for GitHub and Jira, API key for Claude. Once connected, configure per-project settings such as auto-push, repository mapping, or default Jira project.

Building Custom Skills

Partners and developers can build their own Skills to extend bugAgent. Custom Skills connect external tools via OAuth or API keys, feeding data into the context engine alongside the built-in integrations.

  • Submit an integration — reach out via the homepage or email support@bugagent.com to propose a new Skill
  • Build programmatically — use the API reference to build Skills that interact with bugAgent endpoints for reports, automation, and project data
🔗
Skills are available on all plans. Some Skills (Claude analysis) require a Pro or Team subscription. Visit Settings → Integrations to see which Skills are available for your plan.

Free Migration

Switch to bugAgent with zero friction — we handle the heavy lifting.

Moving from another bug tracking or QA tool? The bugAgent team will migrate your data for free. We export your existing bug reports, test scripts, and project configuration, then import everything into your new bugAgent workspace.

What's Included

  • Bug report export — we export your existing bug reports from your current tool and import them into bugAgent with full metadata preserved
  • Playwright automation scripts — we walk you through creating Playwright automation scripts tailored to your product's test scenarios
  • Integration configuration — we help you connect Jira, GitHub, Slack, and Claude so your workflow is fully operational from day one
  • Team onboarding — we set up your team accounts, projects, and permissions to match your existing structure
📧
Ready to migrate? Contact hello@bugagent.com to get started. Migrations are free for all plans and typically complete within a few business days.

Dedicated QA Team

Professional QA engineers embedded in your workflow.

For teams that want hands-on support, bugAgent offers a dedicated QA team service. Our professional QA engineers work directly inside your bugAgent workspace to review, enrich, and act on your bug reports.

What the QA Team Does

  • Bug report review and enrichment — our QA engineers review incoming reports, add reproduction steps, verify severity, and ensure every bug meets a high quality bar before it reaches your developers
  • Automation script development — we develop Playwright automation scripts for your product, covering critical user flows, regression scenarios, and edge cases
  • Custom features and integrations — need a custom Skill, a bespoke reporting dashboard, or a CI/CD pipeline integration? Our team builds it for you
  • Ongoing QA operations — continuous testing, triage, and quality monitoring tailored to your release cadence and priorities
📧
Interested in a dedicated QA team? Contact hello@bugagent.com to discuss your needs and get a custom proposal.

Test Cases

bugAgent includes a complete test case management system inspired by TestRail, built for modern QA teams. Manage your entire testing lifecycle from case creation to execution reports.

Core Concepts

  • Test Cases — individual test scenarios with a name, description, preconditions, and ordered steps. Each step has an action (what to do) and an expected result (what should happen). Drag-and-drop step reordering lets you reorganize steps without recreating them. Cases include priority (critical, high, medium, low), type (functional, regression, smoke, integration, e2e, performance, security, usability, accessibility), tags, and estimated execution time. Use Rapid Mode for fast test case creation when you need to build a large suite quickly
  • Test Suites — collections of related test cases. Suites let you organize cases by feature, sprint, or any grouping. Each suite tracks its case count and last run status. Color-coded cards help you see suite health at a glance
  • Test Runs — an execution session created from a suite. When you create a run, all cases from the suite are snapshotted at that point in time. Assign runs to team members with notifications and track progress as each case is marked pass, fail, blocked, or skip with auto-save
  • Test Reports — aggregated results from completed runs with pass rate, date range filtering, per-suite breakdowns, color-coded cards, and CSV export

Workflow

  1. Create test cases with detailed steps describing the action and expected result
  2. Group related cases into test suites (e.g. "Checkout Regression", "Login Smoke Tests")
  3. Create a test run from a suite and assign it to a team member with notifications
  4. Execute the run: mark each case as passed, failed, blocked, or skipped with notes — results auto-save
  5. When a case fails, instantly create a bug report with full context from the test case — the bug report links back to the test case and test run
  6. Re-run only the failed cases after fixes are deployed
  7. Review reports with pass rates and trends over time, export to CSV

Full Traceability

Every bug report traces back to its origin: manual testing, automation failure, test case run, AI assistant, or FAB SDK. The entire QA lifecycle is connected:

  • Test case runs link to bug reports — when a case fails and you create a bug report, the report includes full context: test case name, steps, expected vs actual, and the test run it came from
  • Bug reports link to test cases and automation scripts — every bug report shows its origin, whether it was filed from a test run, an automation failure, the AI assistant, or the SDK
  • Automation failures create linked bug reports — when a Playwright script fails, the generated bug report links back to the automation script with line-by-line results
  • End-to-end chain — test case → test run → bug report → fix → re-test. No broken chains, no lost context

Access

Test Cases are available through the dashboard, the REST API (full CRUD for cases, suites, runs, and reports), and the MCP server (7 tools for AI agent integration). Duplicate cases with one click, reorder cases within suites via drag-and-drop, and filter by priority, type, status, and tags.

Geo-Snap

Geo-Snap lets you enter any URL and capture screenshots of how a website appears from 50+ countries simultaneously. Powered by real geo-IP routing, every screenshot shows exactly what a visitor in that country would see — including localized content, regional pricing, language variants, and geo-specific redirects.

🌍
Geo-Snap is available on all plans. Free users capture 1 country at a time with 10 saved screenshots. Pro and Team plans support up to 5 countries simultaneously with unlimited saved screenshots.

How It Works

  1. Enter a URL — paste any public website address into the Geo-Snap input
  2. Select countries — choose from 50+ countries (United States, United Kingdom, Germany, Japan, Brazil, Australia, and many more)
  3. Capture — screenshots are taken from real IP addresses in each selected country and delivered within seconds

Key Use Cases

  • Localization verification — confirm that currency, language, date formats, and regional content display correctly for each market
  • Geo-redirect testing — verify that visitors are redirected to the correct regional version of your site (e.g., .co.uk, .de, .jp)
  • Compliance checks — ensure cookie consent banners, GDPR notices, and other regulatory requirements appear in the correct regions
  • Competitive monitoring — see how competitor websites look and price products across different markets

Plan Limits

Free
1 country per capture, 10 saved screenshots
Pro / Team
Up to 5 countries per capture, unlimited saved screenshots

Access

Geo-Snap is available through the dashboard, the REST API (list, create, delete), and the MCP server (create_geo_snap tool for AI agent integration).

Mobile Testing

Mobile Testing lets you upload APK (Android) or IPA (iOS) apps and run automated tests on real BrowserStack devices — Pixel, Galaxy, iPhone, iPad, and more. Record test actions in-browser and let AI generate production-grade scripts, or write your own. Every run produces video recordings, Appium logs, device logs, and network logs. Failed runs automatically create comprehensive bug reports with failure snapshots and step breakdowns.

📱
Mobile Testing is available on all plans. Free users get 1 run per day. Pro and Team plans include unlimited runs.

Record & Generate Scripts

Walk through your app on a virtual device directly in the browser. Log tap, swipe, input, and back actions as you go. When you stop recording, AI (Claude Sonnet 4) generates three production-grade scripts simultaneously:

YAML

Maestro-format YAML with assertVisible after actions, waitForAnimationToEnd for loading screens, retryTapIfNoChange for flaky elements, and descriptive comments grouping the user flow.

Appium Python

pytest with Appium 2.x client, UiAutomator2/XCUITest options, explicit WebDriverWait (never sleep), W3C Actions for gestures, Page Object hints, and docstrings.

Appium JavaScript

WebDriverIO v8+ with async/await, Mocha-style describe/it blocks, waitForDisplayed, W3C touch actions, before/after hooks, JSDoc comments, and modern ES6+ best practices.

Save any script as an automation and run it on real devices immediately.

Supported Platforms

  • Android — upload .apk files. Works directly for both recording and running on real devices (Pixel, Galaxy, OnePlus, Xiaomi).
  • iOS — upload .ipa files for running tests on real iPhones and iPads. For recording, also upload a simulator build (.app zipped) — build from Xcode targeting “iPhone Simulator” and zip the .app bundle.
ℹ️
iOS recording requires two files: The .ipa runs on real devices, while the simulator .app enables in-browser recording. Both represent the same app built for different targets. Upload the simulator build on the app detail page to unlock the Record Test button.

Update App Version

Replace your app binary with a new version without creating a new entry. Click Update App on the app detail page and upload the new APK or IPA. The old file is automatically deleted from storage to save space, and all linked automations and scheduled runs will use the new version on the next execution — no reconfiguration needed. For iOS apps, replacing the IPA also removes the simulator build — you’ll need to re-upload the updated simulator .app to continue recording.

Run on Real Devices

Click Run Now on any automation to launch it on a real BrowserStack device. Select from popular Android and iOS devices — the dropdown automatically filters by your app’s platform. All script types (YAML, Appium Python, Appium JS) run through BrowserStack’s real device farm via the Appium 2.x W3C WebDriver protocol.

Run Detail Page

  • Real-time polling — status auto-updates every 10 seconds while the test is queued or running, with a spinner banner
  • Video recording — watch the full test execution on the device
  • Searchable logs — Appium logs, device logs, and network logs with text search and match highlighting
  • Duration, started, completed timestamps
  • Rerun — one-click rerun with the same device and automation
  • Archive — archive completed runs to keep the list clean. View archived runs with the status filter

Bug Reports from Failures

When a mobile test fails or errors, bugAgent automatically creates a detailed bug report including:

  • Error message in a code block
  • Test step breakdown with pass/fail icons and counts (e.g. “3/5 passed”)
  • Device, platform, OS version in an environment table
  • Failure snapshot as the first attachment (last screenshot = failure state)
  • Video recording of the full run
  • Full logs in a collapsible section (up to 5,000 characters)

The bug report links back to the run detail page and the automation. The run listing shows a red “Bug Report” link for failed runs.

Mobile Analytics

The Analytics dashboard includes a Mobile Testing row showing: total runs, pass/fail rate donut, runs by device bar chart, and Android/iOS platform split. Visible when you have mobile run data.

Scheduling & Notifications

Schedule mobile test runs using cron expressions with timezone support. Configure Slack or email notifications on failure so your team is alerted immediately when regressions are detected.

Plan Limits

Free
1 app, 3 runs/day. Unlimited script generation. Resets at midnight EST.
Pro
3 apps, 20 runs/day. Unlimited script generation. Resets at midnight EST.
Team
10 apps, 75 runs/day. Unlimited script generation. Resets at midnight EST.
Enterprise
Unlimited apps and runs.
Mobile+ Addon
$299/mo for Pro or Team plans. Unlimited apps and runs. No daily limits.

Setup

Configure your BrowserStack credentials (BROWSERSTACK_USERNAME and BROWSERSTACK_ACCESS_KEY) in your environment. Upload apps and create automations via the dashboard, REST API, or MCP tools (upload_mobile_app, create_mobile_automation, run_mobile_test, list_mobile_runs).

Security & Two-Factor Authentication

bugAgent supports two-factor authentication (2FA) for all email/password accounts. Users can enable 2FA from Settings → Security to add an extra layer of protection to their account.

🔒
2FA is available on all plans. Google OAuth users are managed by Google’s security settings.

Supported Methods

Authenticator App (TOTP)

Scan a QR code with Google Authenticator, Authy, 1Password, or any TOTP-compatible app. Enter a 6-digit code each time you log in. The issuer shows as “bugAgent” in your authenticator app. This is the recommended method — no external dependencies and works offline.

SMS Verification

Receive a 6-digit code via text message. Choose from 30+ country codes with a dropdown selector (no need to type the prefix). Requires Twilio configuration on the backend. Codes are delivered via Supabase Auth + Twilio integration.

How It Works

  1. Enable: Go to Settings → Security → choose Authenticator App or SMS → scan QR code or enter phone number → verify with 6-digit code.
  2. Login: Enter email + password → redirected to 2FA verification page → enter code from your app or SMS → access granted.
  3. Google OAuth: 2FA is skipped entirely. Google manages security for OAuth users.
  4. Disable: Settings → Security → Disable Two-Factor Authentication.

Verification Page

The 2FA verification page at login features a 6-digit code input with auto-advance between digits, paste support, and auto-submit when all 6 digits are entered. If both methods are enrolled, users can switch between “Use SMS instead” and “Use authenticator app instead.”

Middleware Protection

The dashboard middleware checks the Authenticator Assurance Level (AAL) on every request. If a user has 2FA enabled but hasn’t completed AAL2 verification, they are automatically redirected to the verification page. This prevents bypassing 2FA by navigating directly to dashboard URLs.

Storage Security

Organization storage usage is tracked at the team level. When usage reaches 80%, an amber warning banner appears across all dashboard pages. At 100%, a red banner indicates uploads are blocked. Contact hello@bugagent.com to increase storage limits.

Enterprise Security

Enterprise plan includes SSO/SAML, audit logs, on-premise deployment option, and multiple organization support. Contact sales for custom security requirements.

Analytics Dashboard

The Analytics Dashboard is a full-screen, interactive reporting suite available on Pro and Team plans. It provides real-time visibility into your team's testing health, bug trends, and QA performance across every project.

📊
Access Analytics from the left sidebar in the dashboard, or programmatically via GET /analytics (API Reference) and the MCP get_stats tool.

Quality Testing Health Score

At the top of the dashboard, a large gauge displays your overall Quality Testing Health Score from 0 to 100. This score is computed from four equally weighted components:

Quality (25%)

Average quality score across all bug reports in the period. Higher-quality reports with complete reproduction steps, evidence, and impact assessments push this component up.

Resolution (25%)

Percentage of bugs resolved or closed within the time range. Teams that close bugs quickly score higher here.

Automation (25%)

Pass rate across all Playwright automation runs. Healthy test suites with few failures drive this component.

Low Severity (25%)

Ratio of low/medium severity bugs vs. critical/high severity. A higher proportion of minor issues indicates a more stable product.

The component with the highest individual score displays a crown icon, highlighting your team's best-performing area at a glance.

Chart Sections

Below the health score, the dashboard includes 12+ interactive chart sections, each providing a different lens on your QA data:

Bug Reports Over Time

Stacked bar chart showing daily or weekly report volume, color-coded by severity (critical, high, medium, low). Spot spikes after deploys or regressions at a glance.

Quality Score Trend

Line chart tracking your team's average bug report quality score over time. Measures whether report completeness and clarity are improving or declining.

Severity Distribution

Donut chart breaking down all reports by severity. Quickly see the ratio of critical vs. low-priority issues in your backlog.

Status Distribution

Donut chart showing reports by status (New, Confirmed, In Progress, Resolved, Closed, etc.). Identify bottlenecks where bugs pile up.

Type Distribution

Horizontal bar chart displaying report counts by type — UI, performance, crash, security, logic, data, network, feature request, and more. See where your product's weaknesses concentrate.

Top Bug Reporters

Leaderboard of your most active testers and agents, ranked by report count. The top reporter receives a crown icon. Useful for recognizing contributors and balancing workloads.

Automation Health

Stacked bar chart showing Playwright automation pass/fail rates over time. Monitor whether your automated test suite is catching regressions or needs maintenance.

Time Tracking

Bar chart showing logged QA hours by day, broken down by category (testing, bug-triage, automation, review, meeting). Track team effort and identify days with unusually low or high activity.

Notes Created

Sparkline showing the volume of testing notes created over the selected period. Correlate note-taking activity with testing sprints and releases.

Resolution Time

Tracks the average time from bug creation to resolution or closure. Identify slow-moving issues and measure the impact of process improvements.

Reports by Project

Comparison chart when multiple projects exist. See which projects are generating the most bugs and where testing resources are concentrated.

Daily Activity Summary

Combined view of reports filed, resolved, notes created, and automation runs per day. The single-pane view for daily standup reviews.

Filtering & Controls

Every chart responds to two global filters at the top of the dashboard:

  • Time Range — select 7, 14, 30, or 90 days. All charts, the health score, and the leaderboard update instantly
  • Project Filter — scope the entire dashboard to a single project, or view aggregated data across all projects

API Access

All analytics data is available programmatically:

  • REST APIGET /analytics returns daily counts, severity/status/type breakdowns, and health score. See the API Reference for parameters and response format
  • MCP — the get_stats tool returns the same data for AI agents. Pass a days parameter (up to 90) to control the lookback window
👑
Analytics is available on Pro ($49/mo) and Team ($499/mo) plans. Free plan users see a summary card with basic counts. Upgrade from Settings → Billing to unlock the full dashboard.

Why bugAgent?

Here's why teams and developers choose bugAgent:

"We ship faster because agents handle triage"

With AI auto-classification across 19 types, reports arrive pre-sorted by type and severity. No more morning triage meetings to categorize yesterday's issues.

🎯

"Our agents finally have a home for quality issues"

Before bugAgent, our coding agents would find issues but had no good way to report them. Now they file enriched bugs, feature requests, and technical debt items in seconds.

🔄

"Zero context switching"

Developers using Claude Code or Cursor never leave their IDE. Bugs are filed, queried, and managed right where the code lives.

📈

"We caught regressions before users did"

CI agents running in our pipeline file bug reports on every regression. By the time a human reviews the PR, the bugs are already documented with full context.

🤝

"Jira sync saved us hours per week"

Enriched reports flow directly to Jira with the right labels, priority, and project. Our PM doesn't have to re-enter anything.

🚀

"The free tier let us validate in a day"

We integrated our agent with bugAgent's free tier in an afternoon. By the next morning, we had 40+ classified bug reports from our test suite.

Developer Resources

Everything you need to build with bugAgent:

Frequently Asked Questions

Quick answers to the most common questions about bugAgent.

What is bugAgent?

bugAgent is an AI-powered platform for managing software quality. It handles bug reports, feature requests, enhancements, technical debt, documentation issues, DevOps tasks, UX improvements, and integration requests. Every report is automatically classified across 19 types, enriched with environment data, and optionally synced to Jira. It works through a REST API, the Model Context Protocol (MCP), a browser SDK, email, or the dashboard's AI Assistant.

How does bugAgent work?

You connect bugAgent to your workflow using one of three methods: the MCP server (for AI coding tools like Claude, Cursor, and Windsurf), the REST API (for programmatic access from CI/CD, scripts, or custom apps), or the browser SDK (a single script tag that records user sessions and lets users file reports from a floating toolbar).

When a report is submitted — whether by an AI agent, a developer, or an end user — bugAgent automatically classifies it by type and severity, captures environment data (browser, OS, viewport, console errors, network failures), checks for duplicates, and formats a structured report. If Jira is connected, the report syncs automatically with mapped fields, priority, and labels.

How do I get started?

Sign up at app.bugagent.com to start your 14-day free trial with all Team features — no credit card required. You get full access to automations, performance testing, security scanning, mobile testing, and more. After 14 days, choose a plan or continue on the free tier. From there:

  • For AI agents: Add the MCP server with claude mcp add bugagent or use the REST API with an API key from your dashboard.
  • For browser session replay: Add a single <script> tag to your site with your API key. Users can file reports from the floating toolbar. Note: you may need to allow popups from app.bugagent.com in your browser settings, as screen recording runs in a separate popup window.

Setup takes under a minute. See the Get Started section above for detailed guides.

Can I add my team?

Yes. Every bugAgent account belongs to an organization (team). You can invite team members from Settings → Team in the dashboard. Each member gets their own login and can create, view, and manage reports within the shared organization. Team members can have different roles — owner, admin, manager, contributor, or tester — with appropriate permissions for each. Owners and managers can remove members; all data created by a removed user is reassigned and their profile is preserved as a ghost stub for historical integrity and re-invite capability.

Can I have multiple projects and organizations?

Yes to both. Within each organization you can create unlimited projects to separate reports by product, repository, or team. Each project has its own slug, default settings, and Jira mapping. You can also create or join multiple organizations — useful if you work with different companies or clients. Switch between organizations from the dashboard header. A Default Project is automatically created when you sign up or create a new organization, so you always have at least one. Owners and managers can delete projects, which permanently removes all associated data (bug reports, automations, test cases, mobile apps, schedules, geo snaps, notes, and time entries). You cannot delete your last project.

What does pricing look like?

bugAgent offers four plans:

  • Free ($0) — Unlimited bug reports and test cases, 1 automation script, 1 scheduled task, 3 projects, 1 user, AI classification, Jira sync, AI Assistant, limited API and MCP access, 100 MB media storage, and community support.
  • Pro ($49/mo) — Unlimited reports, unlimited test cases, unlimited automations, unlimited scheduled tasks, unlimited projects, up to 3 users, Session Replay, Slack integration, 3 GB media storage, full API and MCP access, and email support.
  • Team ($499/mo) — Unlimited reports, unlimited projects, unlimited users, Session Replay, all integrations, team dashboard, custom workflows, unlimited storage, full API and MCP access, and Slack support.
  • Enterprise (custom) — Everything in Team plus SSO/SAML, dedicated AI Assistant, advanced Session Replay, custom MCP tools, SLA guarantee, and dedicated support.

Annual billing saves 10% ($45/mo for Pro, $449/mo for Team). All plans include a 14-day free trial with all Team features, SSL encryption, MCP protocol access, AI Assistant, and email notifications.

Add-ons (available for any paid plan): Mobile+ ($299/mo) for unlimited mobile apps and runs, Performance+ ($299/mo) for unlimited performance tests, Security+ ($499/mo) for unlimited security scans. Add-ons start billing immediately, even during your trial. See the Pricing page for full details.

How do I cancel my subscription?

You can cancel anytime from Settings → Billing in the dashboard. Click "Manage Subscription" to open the Stripe billing portal where you can cancel, downgrade, or update your payment method. When you cancel, your plan stays active until the end of the current billing period. Your data is retained — you can reactivate at any time. No cancellation fees.

What types of reports can bugAgent handle?

bugAgent supports 19 report types across two categories:

  • Bug types: UI, Performance, Crash, Security, Logic, Data, Network, Functional, UI/UX, Data Integrity, Accessibility, Compatibility
  • Non-bug types: Feature Request, Enhancement, Technical Debt, Documentation, DevOps, UX Improvement, Integration

Reports are auto-classified based on title and description using AI pattern matching. You can also set the type manually via the API, MCP, or dashboard.

Does bugAgent integrate with Jira?

Yes. Connect your Jira Cloud instance from Settings → Integrations in the dashboard. Once connected, reports can be synced to Jira with one click or automatically via the API. bugAgent maps fields (type, severity, description, environment), sets Jira priority and labels, and uploads any attachments. Synced issues show the Jira key and a direct link on the report detail page.

Does bugAgent integrate with GitHub?

Yes. Connect your GitHub account from Settings → Integrations in the dashboard. Once connected, map your bugAgent projects to GitHub repos. Playwright automation scripts are automatically pushed to tests/bugagent/{name}.spec.ts in the mapped repo. Scripts update on edit and are removed when archived. If a SHA conflict occurs (someone edited the file outside bugAgent), the dashboard shows the conflict and lets you force-push to resolve it. Available on Pro and Team plans.

Is my data secure?

Yes. All data is encrypted in transit (TLS/SSL) and at rest. bugAgent uses Supabase with Row Level Security (RLS), ensuring users can only access data within their own organization. API keys are scoped to specific permissions, and session replay data is team-isolated. We never share or sell your data. See our Privacy Policy for full details.

Support

We're here to help you build with bugAgent. Reach out anytime: