TestDino
Test Explorer

Find your flakiest specs in seconds, not hours.

Sort specs by flaky rate, find the worst offenders, and drill into individual test health.

Free up to 5,000 executions/month. Setup in under 5 minutes.

TestDino test explorer dashboard
Playwright-native reporting
AI-powered test analysis
Trusted by teams at

Hundreds of tests and no way
to see the big picture

Something is unstable in your suite but you do not know where to start looking.

No overview of which spec files are healthy and which are not

You know some tests are failing or flaky, but you cannot see which spec files are the worst offenders. Without a spec-level health view, you are guessing where to focus your stabilization effort.

Pass rates and flaky rates exist nowhere in your workflow

Playwright gives you pass/fail per test per run. But what is the pass rate of your login spec over the last 30 runs? That data exists in scattered CI logs but nobody has aggregated it.

Slow tests hide in plain sight

Some spec files take 5 minutes to run while others take 10 seconds. But you have never sorted your specs by duration because no tool shows you that view. Slow tests waste CI time every single run.

No way to track suite health over time

Is your test suite getting more stable or less stable? Are flaky rates going up or down? Without trend data, you cannot answer these questions. You cannot improve what you do not measure.

Your suite is unstable.You do not know where to look.

Browse specs and tests

Test Explorer
Test Explorer
Spec health in one place
Search test explorer
Time Period
Tags
Platforms
Environment
Showing 3 of 3 specsLast updated: Just now
Spec / Test CaseExecutionsFailureFlakyAvgLast Run

Click a test to inspect

Test Details
Select a test
Select a test to inspect run history and repeated failures.

How the test explorer works

TestDino aggregates results across every run to build spec-level and test-level health metrics. Sort, filter, and drill down to understand exactly where your suite needs work.

Add the TestDino reporter

One line in your Playwright config. The test explorer starts building health metrics for every spec file and test case from your very first run.

playwright.config.ts
reporter: [
  ['html', { outputDir: './playwright-report' }],
  ['json', { outputFile: './playwright-report/report.json' }],
]
Upload after tests run
npx tdpw upload <report-dir> --token=$TESTDINO_API_KEY

Metrics aggregate across runs

After each run, TestDino recalculates pass rates, flaky rates, average durations, and failure counts for every spec file and individual test case. The more runs you have, the more accurate the picture.

Metrics aggregate across runs

Sort specs to find problems

Open the test explorer and sort spec files by flaky rate to find unstable areas. Sort by duration to find slow specs eating CI time. Click into any spec to see the same metrics at the individual test level.

Sort specs to find problems

Explore test data via MCP

Connect the TestDino MCP server to your editor and pull recent failures by tag, spec file, or status, and inspect individual test case history without opening the dashboard.

Explore test data via MCP

Teams love what we built

See why developers choose TestDino to ship faster and debug smarter

Shrinath Rao

Over 30 flaky tests and no structured way to track them, just CI artifacts and morning guesswork. TestDino's "Most Flaky Tests" feature broke this pattern. We can see failure trends now and pull up video recordings of exactly what went wrong. The TestDino MCP server is the magic piece on top, I ask my Claude agent about a failure and it pulls full context from TestDino without switching tabs. We went from 30-something flaky tests down to 3 or 4.

70%

Fewer flaky test reruns

3x

Faster failure triage

Shrinath Rao

Lead Quality Assurance Engineer

Johan Frølich

Migrating to TestDino from Currents was an easy decision. The features are stronger, the cost is lower, and the interface makes debugging far less painful. Flaky test detection and AI failure classification have simplified debugging and reduced our CI costs by cutting down reruns and noisy failures.

40%

Reduction in CI costs

90%

Less time triaging failures

Johan Frølich

CTO & Co-founder

Spec-level and test-level
health metrics at a glance

Spec-level health metrics

Each spec file shows executions, failure rate, flaky rate, average duration, platform, tags, recent status, and last run timestamp in sortable columns.

Individual test case drill-down

Click any spec file to view health metrics for each test case inside it and spot what lowers your pass rate.

Sortable columns for prioritization

Sort spec files by flaky rate, pass rate, duration, or failure count. The worst offenders surface immediately.

Trend lines over time

Pass rates and flaky rates are tracked over time so you can confirm whether your suite is stabilizing or regressing.

Duration tracking per spec and test

Average execution duration for every spec and test case. Identify tests taking too long and specs with creeping duration.

Filter by time period, environment, and platform

Scope metrics to the last 7, 14, 30, 60, or 90 days. Filter by environment, browser, or tags. Use regex search to find specific spec files or test titles.

What you get with the test explorer

Health metrics from spec-level overview to individual test cases.

Individual test case metrics

Individual test case metrics

Drill into any spec file to see the same health metrics for every test case inside it. The test-level view shows you exactly where the problem is so you fix the right thing instead of refactoring the whole spec.

Hierarchical and flat views with sortable columns

Hierarchical and flat views with sortable columns

Switch between a hierarchical view grouped by spec file and a flat view listing every test. Sort by executions, failure rate, flaky rate, average duration, or last run timestamp to find problems fast.

Test case details with execution history and error aggregation

Test case details with execution history and error aggregation

Click any test to open a details drawer with two tabs. Overview shows execution history and platform/environment breakdowns. Errors shows aggregated error messages with first/last occurrence and frequency.

FAQs

You get basic pass/fail data from your first run. After 5-10 runs, flaky rates and pass rates become statistically meaningful. Most teams see a reliable health picture within their first week.