Skip to content

Load Test Results Overview

Your load test just finished. You've got hundreds of thousands of transaction results (maybe millions), performance metrics across time, error logs, server resource data. What matters? Where do you start?

This guide explains the results interface: what you see, what it means, and where to look first.


What Happens When a Load Test Completes

After a load test completes, Load Tester automatically switches from live monitoring mode to results analysis mode. Three views, three purposes:

1. Load Test Results View (Live Monitoring)

This view runs during a load test. It's the operator's dashboard while the test executes: a grid of live charts updating in real time (the default is nine, configurable to 2 / 4 / 6 / 8 / 12 via the chart layout control), plus current/next step indicators driving the test forward. After the test ends, it persists with the final state visible, but for analysis use the Embedded Analytics Dashboard (below). The Results View's only practical post-test purpose is drilling into individual errors via its Errors tab, and that same error data is also surfaced in the Dashboard.

What you see:

  • Summary section (top): Elapsed time, estimated duration, Report button, Dashboard button
  • Live metric charts (middle): Users, Avg Page Duration, Max Page Duration, Total Pages Failed, Hits/sec, Pages/sec, Hits (cumulative), Bandwidth Out, Cases/min, all updating live during the test. The chart layout is configurable (2 / 4 / 6 / 8 / 9 / 12 charts) and any chart slot can be swapped to display a different metric.
  • Current/Next Step indicators (bottom, during test only): Load level progression
  • Errors tab (below charts): Drill into individual errors and compare recorded vs replay requests/responses

Why this view matters: While the test runs, you watch metrics here and intervene if something goes wrong (halt the test, adjust load levels). When the test completes, click Dashboard to move into the interactive analysis surface.


2. Navigator View (Test Results Tree)

The Navigator shows all your load test results organized chronologically:

Structure:

📁 Test Case Name
  ├── 📊 Load Test Result (2026-02-10 14:35) ← Most recent
  ├── 📊 Load Test Result (2026-02-09 11:22)
  └── 📊 Load Test Result (2026-02-08 16:45)

Why this view matters: Compare multiple test runs. Did the performance improve after your optimization? Did adding 2 web servers double capacity? The Navigator shows your testing history.

What to do: Double-click a result to open its report. Right-click for options (Open Test Report, Delete, Export).


3. Embedded Analytics Dashboard (Interactive Analysis)

Click "Dashboard" in the Results View (or double-click a test in Navigator) to open the interactive analytics dashboard:

Embedded Analytics Dashboard showing interactive metrics, drill-down, and AI insights

What you see:

  • Interactive charts: Real-time metrics with zoom, filter, and drill-down
  • AI-powered insights: Ask questions like "What's causing slow response times?"
  • Health matrix: Color-coded performance health across all metrics
  • Bottleneck detection: Automatic identification of CPU, memory, database, network bottlenecks
  • Transaction drill-down: Click any page to see individual transaction performance
  • Server metrics correlation: CPU, memory, disk, network overlaid with response times
  • Error analysis: Interactive error timeline with filtering and AI explanations

Why this view matters: This is where you diagnose root causes interactively. The Results View shows symptoms (response times spiked), the Dashboard lets you drill down to find causes (database CPU hit 100% at 300 VUs).

What to do: Ask the AI "What's the bottleneck?" or drill down interactively: click charts to zoom, filter by time range, compare metrics.


The Three Views: When to Use Each

View Purpose When to Use
Load Test Results View Live runtime monitoring During test (watch metrics, halt or adjust load), After test (error drill-down via Errors tab)
Navigator Test history and comparison Finding specific test results, Comparing multiple runs
Embedded Analytics Dashboard Interactive analysis with AI Investigating bottlenecks, Drill-down analysis, AI-powered insights
Legacy Reports Static HTML for archival Offline sharing, Archival, Executive PDFs

Mental model: Results View is the speedometer (quick glance). Navigator is the filing cabinet (find your tests). Dashboard is the diagnostic lab (interactive deep-dive). Legacy Reports are the printouts (static snapshots for sharing).


Understanding the Results View Interface

The Load Test Results View has three sections, each serving a specific purpose.

Summary Section (Top)

Fields you'll see:

Field What It Shows Why It Matters
Elapsed Time Total test duration Verify test ran for expected length
Estimated Duration Configured test duration Catch early termination (e.g., test stopped at 30 min instead of 60 min)
Dashboard button Opens Embedded Analytics Dashboard Gateway to interactive deep analysis
Report ▼ dropdown Opens legacy static reports For archival/offline sharing

During an active test, this section also shows:

  • Current Step: Which load level is active (e.g., "Ramping to 200 VUs")
  • Next Step: What happens next (e.g., "Hold at 200 VUs")
  • Next Step Begins In: Countdown timer (e.g., "4:32 remaining")
  • Edit button: Modify running test plan (add/remove load levels)

Know what's happening now and what's coming next. If the test is holding at 100 VUs for 10 more minutes, you can grab coffee. If it's about to ramp to 500 VUs, stay at your desk.


Metric Charts Section (Middle)

The default live metric grid answers the critical operator questions while the test is running:

Default chart Question it answers
Users What's the current load level? (Does it match the configured profile?)
Avg Page Duration Is the application fast on average?
Max Page Duration What's the worst case any user is experiencing right now?
Total Pages Failed Are users seeing errors? (Should stay near zero)
Hits/sec Is the server handling more requests as users ramp up? (If hits plateau before users do, the server has saturated)
Pages/sec Same question as Hits/sec, but at page granularity
Hits Cumulative hits since the test started (useful as a sanity check on throughput)
Bandwidth Out Is the network saturated? (Should not plateau before user count does)
Cases/min How many full test cases are completing per minute? (Indirect indicator of end-to-end throughput)

Server-side CPU, memory, and disk metrics live in the Servers View, not in the Results View. The two views are designed to be watched side by side.

Chart controls (toolbar buttons):

Button Function Use Case
Display Properties Change number of charts shown (2 / 4 / 6 / 8 / 9 / 12) Show more or fewer metrics
Chart Properties Change which metric a chart displays Swap "Bandwidth Out" for an OS-level metric, for example
Chart Maximize Expand single chart to full view Focus on one metric

Your bottleneck determines what you watch. If database CPU is the problem, swap out "Bandwidth" (irrelevant) for "Database CPU" (critical). If you suspect memory leaks, add a Memory % chart.


Chart Interpretation: What "Good" Looks Like

Understanding chart patterns reveals performance behavior:

Response Time Chart - The Truth

Good (performance scales):

Response Time (ms)
200 |███████████████████████████████
    |
  0 +-------------------------------
    0  100  200  300  400  500 (VUs)

Response times stay flat as VUs increase. Server has capacity to spare.


Acceptable (linear degradation):

Response Time (ms)
600 |                      ██████████
400 |              ████████
200 |      ████████
  0 |██████
    +----------------------------------
    0  100  200  300  400  500 (VUs)

Response times increase gradually. Normal resource contention, still manageable (under 2000ms).


Problem (capacity exceeded):

Response Time (ms)
5000|                    ████
2000|                ████
 500|        ████████
   0|████████
    +----------------------------
    0  100  200  300  400 (VUs)

Response times spike suddenly at around 350 VUs. Hard limit reached: connection pool exhausted, CPU maxed, something broke.


Hits/Sec Chart - Server Throughput

Good (scaling linearly):

Hits/Sec
2000|                      ██████████
1500|              ████████
1000|      ████████
 500|██████
    +----------------------------------
    0  100  200  300  400  500 (VUs)

Hits/sec increases with VUs. Server can process more requests as load increases.


Problem (throughput plateau):

Hits/Sec
2500|        ████████████████████████
2000|    ████
1000|████
    +----------------------------------
    0  100  200  300  400  500 (VUs)

Hits/sec plateaus at around 2500. Server is maxed out and can't process more requests even though VUs keep increasing.


Errors/Sec Chart - Application Health

Good (zero errors):

Errors/Sec
100|
 50|
  0|════════════════════════════════
    +----------------------------------
    0  100  200  300  400  500 (VUs)

Zero errors. Application working correctly under load.


Problem (errors appear under load):

Errors/Sec
100|                    ████████████
 50|                ████
  0|════════════════
    +----------------------------------
    0  100  200  300  400  500 (VUs)

Errors start at around 300 VUs. Application is breaking under load (503 Service Unavailable, timeouts, etc.).


Ask the AI to Interpret Your Results

If you're not sure what your results mean:

I just finished a load test ramping from 100 to 500 VUs over 60 minutes.
Response times stayed at 200ms until 350 VUs, then jumped to 4000ms.
Errors started appearing at 350 VUs (503 errors). Server CPU was at 45%.
What's the bottleneck?

The AI can:

  • Analyze chart patterns to identify capacity limits, degradation points, and failure modes
  • Correlate metrics to diagnose root causes (e.g., low CPU + 503 errors = connection pool exhaustion)
  • Explain what each pattern means for your capacity planning
  • Recommend next steps (increase connection pools, optimize queries, add servers)
  • Prioritize which issues to fix first (critical bugs vs. optimization opportunities)

Opening the Embedded Analytics Dashboard

The Dashboard is where interactive analysis happens. Three ways to open it:

Method 1: From Results View (During or After Test)

  1. Look at Load Test Results View (auto-opens during test)
  2. Click "Dashboard" button (top-right)

Use when: Test just finished, you're already looking at Results View.


Method 2: From Navigator (Anytime)

  1. Open Navigator view (Window → Show View → Navigator)
  2. Expand test case to see all load test results
  3. Double-click a load test result

Use when: Analyzing older tests, comparing multiple test runs.


Method 3: From Menu

  1. Window → Show View → Embedded Analytics Dashboard

Use when: You want the dashboard always visible (can dock it as a persistent view).


Legacy Reports Still Available

For static HTML reports (archival, offline sharing), click the Report ▼ dropdown and select "Legacy Report". See Legacy Reports for details.


What to Look For First

When results open, follow this triage workflow:

1. Glance at Response Time Chart

Question: Did response times spike?

  • Flat or gradual increase → Performance good, likely no bottleneck
  • ⚠️ Sudden spike → Capacity limit reached, investigate further

If flat, you're probably done. If spike, note the VU count when it happened. That's your capacity limit.


2. Check Errors/Sec Chart

Question: Are there errors?

  • Zero errors → Application working correctly
  • ⚠️ Errors > 0 → Application broken under load, investigate error types

If errors exist, open the Report and check the Errors section for error messages. 401 Unauthorized? 503 Service Unavailable? Error types reveal root causes.


3. Correlate Server Metrics

Question: Which server resource maxed out?

  • CPU at 95%+ → CPU bottleneck (optimize code or add CPU capacity)
  • Memory at 95%+ → Memory bottleneck or leak (optimize memory usage)
  • Database CPU at 95%+ → Database bottleneck (optimize queries, add DB capacity)
  • Bandwidth plateaued → Network bottleneck (upgrade NIC, add CDN)

Server metrics explain WHY response times spiked. Fix the resource that maxed out first.


4. Determine Capacity Limit

Question: How many users can the system handle?

Look for the VU count where degradation accelerated:

  • Response times doubled at 200 VUs → Soft limit (degrading but usable)
  • Response times spiked 10x at 350 VUs → Hard limit (broken)
  • Errors appeared at 350 VUs → Failure threshold

If capacity meets requirements (say your target was 500 concurrent users and you got 600), you're done. If not, identify the bottleneck and fix it.


Next Steps: Deep Analysis

After the overview, dive into specific analysis tasks:

Or: If you found a bottleneck, jump to server-side optimization:


Related Topics: