Performance Testing Metrics: A Comprehensive Guide
Performance testing involves monitoring various metrics to assess system behavior under different conditions, including response time, throughput, CPU usage, memory utilization, and error rates.
0 views collected around this technical thread.
Performance testing involves monitoring various metrics to assess system behavior under different conditions, including response time, throughput, CPU usage, memory utilization, and error rates.
This article explains the meanings, calculations, and differences of key high‑performance metrics such as system throughput, QPS, TPS, response time, and concurrency, illustrating each concept with real‑world examples and visual diagrams to help engineers evaluate and optimize large‑scale systems.
This article explains key performance testing parameters such as concurrent users, TPS, response time, virtual users, and data volume, outlines essential monitoring metrics, details preparation steps and simple API testing procedures, and compares popular load‑testing tools like JMeter, Locust, and LoadRunner.
This article provides a comprehensive guide to RunnerGo's performance testing module, detailing plan management, various load test modes such as concurrency, step, error‑rate, response‑time, and requests‑per‑second, along with configuration options, execution workflow, and links to the open‑source repository.
This article explains the key performance metrics of backend systems—including QPS, TPS, response time, concurrency, and throughput—provides their definitions, relationships, practical calculation examples, and guidance on optimal thread counts for achieving balanced performance.
This article presents a comprehensive performance testing case study of a multi‑table database query interface, detailing common performance metrics, response‑time classifications, step‑by‑step load‑test results, root‑cause analysis of CPU, memory, and connection‑pool bottlenecks, and demonstrates how tuning Druid’s maxActive and maxWait parameters optimizes throughput and reduces latency.
This article explains how to filter early response times and compute detailed latency metrics—including percentiles, average, and median—within the FunTester load‑testing framework by redesigning data collection structures and adding a unified statistics method.
The article outlines seven essential performance metrics—response time and throughput, average load, error rate, GC pause time, business indicators, uptime, and log size—explaining their significance for application health and recommending popular monitoring tools for each metric.
The article defines MySQL response time as total request‑to‑reply latency, illustrates it with a travel analogy, and then details two monitoring solutions—Percona’s tcprstat utility that captures TCP‑level timings and the Query Response Time audit plugin that records latency distribution in INFORMATION_SCHEMA tables for analysis and Grafana visualization.
This article explains the concept of response time (RT) for MySQL queries, illustrates how network delays and server load affect RT, and provides two practical monitoring methods—using Percona's tcprstat tool and the MySQL Query Response Time plugin—along with installation commands, sample outputs, and visualization tips.
High concurrency refers to a short period where a system receives a massive number of requests, such as ticket‑booking spikes, and is measured by metrics like response time, throughput, QPS, and concurrent users, while multithreading is merely one programming technique to handle such load, not synonymous with high concurrency.
This article explains the concept of response time in mobile applications and provides a detailed, image‑rich workflow—including video recording with ReplayKit, simulated interactions, video parsing, log extraction, timestamp alignment, and stability detection—to accurately measure and reduce response latency.
This article provides a comprehensive overview of performance testing, explaining its definition, key metrics such as throughput, response time and concurrent users, the relationship between these indicators, optimal testing timing, a step‑by‑step workflow from requirement analysis to reporting, and practical tips for effective load testing and system optimization.
This article explains fundamental performance concepts—defining response time and throughput, using axiomatic methods, analyzing bottlenecks with sequence diagrams and profiling, applying Amdahl’s Law, and guiding capacity planning to build reliable, high‑performance applications.
The article explains why accurate response‑time testing is essential for mobile apps, compares four common measurement methods, introduces a frame‑analysis tool that extracts video frames, deduplicates them, and lets users select start and end frames to compute precise cold‑ and hot‑start times, and discusses its advantages and limitations.
To achieve low‑cost, accurate mobile app response‑time measurement, this article evaluates existing methods, identifies their drawbacks, and proposes an automated solution that combines external camera screen capture with a perceptual‑hash image‑comparison algorithm, detailing implementation steps, hardware setup, and trade‑offs.
This article explains key website performance metrics such as response time, concurrency, and throughput, presents typical values for various operations, and outlines practical optimization strategies for front‑end, application‑server, and storage layers, including caching, CDN, reverse proxy, clustering, and code improvements.
This guide explains how to configure a Windows-based testing environment using Cygwin, install necessary tools, create shortcuts, access Windows drives, and execute shell scripts to record videos, extract frames, compare images, and calculate average response times for Android applications.
This article provides a conceptual summary of SQL Server performance tuning, explaining the goals of optimizing response time and throughput, the importance of baselines and costs, factors influencing performance, and a step‑by‑step DETECT methodology for systematic optimization.