Performance Testing Terminology
Performance Testing Terminology
In order to conduct and analyze a good load test, you need to understand these basic terms.
Join the DZone community and get the full member experience.Join For Free
Built by operators for operators, the Sensu monitoring event pipeline empowers businesses to automate their monitoring workflows and gain deep visibility into their multi-cloud environments. Get started for free today.
Performance testing is the process of determining the speed, responsiveness, and stability of a computer, network, software program, or a device under a workload. In order to conduct and analyze a good load test, you need to understand the performance testing terminology including connect time, latency, etc.
Let’s elaborate on what they mean.
Connect time is the time taken to establish a TCP connection between the server and the client. TCP/IP model guarantees delivery of data by its nature by using the TCP Handshake. If TCP Handshake is successful, then the client can send further requests. That data send/receive task is not on the HTTP layer. If there’s no TCP connection made between the server and the client, the client can’t talk to the server. This can happen if the server is not live or busy responding to other requests.
Latency is measured by the time taken for information to get to its target and back again. It’s a round trip. Sometimes, latency means delay. which is a very problematic issue when working with remote data centers. Data hops through nodes till it’s sent from the server so the bigger the distance the more the delay. That’s why those nodes will increase the response time and violate your service level agreements (SLA’s). That’s why dealing with latency is the hardest one. JMeter measures the latency from the first moment of sending the request until the first byte is received. So, in JMeter, Connect time is included when calculating Latency Time. There’s also the network latency to express the time for a packet of data to get from one designated point to another.
Elapsed time is measured by the time from the first moment of sending the data and the time of the last byte of the received response.
So, our basic math tells that:
Latency time – Connect time = Server Processing Time
Elapsed time – Latency time = Download Time
Throughput is the number of units of work that can be handled per unit of time. From a performance testing perspective, it can count requests per second, request calls per day, hits per second, bytes per second. JMeter allows you to create assertions for response time, so you can set your request’s fail\success status according to its result.
Performance thresholds are the KPIs or the maximum acceptable values of a metric. There are many metrics to be identified for a performance test project. That can be response time, throughput, and resource-utilization levels like processor capacity, memory, disk I/O, and network I/O.
Login service shouldn’t execute more than three seconds.
Server should handle 40 payment requests per second.
CPU shouldn’t be higher than 60 percent.
Saturation means the point where a resource is at the maximum utilization point. At this point, if we are talking about a server, it cannot respond to any more requests.
I hope this information will clear your mind about performance testing terminology and it would help you analyze performance test reports in a more efficient way.
Happy Load Testing!
Published at DZone with permission of Canberk Akduygu . See the original article here.
Opinions expressed by DZone contributors are their own.