Load testing is one of the most common ways to test the resiliency of your applications. In this blog we show how recording production data with Speedscale and exporting to a K6 load tests gives you the best of both worlds. Whether or not it's important for your organization, there are clear benefits to be had from implementing these types of tests. By doing so, you can: When it comes to load testing, two of the most modern tools are Speedscale and K6. While there are many reasons for choosing one over the other, there are also benefits to using them together. If you want to know what the main differences are, check out the in-depth comparison.
Speedscale creates load tests from recorded traffic so generating load is pretty core to what we do. As a brief overview, we record traffic from your service in one environment and replay it in another, optionally increasing load several fold. During a replay the Speedscale load generator makes requests against the system under test (SUT), with the responses from external dependencies like APIs or a payment processor optionally mocked out for consistency. Your service is the SUT here. Currently the load generator runs as a single process, usually inside a pod in Kubernetes. So how fast is this thing, and how did we get to where we are today?
In this Postman load testing tutorial, you'll learn how to run a large scale load test in Kubernetes using your existing Postman collections. Because HTTP services don't have a graphical user interface, it's common to build collections of requests using Postman during the development process. These collections are useful for running quick functionality tests as you develop each endpoint. However, as the service grows you eventually need to test it in a more realistic way with larger volume. This is called a load or stress test. Speedscale is a Production Data Simulation Platform that includes this stress/load testing capability out of the box.