No results for

Powered byAlgolia

Running distributed tests

It has already been established that k6 can run large load tests from a single instance, but what about multiple instances running a single test?

Several reasons why you may wish to run a distributed test include:

  • Your system under test (SUT) should be accessed from multiple IP addresses.
  • A fully optimized node cannot produce the load required by your extremely large test.
  • Kubernetes is already your preferred operations environment.

For scenarios such as these, we've created the k6-operator.

Introducing k6-operator

k6-operator is an implementation of the operator pattern in Kubernetes, defining custom resources in Kubernetes. The intent is to automate tasks that a human operator would normally do; tasks like provisioning new application components, changing configurations, or resolving run-time issues.

The k6-operator defines the custom TestRun resource type and listens for changes to, or creation of, TestRun objects. Each TestRun object references a k6 test script, configures the environment, and specifies the number of instances, as parallelism, for a test run. Once a change is detected, the operator will react by modifying the cluster state, spinning up k6 test jobs as needed.

Get started with k6-operator

Let's walk through the process for getting started with the k6-operator. The only requirement being access to a Kubernetes cluster and having the appropriate access and tooling.

Experiment in Docker

This process can be performed on a local Kubernetes cluster running within Docker! Using kind or k3d are awesome options to experiment with the process.

1. Install the operator

The first step to running distributed tests in Kubernetes is to install the operator if not already installed in the cluster. At this time, installation does require having the project source code downloaded onto your system. Installation commands must be run from the source directory.

Prerequisites

Besides privileged access to a Kubernetes cluster, installation will require that the system performing the installation has the following tools installed:

From your command-line, execute the following:

git clone https://github.com/grafana/k6-operator && cd k6-operator

Ensure that your kubectl tool is set for the appropriate Kubernetes cluster. Then, from the k6-operator directory, you may now perform the installation:

make deploy

By default, the operator will be installed into a new namespace, k6-operator-system. You can verify the successful installation by listing available resources within the namespace:

kubectl get pod -n k6-operator-system

After a few moments, your resulting status should become Running as shown below:

Output
NAME READY STATUS RESTARTS AGE
k6-operator-controller-manager-7664957cf7-llw54 2/2 Running 0 160m

You are now ready to start create and execute test scripts!

2. Create a test script

Creating k6 test scripts for Kubernetes is no different from creating the script for the command-line. If you haven’t already created test cases for your system, then we suggest having a read through one of our guides for creating tests for websites and APIs/microservices:

In general, it is advised to start small and expand on your scripts over iterations. So let's start simple and create a test.js with the following content:

test.js
import http from 'k6/http';
import { sleep } from 'k6';
export const options = {
vus: 10,
duration: '10s',
};
export default function () {
http.get('https://test.k6.io/');
sleep(1);
}

Best practice

While creating scripts, run them locally before publishing to your cluster. This can give you immediate feedback if you have errors in your script.

Let's go ahead and verify our script is valid by performing a brief test:

k6 run test.js

We should see a successful execution and resulting output summary.

3. Add test scripts

In order for your test scripts to be run, they must be published into your cluster. Two main methods for this are using ConfigMap or PersistentVolume resources.

Add as a ConfigMap

Using a ConfigMap is a quick and straightforward mechanism for adding your test scripts to Kubernetes. The kubectl tool provides a convenient method to create a new ConfigMap from a local script.

Let's create our ConfigMap as my-test with the content of our test.js script we created in the previous step:

kubectl create configmap my-test --from-file test.js
attention

Limitations exist on how large your test script can be when deployed within a ConfigMap. Kubernetes imposes a size limit of 1,048,576 bytes (1 MiB) for the data, therefore if your test scripts exceed this limit, you'll need to mount a PersistentVolume.

Check the motivations for when you should use a ConfigMap versus a PersistentVolume.

You should see confirmation with configmap/my-test created.

Add inside a PersistentVolume

Setting up a PersistentVolume is beyond the scope of this guide, but enables access to a shared filesystem from your Kubernetes cluster via PersistentVolumeClaim.

When using this option, organize your test scripts in the applicable filesystem just as you would locally. This mechanism is ideal when breaking up monolithic scripts into reusable modules.

As seen on k6 Office Hours

Organizing your test scripts was part of the discussion during episode #76 of k6 Office Hours.

attention

When using a PersistentVolume, the operator will expect all test scripts to be contained within a directory named /test/.

To learn more about creating PersistentVolume and PersistentVolumeClaim resources, review the Kubernetes documentation.

4. Create a custom resource

During installation, the TestRun Custom Resource definition was added to the Kubernetes API. The data we provide in the custom resource TestRun object should contain all the information necessary for the k6-operator to start a distributed load test.

Specifically, the main elements defined within the TestRun object relate to the name and location of the test script to run, and the amount of parallelism to utilize.

note

The TestRun custom resource provides many configuration options to control the initialization and execution of tests. For the full listing of possible options, please refer to the project source and README.

The following examples will show some common variations for the custom resource:

Script in a ConfigMap

When the test script to be executed is contained within a ConfigMap resource, we specify the script details within the configMap block of YAML. The name is the name of the ConfigMap and the file is the key-value for the entry.

Let's create the file run-k6-from-configmap.yaml with the following content:

run-k6-from-configmap.yaml
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-from-configmap
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js

Recall when the script was added as a ConfigMap for our configuration values. We created the ConfigMap named my-test. The test script content was added to the map using the filename as the key-value, therefore the file value is test.js.

The amount of parallelism is up to you; how many pods do you want to split the test amongst? The operator will split the workload between the pods using execution segments.

attention

It is important that the ConfigMap and CustomResource are created in the same Namespace.

Script in a PersistentVolume

If the test script to be executed is contained within a PersistentVolume, creation of a PersistentVolumeClaim will be required. We won't go into the details of PersistentVolumes and PersistentVolumeClaims, but to learn more, you should review the Kubernetes documentation.

Assume we've created a PersistentVolumeClaim named my-volume-claim against a PersistentVolume containing the test script /test/test.js, we can create the file run-k6-from-volume.yaml with the following content:

run-k6-from-volume.yaml
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-from-volume
spec:
parallelism: 4
script:
volumeClaim:
name: my-volume-claim
# File is relative to /test/ directory within volume
file: test.js
attention

It is important that the PersistentVolumeClaim and CustomResource are created in the same Namespace.

Configure the environment

Not everything should be included directly in your scripts. Well written scripts will allow for variability to support multiple scenarios and to avoid hard-coding values that tend to change. These could be anything from passwords to target urls, in addition to system options.

We can pass this data as environment variables for use with each pod executing your script. This can be defined explicitly within the TestRun resource, or by referencing a ConfigMap or Secret.

run-k6-with-vars.yaml
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-with-vars
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js
runner:
env:
- name: MY_CUSTOM_VARIABLE
value: 'this is my variable value'
envFrom:
- configMapRef:
name: my-config-vars
- secretRef:
name: my-secrets-vars
note

The above YAML introduces the runner section. This section applies to each pod that will be running a portion of your test, based upon the desired parallelism.

Now, with the referenced resources, our test scripts can use environment variables as in the following:

export function setup() {
console.log(`Variable is set as: ${__ENV.MY_CUSTOM_VARIABLE}`);
}

Change command-line arguments

k6 options can be specified in many ways, one being the command-line. Specifying options via command-line can still be accomplished when using the operator as shown with the following example:

run-k6-with-args.yaml
apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-with-args
spec:
parallelism: 4
script:
configMap:
name: my-test
file: test.js
arguments: --tag testid=run-k6-with-args --log-format json

With the above arguments, we're adding a test-wide custom tag to metrics and changing the output format of logs to JSON.

note

Be sure to visit the options reference for a listing of available options.

5. Run your test

Tests are executed by applying the custom resource TestRun to a cluster where the operator is running. The test configuration is applied as in the following:

kubectl apply -f /path/to/your/k6-resource.yaml

After completing a test run, you need to clean up the test jobs created. This is done by running the following command:

kubectl delete -f /path/to/your/k6-resource.yaml

Best Practice

If you make use of real-time results output, e.g. Prometheus Remote Write, use the cleanup option to automatically remove resources upon test completion as with the following example:

apiVersion: k6.io/v1alpha1
kind: TestRun
metadata:
name: run-k6-with-realtime
spec:
parallelism: 4
# Removes all resources upon completion
cleanup: post
script:
configMap:
name: my-test
file: test.js
arguments: -o experimental-prometheus-rw

6. When things go wrong

Sadly nothing works perfectly all the time, so knowing where you can go for help is important.

Be sure to search the k6-operator category in the community forum. k6 has a growing and helpful community of engineers working with k6-operator, so there's a good chance your issue has already been discussed and overcome. It's also in these forums where you'll be able to get help from members of the k6 development team.

See also

Here are some additional resources to help on your learning journey: