Introduce Network Fault Into an Application Running on Kubernetes
This guide will walk you through emulating faults into an application running in a Kubernetes cluster.
Run lueur's proxy as a Deployment
-
Deploy lueur's demo application in the cluster
This steps serves only the purpose of demonstrating lueur's working in a Kubernetes cluster. You can safely ignore it if you have another application you wish to try.
lueur-demo.yaml--- apiVersion: v1 kind: ServiceAccount metadata: name: lueur-demo labels: app: lueur-demo automountServiceAccountToken: false --- apiVersion: v1 kind: Service metadata: name: lueur-demo labels: app: lueur-demo spec: selector: app: lueur-demo ports: - protocol: TCP port: 7070 targetPort: 7070 --- apiVersion: apps/v1 kind: Deployment metadata: name: lueur-demo labels: app: lueur-demo spec: replicas: 1 selector: matchLabels: app: lueur-demo template: metadata: labels: app: lueur-demo annotations: sidecar.istio.io/inject: "false" spec: serviceAccountName: lueur-demo securityContext: runAsUser: 65532 runAsGroup: 65532 fsGroup: 65532 containers: - name: lueur-demo image: rebound/lueur:latest imagePullPolicy: Always tty: true args: - demo - run - "0.0.0.0" - "7070" ports: - containerPort: 7070 name: http securityContext: allowPrivilegeEscalation: false readOnlyRootFilesystem: true privileged: false capabilities: drop: - ALL
Apply it as follows:
-
Deploy lueur's proxy Kubernetes Resources
Below is an example of running lueur's proxy as a deployment, with a single replica.
lueur-proxy.yaml--- apiVersion: v1 kind: ServiceAccount metadata: name: lueur-proxy labels: app: lueur-proxy automountServiceAccountToken: false --- apiVersion: v1 kind: ConfigMap metadata: name: lueur-proxy-config labels: app: lueur-proxy data: LUEUR_UPSTREAMS: "http://lueur-demo:7070" # (1)! LUEUR_WITH_LATENCY: "true" # (2)! LUEUR_LATENCY_MEAN: "300" --- apiVersion: v1 kind: Service metadata: name: lueur-proxy labels: app: lueur-proxy spec: selector: app: lueur-proxy ports: - protocol: TCP port: 8080 targetPort: 8080 --- apiVersion: apps/v1 kind: Deployment metadata: name: lueur-proxy labels: app: lueur-proxy spec: replicas: 1 selector: matchLabels: app: lueur-proxy template: metadata: labels: app: lueur-proxy annotations: sidecar.istio.io/inject: "false" # (3)! spec: serviceAccountName: lueur-proxy securityContext: runAsUser: 65532 runAsGroup: 65532 fsGroup: 65532 containers: - name: lueur-proxy image: rebound/lueur:latest imagePullPolicy: Always tty: true args: - --log-stdout - --log-level - debug - run - --no-ui # (4)! - --proxy-address - "0.0.0.0:8080" # (5)! ports: - containerPort: 8080 name: http envFrom: - configMapRef: name: lueur-proxy-config securityContext: allowPrivilegeEscalation: false readOnlyRootFilesystem: true privileged: false capabilities: drop: - ALL
- Comma-seperated list of hosts that the proxy is allowed to impact. We resolve to the demo application via its Kubernetes service name.
- Enable a latency fault, read the reference for more details on environment variables
- Not really needed but in case you run in a Istio-aware environment, tell Istio not to add any sidecar to the proxy
- Disable the proxy terminal's UI which isn't really useful in this environment
- Make the lueur proxy address listen on a non-loopback interface to be reachable
Apply it as follows:
-
Make a HTTP request to the demo service via the proxy
First, start a throwaway curl pod. This will start a shell from it:
Once the pod is started and its shell available, you can run the following command from it:
curl -w "\nConnected IP: %{remote_ip}\nTotal time: %{time_total}s\n" -x http://lueur-proxy:8080 http://lueur-demo:7070 <h1>Hello, World!</h1> Connected IP: 10.152.183.146 Total time: 0.315056s
This resolves both the proxy and the demo application from within the cluster, demonstrating a latency of roughly
315ms
.Once you exist the pod, its resources will be automatically released.
Run lueur's scenario as a Job
-
Deploy lueur's demo application in the cluster
This steps serves only the purpose of demonstrating lueur's working in a Kubernetes cluster. You can safely ignore it if you have another application you wish to try.
lueur-demo.yaml--- apiVersion: v1 kind: ServiceAccount metadata: name: lueur-demo labels: app: lueur-demo automountServiceAccountToken: false --- apiVersion: v1 kind: Service metadata: name: lueur-demo labels: app: lueur-demo spec: selector: app: lueur-demo ports: - protocol: TCP port: 7070 targetPort: 7070 --- apiVersion: apps/v1 kind: Deployment metadata: name: lueur-demo labels: app: lueur-demo spec: replicas: 1 selector: matchLabels: app: lueur-demo template: metadata: labels: app: lueur-demo annotations: sidecar.istio.io/inject: "false" spec: serviceAccountName: lueur-demo securityContext: runAsUser: 65532 runAsGroup: 65532 fsGroup: 65532 containers: - name: lueur-demo image: rebound/lueur:latest imagePullPolicy: Always tty: true args: - demo - run - "0.0.0.0" - "7070" ports: - containerPort: 7070 name: http securityContext: allowPrivilegeEscalation: false readOnlyRootFilesystem: true privileged: false capabilities: drop: - ALL
Apply it as follows:
-
Load a lueur scenario as a Kubernetes ConfigMap
Let's play a simple scenario with a single test call executed 4 times in total: 12 baseline call without latency applied and three calls with latencies gradually increasing by
30ms
steps.scenario.yaml--- title: "Latency Increase By 30ms Steps From Downstream" description: "" scenarios: - call: method: GET url: http://lueur-demo:7070/ping context: upstreams: - https://postman-echo.com faults: - type: latency mean: 80 stddev: 5 direction: ingress side: client strategy: mode: Repeat step: 30 count: 3 add_baseline_call: true expect: status: 200 response_time_under: 490
To load this scenario as a configmap, run the next command:
-
Deploy lueur's scenario as a Kubernetes Job
Below is an example of running lueur's scenarior as a job without retry.
lueur-scenario.yaml--- apiVersion: v1 kind: ServiceAccount metadata: name: lueur-scenario labels: app: lueur-scenario automountServiceAccountToken: false --- apiVersion: batch/v1 kind: Job metadata: name: lueur-scenario labels: app: lueur-scenario spec: backoffLimit: 0 # (1)! template: metadata: labels: app: lueur-scenario annotations: sidecar.istio.io/inject: "false" spec: serviceAccountName: lueur-scenario restartPolicy: Never securityContext: runAsUser: 65532 runAsGroup: 65532 fsGroup: 65532 containers: - name: lueur-scenario image: rebound/lueur:latest imagePullPolicy: Always tty: true args: - scenario - run - --scenario - rebound/scenario.yaml - --result - result.json # (2)! - --report - report.json # (3)! volumeMounts: - name: lueur-scenario-file mountPath: /home/nonroot/rebound/scenario.yaml # (4)! readOnly: true securityContext: allowPrivilegeEscalation: false readOnlyRootFilesystem: false privileged: false capabilities: drop: - ALL volumes: - name: lueur-scenario-file configMap: name: lueur-scenario-file items: - key: scenario.yaml path: scenario.yaml
- Do not restart the job if it failed
- Results contain the detailed events of the tests and all the applied faults
- A report is a rough analysis of the results made by lueur
- Mount the scenario into the job's container
Apply it as follows: