This repository contains configurations used for scale tests. Deploying things is a semi-manual for now, but the configurations should make it easy to build a given environment quickly.
How the scale tests are organized:
opstrace
cluster to test against, seeopstrace-*.yaml
for configs.- EKS
driver
cluster to send test data, with separate nodepools forprometheus
instances andavalanche
instances, seedriver-*.yaml
for configs.- The
prometheus
instances EACH scrape ALL of theavalanche
instances, but each prometheus appends its pod name to the labels to ensure that each is treated as a distinct set of data. So the resulting metrics are Nprometheus * Navalanche. - The
prometheus
instances in thedriver
cluster are configured toremote_write
to theopstrace
cluster.
- The
sudo pacman -S eksctl
- Download latest
opstrace
CLI. This ensures we use the latest passing build of Opstrace.
curl -L https://opstrace-ci-main-artifacts.s3-us-west-2.amazonaws.com/cli/main/latest/opstrace-cli-linux-amd64-latest.tar.bz2 | tar xjf -
- Create cluster in
us-west-2
./opstrace create aws $USER-$NUMNODES \
--log-level debug \
-c ./opstrace-cluster-PICKONE.yaml \
--write-kubeconfig-file ~/.kube/config-opstrace
- Apply cortex config overrides (increase series limits)
kubectl apply -f opstrace-kube-overrides.yaml
# restart controller pod if it's already running, some overrides are only checked once at startup
kubectl delete pod -n kube-system opstrace-controller-HASH
- Add people to opstrace cluster via UI
- Create driver cluster in same region as opstrace cluster
eksctl create cluster -f ./driver-eksctl-PICKONE.yaml
If there are problems, take a look at the CloudFormation dashboard first - eksctl is a thin veneer over some CF templates
In the test we have N looker instances writing directly to Opstrace.
The looker
pods are deployed as a StatefulSet
only to have stable pod names, so that metric labels do not change if the looker pods are restarted, reconfigured, or evicted. For example a mass eviction can lead to an unexpected increase in distinct series labels, because the looker pod name is included in the series.
- Deploy looker instances into driver cluster.
CLUSTER_NAME=ship
TENANT=staging
sed "s/__AUTH_TOKEN__/$(cat tenant-api-token-$TENANT)/g" ./driver-looker.template.yaml | \
sed "s/__CLUSTER_NAME__/$CLUSTER_NAME/g" | \
sed "s/__TENANT__/$TENANT/g" | \
kubectl apply -f -
In the test we have N avalanche instances being polled by M prometheus instances.
The avalanche
pods are deployed as a StatefulSet
only to have stable pod names, so that metric labels do not change if the avalanche pods are restarted, reconfigured, or evicted. For example a mass eviction can lead to an unexpected increase in distinct series labels, because the avalanche pod name is included in the series.
The prometheus
pods are also deployed as a StatefulSet, both in order to have stable pod names, and because they need local storage.
- Deploy avalanche instances into driver cluster. Could someday try cortex-tools' benchtool since it has richer metric type support, but it only has basic auth.
kubectl apply -f ./driver-kube-avalanche.yaml
- Deploy per-tenant prometheus instances into driver cluster (will send data to
$TENANT
in opstrace cluster):
CLUSTER_NAME=nick-nine-nodes
TENANT=default
sed "s/__AUTH_TOKEN__/$(cat tenant-api-token-$TENANT)/g" ./driver-kube-prom.template.yaml | \
sed "s/__CLUSTER_NAME__/$CLUSTER_NAME/g" | \
sed "s/__TENANT__/$TENANT/g" | \
kubectl apply -f -
- (optional) Check prometheus dashboard
kubectl port-forward deployments/tenant-prometheus-$TENANT ui
EACH prometheus instance will be scraping ALL avalanche instances: Nprometheus * Navalanche
Go to https://system.yourcluster.opstrace.io/grafana
and check the Opstrace Overview Dashboard
.
- Infra section: Want to stay under limits of around 60% CPU and RAM
- Cluster section: Pay attention to Active Series
Get kubectl config:
aws eks update-kubeconfig --region CLUSTER-REGION --name CLUSTER-NAME --kubeconfig ~/.kube/outpath
After adding a nodegroup to the file, deploy it to an existing cluster:
eksctl create nodegroup -f driver-eksctl.yaml
Change nodegroup size:
COUNT=10
eksctl scale nodegroup -r us-west-1 --cluster nick-scaletest-driver -n t3medium --nodes $COUNT --nodes-min $COUNT --nodes-max $COUNT
To delete an eksctl nodegroup: Delete the corresponding CloudFormation stack in the AWS dashboard. The nodes will automatically be decommissioned and any resident pods will be moved off of them
To delete an eksctl cluster: Delete all the CloudFormation stacks for the nodegroups, THEN delete the stack for the cluster. Attempts to delete the cluster before the stacks SHOULD fail.