Migrate from Linux install
StackState Self-hosted v5.0.x
This document describes how to migrate data from the Linux install of StackState to the Kubernetes install.
To migrate from the Linux install to the Kubernetes install of StackState, the following high level steps need to be performed:
Incoming data from agents (Kafka) and node synchronisation data (Zookeeper) will not be copied.
After the migration:
- 1.Run both instances of StackState side by side for a number of days to ensure that the new instance runs correctly.
- 2.Stop the Linux install for StackState.
- 3.Remove the Linux install for StackState.
Before you start the migration procedure, make sure you have the following information and tools available:
- Access to:
- The Linux machines running your old StackState installation.
- The Kubernetes cluster running your new StackState installation.
values.yamlfile used to install your StackState installation on Kubernetes.
To export the StackGraph data, execute the regular StackState Linux backup procedures as described below.
- 1.Ensure that the StackGraph node is up and running.
- 2.Login to the StackState node as user
- 3.Stop the StackState service:systemctl stop stackstate.service
- 4.Create a directory to store the exported data:sudo -u stackstate mkdir -p /opt/stackstate/migration
- 5.Export the StackGraph data by creating a backup:sudo -u stackstate /opt/stackstate/bin/sts-standalone.sh export \--file /opt/stackstate/migration/sts-export.graph --graph default
- 6.Copy the file
/opt/stackstate/migration/sts-export.graphto a safe location.
- 7.Start the StackState service:systemctl start stackstate.service
- 1.Enable the MinIO component by adding the following YAML fragment to the
values.yamlfile that is used to install StackStatebackup:enabled: truestackGraph:scheduled:enabled: falseelasticsearch:restore:enabled: falsescheduled:enabled: falseminio:accessKey: MINIO_ACCESS_KEYsecretKey: MINIO_SECRET_KEYpersistence:enabled: trueInclude the credentials to access the MinIO instance:
MINIO_ACCESS_KEYwith 5 to 20 alphanumerical characters.
MINIO_SECRET_KEYwith 8 to 40 alphanumerical characters.
The Helm values `backup.stackGraph.scheduled.enabled`, `backup.elasticsearch.restore.enabled` and `backup.elasticsearch.scheduled.enabled` have been set to `false` to prevent scheduled backups from overwriting the backups that we will upload to MinIO.
- 1.Run the appropriate
helm upgradecommand for your installation to enable MinIO.
- 2.Start a port-forward to the MinIO service in your StackState instance:kubectl port-forward service/stackstate-minio 9000:9000
- 3.In a new terminal window, configure the MinIO client to connect to that MinIO service:mc alias set minio-backup http://localhost:9000 ke9Dm7eFhk9kP53rXlUI mNOWCpoYrhwati7QcOrEwnI7Mtcf0jxg2JzNOMk6
- 4.Verify that access has been configured correctly:mc ls minio-backupThe output should be empty, as we have not created any buckets yet.If the output is not empty, the automatic backup functionality has been enabled. Disable the automatic backup functionality and configure MinIO as described above (i.e. not as a gateway to AWS S3 or Azure Blob Storage and without any local storage).
- 5.Create the bucket that is used to store StackGraph buckets:mc mb minio-backup/sts-stackgraph-backupThe output should look like this:Bucket created successfully `minio-backup/sts-stackgraph-backup`.
- 6.Upload the backup file created in the previous step when StackGraph data was exported from the Linux install:mc cp sts-export.graph minio-backup/sts-stackgraph-backup/The output should look like this:sts-export.graph: 15.22 KiB / 15.22 KiB ▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓▓ 42.61 KiB/s 0s
- 7.Verify that the backup file was uploaded to the correct location:./restore/list-stackgraph-backups.shThe output should look like this:job.batch/stackgraph-list-backups-20210222t122522 createdWaiting for job to start...Waiting for job to start...=== Listing StackGraph backups in bucket "sts-stackgraph-backup"...sts-export.graph===job.batch "stackgraph-list-backups-20210222t122522" deletedMost importantly, the backup file uploaded in the previous step should be listed here.
- 8.Restore the backup:./restore/restore-stackgraph-backup.sh sts-export.graphThe output should look like this:job.batch/stackgraph-restore-20210222t171035 createdWaiting for job to start...Waiting for job to start...=== Downloading StackGraph backup "sts-export.graph" from bucket "sts-stackgraph-backup"...download: s3://sts-stackgraph-backup/sts-export.graph to ../../tmp/sts-export.graph=== Importing StackGraph data from "sts-export.graph"...WARNING: An illegal reflective access operation has occurredWARNING: Illegal reflective access by org.codehaus.groovy.vmplugin.v7.Java7$1 (file:/opt/docker/lib/org.codehaus.groovy.groovy-2.5.4.jar) to constructor java.lang.invoke.MethodHandles$Lookup(java.lang.Class,int)WARNING: Please consider reporting this to the maintainers of org.codehaus.groovy.vmplugin.v7.Java7$1WARNING: Use --illegal-access=warn to enable warnings of further illegal reflective access operationsWARNING: All illegal access operations will be denied in a future release===job.batch "stackgraph-restore-20210222t171035" deleted
- 9.Remove the YAML snippet added in step 1 and run the appropriate
helm upgradecommand for your installation to disable MinIO.
- To access the Elasticsearch instance that runs as part of the Kubernetes installation for StackState, execute the following command:kubectl port-forward service/stackstate-elasticsearch-master 9200:9200and access it on
- To modify the
elasticsearch.ymlconfiguration file, use the Helm chart value
stackstate.elasticsearch.esConfig.For example:stackstate:elasticsearch:esConfig:elasticsearch.yml: |reindex.remote.whitelist: oldhost:9200