automation-suite
2022.4
false
- Overview
- Requirements
- Installation
- Post-installation
- Cluster administration
- Managing products
- Managing the cluster in ArgoCD
- Setting up the external NFS server
- Automated: Enabling the Backup on the Cluster
- Automated: Disabling the Backup on the Cluster
- Automated, Online: Restoring the Cluster
- Automated, Offline: Restoring the Cluster
- Manual: Enabling the Backup on the Cluster
- Manual: Disabling the Backup on the Cluster
- Manual, Online: Restoring the Cluster
- Manual, Offline: Restoring the Cluster
- Additional configuration
- Migrating objectstore from persistent volume to raw disks
- Monitoring and alerting
- Migration and upgrade
- Migration options
- Step 1: Moving the Identity organization data from standalone to Automation Suite
- Step 2: Restoring the standalone product database
- Step 3: Backing up the platform database in Automation Suite
- Step 4: Merging organizations in Automation Suite
- Step 5: Updating the migrated product connection strings
- Step 6: Migrating standalone Insights
- Step 7: Deleting the default tenant
- B) Single tenant migration
- Product-specific configuration
- Remapping the organization IDs
- Migrating Looker data to Automation Suite
- Performing Insights database maintenance
- Configuring Log Rules
- Best practices and maintenance
- Troubleshooting
- How to Troubleshoot Services During Installation
- How to Uninstall the Cluster
- How to clean up offline artifacts to improve disk space
- How to clear Redis data
- How to enable Istio logging
- How to manually clean up logs
- How to clean up old logs stored in the sf-logs bucket
- How to disable streaming logs for AI Center
- How to debug failed Automation Suite installations
- How to delete images from the old installer after upgrade
- How to automatically clean up Longhorn snapshots
- How to disable TX checksum offloading
- How to address weak ciphers in TLS 1.2
- Unable to run an offline installation on RHEL 8.4 OS
- Error in Downloading the Bundle
- Offline installation fails because of missing binary
- Certificate issue in offline installation
- First installation fails during Longhorn setup
- SQL connection string validation error
- Prerequisite check for selinux iscsid module fails
- Azure disk not marked as SSD
- Failure After Certificate Update
- Automation Suite not working after OS upgrade
- Automation Suite Requires Backlog_wait_time to Be Set 1
- Volume unable to mount due to not being ready for workloads
- RKE2 fails during installation and upgrade
- Failure to upload or download data in objectstore
- PVC resize does not heal Ceph
- Failure to Resize Objectstore PVC
- Rook Ceph or Looker pod stuck in Init state
- StatefulSet volume attachment error
- Failure to create persistent volumes
- Storage reclamation patch
- Backup failed due to TooManySnapshots error
- All Longhorn replicas are faulted
- Setting a timeout interval for the management portals
- Update the underlying directory connections
- Cannot Log in After Migration
- Kinit: Cannot Find KDC for Realm <AD Domain> While Getting Initial Credentials
- Kinit: Keytab Contains No Suitable Keys for *** While Getting Initial Credentials
- GSSAPI Operation Failed With Error: An Invalid Status Code Was Supplied (Client's Credentials Have Been Revoked).
- Alarm Received for Failed Kerberos-tgt-update Job
- SSPI Provider: Server Not Found in Kerberos Database
- Login Failed for User <ADDOMAIN><aduser>. Reason: The Account Is Disabled.
- ArgoCD login failed
- Failure to get the sandbox image
- Pods not showing in ArgoCD UI
- Redis Probe Failure
- RKE2 Server Fails to Start
- Secret Not Found in UiPath Namespace
- After the Initial Install, ArgoCD App Went Into Progressing State
- MongoDB pods in CrashLoopBackOff or pending PVC provisioning after deletion
- Unexpected Inconsistency; Run Fsck Manually
- Degraded MongoDB or Business Applications After Cluster Restore
- Missing Self-heal-operator and Sf-k8-utils Repo
- Unhealthy Services After Cluster Restore or Rollback
- RabbitMQ pod stuck in CrashLoopBackOff
- Prometheus in CrashloopBackoff state with out-of-memory (OOM) error
- Missing Ceph-rook metrics from monitoring dashboards
- Pods cannot communicate with FQDN in a proxy environment
- Using the Automation Suite Diagnostics Tool
- Using the Automation Suite support bundle
- Exploring Logs

Automation Suite installation guide
Last updated Mar 17, 2025
Use this procedure to migrate Looker data from Insights Standalone to Insights Automation Suite.
- Open an SSH client and run the following command
to authenticate.
ssh <username>@<hostname>
ssh <username>@<hostname>
- Create a
backup.
image=$(sudo docker container inspect looker-container -f '{{ .Image }}') backupPath="$HOME/insights/backup" && mkdir "$backupPath" -p sudo docker run -u root --rm --platform linux --volumes-from looker-container -v "$backupPath":/backup "$image" bash -c "tar cvf /backup/looker_backup.tar --absolute-names /app/workdir/.db /mnt/lookerfiles /app/workdir/looker.key"
image=$(sudo docker container inspect looker-container -f '{{ .Image }}') backupPath="$HOME/insights/backup" && mkdir "$backupPath" -p sudo docker run -u root --rm --platform linux --volumes-from looker-container -v "$backupPath":/backup "$image" bash -c "tar cvf /backup/looker_backup.tar --absolute-names /app/workdir/.db /mnt/lookerfiles /app/workdir/looker.key"Note: Starting with the 2022.10 version, thelooker.key
is included in the migration backup.
- Exit the SSH.
exit
exit - Copy the files to the local machine using Secure
Copy Protocol (SCP).
scp <username>@<hostname>:~\insights\backup\looker_backup.tar <path-to-download-dir>
scp <username>@<hostname>:~\insights\backup\looker_backup.tar <path-to-download-dir>
- Download the
kubeconfig
file. The following screenshot shows an Azure deployment template for Automation Suite. - Open an SSH client and run the following command to get the Insights Looker pod name
$Env:KUBECONFIG="C:\Users\username\Downloads\output.yaml" # path to the optput.yaml kubectl get pods -n uipath
$Env:KUBECONFIG="C:\Users\username\Downloads\output.yaml" # path to the optput.yaml kubectl get pods -n uipath -
The Insights Looker pod is now listed (e.g.,
insights-insightslooker-c987df55c-gngqd
). - Set a variable name for the Insights Looker pod.
$PodName = "insights-insightslooker-74db798bc5-dt68p"
$PodName = "insights-insightslooker-74db798bc5-dt68p" - Create a backup for the Automation Suite data. The backup file will be stored on a machine with console where command is initiated.
Please make sure to delete created backup because it contains sensitive information.
kubectl cp ${PodName}:/app/workdir/.db/ .\sf_db_backup -n uipath -c insightslooker kubectl cp ${PodName}:/mnt/lookerfiles/ .\sf_lookerfiles_backup -n uipath -c insightslooker
kubectl cp ${PodName}:/app/workdir/.db/ .\sf_db_backup -n uipath -c insightslooker kubectl cp ${PodName}:/mnt/lookerfiles/ .\sf_lookerfiles_backup -n uipath -c insightslooker - Unzip the TAR file from the Insights Linux Server.
7z x .\looker_backup.tar -olooker_backup # looker_backup is the output folder
7z x .\looker_backup.tar -olooker_backup # looker_backup is the output folder - (Conditionally required if the Automation Suite password is different from the one used in the Standalone deployment model)
Edit the following files before the migration and update url and/or password in both looker.log and looker.script files.
Open
sf_db_backup/looker.log
and search forhost_url
to find the string that starts withhttps://
.Search forINSERT INTO "CREDENTIALS_EMAIL" VALUES(1,1,'[email protected]'
insf_db_backup/looker.log
.Copy the string that starts with$2a$12$
and overwrite all occurrences inlooker_backup/app/.db/looker.log
.Note:- Turn off regex in the editor if you cannot find these lines.
- If there are multiple occurrences of
INSERT INTO "CREDENTIALS_EMAIL" VALUES(1,1,'[email protected]'
you need to update all of them.
- Change the working directory to the
looker_backup
folder. In this folder you will see two sub-folders namedapp
andmnt
.cd looker_backup
cd looker_backup - Copy files to the insights Looker pod and then restart the deployment.
kubectl cp .\app\workdir ${PodName}:/app -n uipath -c insightslooker kubectl cp .\mnt\lookerfiles ${PodName}:/mnt -n uipath -c insightslooker kubectl rollout restart statefulset insights-insightslooker -n uipath
kubectl cp .\app\workdir ${PodName}:/app -n uipath -c insightslooker kubectl cp .\mnt\lookerfiles ${PodName}:/mnt -n uipath -c insightslooker kubectl rollout restart statefulset insights-insightslooker -n uipath