- Overview
- Requirements
- Installation
- Post-installation
- Cluster administration
- Managing products
- Managing the cluster in ArgoCD
- Setting up the external NFS server
- Automated: Enabling the Backup on the Cluster
- Automated: Disabling the Backup on the Cluster
- Automated, Online: Restoring the Cluster
- Automated, Offline: Restoring the Cluster
- Manual: Enabling the Backup on the Cluster
- Manual: Disabling the Backup on the Cluster
- Manual, Online: Restoring the Cluster
- Manual, Offline: Restoring the Cluster
- Additional configuration
- Migrating objectstore from persistent volume to raw disks
- Monitoring and alerting
- Migration and upgrade
- Migration options
- Step 1: Moving the Identity organization data from standalone to Automation Suite
- Step 2: Restoring the standalone product database
- Step 3: Backing up the platform database in Automation Suite
- Step 4: Merging organizations in Automation Suite
- Step 5: Updating the migrated product connection strings
- Step 6: Migrating standalone Insights
- Step 7: Deleting the default tenant
- B) Single tenant migration
- Product-specific configuration
- Best practices and maintenance
- Troubleshooting
- How to Troubleshoot Services During Installation
- How to Uninstall the Cluster
- How to clean up offline artifacts to improve disk space
- How to clear Redis data
- How to enable Istio logging
- How to manually clean up logs
- How to clean up old logs stored in the sf-logs bucket
- How to disable streaming logs for AI Center
- How to debug failed Automation Suite installations
- How to delete images from the old installer after upgrade
- How to automatically clean up Longhorn snapshots
- How to disable TX checksum offloading
- How to address weak ciphers in TLS 1.2
- Unable to run an offline installation on RHEL 8.4 OS
- Error in Downloading the Bundle
- Offline installation fails because of missing binary
- Certificate issue in offline installation
- First installation fails during Longhorn setup
- SQL connection string validation error
- Prerequisite check for selinux iscsid module fails
- Azure disk not marked as SSD
- Failure After Certificate Update
- Automation Suite not working after OS upgrade
- Automation Suite Requires Backlog_wait_time to Be Set 1
- Volume unable to mount due to not being ready for workloads
- RKE2 fails during installation and upgrade
- Failure to upload or download data in objectstore
- PVC resize does not heal Ceph
- Failure to Resize Objectstore PVC
- Rook Ceph or Looker pod stuck in Init state
- StatefulSet volume attachment error
- Failure to create persistent volumes
- Storage reclamation patch
- Backup failed due to TooManySnapshots error
- All Longhorn replicas are faulted
- Setting a timeout interval for the management portals
- Update the underlying directory connections
- Cannot Log in After Migration
- Kinit: Cannot Find KDC for Realm <AD Domain> While Getting Initial Credentials
- Kinit: Keytab Contains No Suitable Keys for *** While Getting Initial Credentials
- GSSAPI Operation Failed With Error: An Invalid Status Code Was Supplied (Client's Credentials Have Been Revoked).
- Alarm Received for Failed Kerberos-tgt-update Job
- SSPI Provider: Server Not Found in Kerberos Database
- Login Failed for User <ADDOMAIN><aduser>. Reason: The Account Is Disabled.
- ArgoCD login failed
- Failure to get the sandbox image
- Pods not showing in ArgoCD UI
- Redis Probe Failure
- RKE2 Server Fails to Start
- Secret Not Found in UiPath Namespace
- After the Initial Install, ArgoCD App Went Into Progressing State
- MongoDB pods in CrashLoopBackOff or pending PVC provisioning after deletion
- Unexpected Inconsistency; Run Fsck Manually
- Degraded MongoDB or Business Applications After Cluster Restore
- Missing Self-heal-operator and Sf-k8-utils Repo
- Unhealthy Services After Cluster Restore or Rollback
- RabbitMQ pod stuck in CrashLoopBackOff
- Prometheus in CrashloopBackoff state with out-of-memory (OOM) error
- Missing Ceph-rook metrics from monitoring dashboards
- Pods cannot communicate with FQDN in a proxy environment
- Using the Automation Suite Diagnostics Tool
- Using the Automation Suite support bundle
- Exploring Logs
Managing products
You can enable and disable products in Automation Suite at any point post-installation.
cluster_config.json
file stored in the
/opt/UiPathAutomationSuite
folder, and rerun the service
installer step.
cluster_config.json
file:
"actioncenter": {
"enabled": "true" //Set to "false" to disable the Action Center
}
"actioncenter": {
"enabled": "true" //Set to "false" to disable the Action Center
}
cluster_config.json
file:
"apps": {
"enabled": "true" //Set to "false" to disable the Apps
}
"apps": {
"enabled": "true" //Set to "false" to disable the Apps
}
enable
flag and an SQL database.
sql_connection_string_template_jdbc
in cluster_config.json
, then the default AI Center database name is AutomationSuite_AICenter
.
sql_connection_str
inside the AI Center field. This overrides the default database and connection string template set in sql_connection_string_template_jdbc
.
cluster_config.json
:
"aicenter": {
"enabled": "true", //Set to "false" to disable the AICenter
"sql_connection_str": "" //Optional and only require to override the default database name
}
"aicenter": {
"enabled": "true", //Set to "false" to disable the AICenter
"sql_connection_str": "" //Optional and only require to override the default database name
}
automation_hub
section of the cluster_config.json
file, set the enabled
flag
to true
:
"automation_hub": {
"enabled": "true" //Set to "false" to disable Automation Hub
"sql_connection_str": "" //Optional and only require to override the default database name
}
"automation_hub": {
"enabled": "true" //Set to "false" to disable Automation Hub
"sql_connection_str": "" //Optional and only require to override the default database name
}
enabled
flag to false
in the
automation_hub
section of the cluster_config.json
file.
enable
flag and an SQL database.
sql_connection_string_template
in cluster_config.json
, then the default database name for
Automation Ops is AutomationSuite_Platform
.
sql_connection_str
inside the Automation Ops field. This overrides the default database and connection string template set in sql_connection_string_template
.
cluster_config.json
:
"automation_ops": {
"enabled": "true", //Set to "false" to disable the Automation Ops
"sql_connection_str": "" //Optional and only require to override the default database name
}
"automation_ops": {
"enabled": "true", //Set to "false" to disable the Automation Ops
"sql_connection_str": "" //Optional and only require to override the default database name
}
enable
flag and an SQL database.
sql_connection_string_template
in cluster_config.json
, then the default database name is
AutomationSuite_DataService
.
sql_connection_str
inside the Data Service field. This overrides the default database and connection string template set in sql_connection_string_template
.
cluster_config.json
:
"dataservice": {
"enabled": "true", //Set to "false" to disable the Data Service,
"sql_connection_str": "" //Optional and only require to override the default database name
}
"dataservice": {
"enabled": "true", //Set to "false" to disable the Data Service,
"sql_connection_str": "" //Optional and only require to override the default database name
}
enable
flag and an SQL database.
sql_connection_string_template_odbc
in
cluster_config.json
, then the default database name
is AutomationSuite_DU_Datamanager
.
sql_connection_str
inside the Document Understanding field. This overrides the default database and connection string template set in sql_connection_string_template_odbc
.
If you perform an offline installation, you need to download the offline bundle for Document Understanding as well. Before enabling Document Understanding, please refer to Install and Use Document Understanding.
cluster_config.json
:
"documentunderstanding": {
"enabled": true,
"sql_connection_str": "***" // dotnet connection string,
"datamanager": {
"sql_connection_str": "***" // odbc connection string
"pyodbc_sql_connection_str": "***" // python sql connection string
}
}
"documentunderstanding": {
"enabled": true,
"sql_connection_str": "***" // dotnet connection string,
"datamanager": {
"sql_connection_str": "***" // odbc connection string
"pyodbc_sql_connection_str": "***" // python sql connection string
}
}
enable
flag and an SQL database.
sql_connection_string_template
in cluster_config.json
, then the default database name is AutomationSuite_Insights
.
sql_connection_str
inside the Insights field. This overrides the default database and connection string template set in sql_connection_string_template
.
Insights has an optional SMTP configuration to enable receiving email notifications. Refer to Advanced installation experience for more details.
cluster_config.json
:
"insights": {
"enabled": "true", //Set to "false" to disable the Insights,
"sql_connection_str": "" //Optional and only require to override the default database name
}
"insights": {
"enabled": "true", //Set to "false" to disable the Insights,
"sql_connection_str": "" //Optional and only require to override the default database name
}
enable
flag and an SQL Database.
sql_connection_string_template
in cluster_config.json
, then the default database name is
AutomationSuite_Task_Mining
.
To change the name to something other than the default, you can update the sql_connection_str inside the Task Mining field to override the default database and connection string template set in sql_connection_string_template
Task Mining requires a few additional steps, including the provisioning of a dedicated agent node for AI-specific workloads.
Before enabling Task Mining, please refer Task Mining installation section Adding a dedicated agent node for Task Mining.
cluster_config.json
:
"task_mining": {
"enabled": "true", //Set to "false" to disable the Task Mining,
"sql_connection_str": "" //Optional and only require to override the default database name
}
"task_mining": {
"enabled": "true", //Set to "false" to disable the Task Mining,
"sql_connection_str": "" //Optional and only require to override the default database name
}
Un-tainting for workloads
If desired, the Task Mining node can be repurposed as an agent node. To accomplish this, run the following commands:
- Check an actual taint using
kubectl describe node task-mining_nodename
command. - Run the command
kubectl taint nodes task-mining_nodename task.mining/cpu=present:NoSchedule-
. - Check taint has disappeared using
kubectl describe node task-mining_nodename
command.
enable
flag and an SQL database.
sql_connection_string_template
in cluster_config.json
, then the default database name is AutomationSuite_Test_Manager
.
sql_connection_str
inside the Test Manager field. This overrides the default database and connection string template set in sql_connection_string_template
.
cluster_config.json
:
"test_manager": {
"enabled": "true", //Set to "false" to disable the Test Manager,
"sql_connection_str": "" //Optional and only require to override the default database name
}
"test_manager": {
"enabled": "true", //Set to "false" to disable the Test Manager,
"sql_connection_str": "" //Optional and only require to override the default database name
}
cluster_config.json
, run the following command on the installer to update the service configuration.
-
For online installations:
./install-uipath.sh -i cluster_config.json -o output.json -s --accept-license-agreement
./install-uipath.sh -i cluster_config.json -o output.json -s --accept-license-agreement -
For offline installations:
./install-uipath.sh -i cluster_config.json -o output.json -s --install-type offline --accept-license-agreement
./install-uipath.sh -i cluster_config.json -o output.json -s --install-type offline --accept-license-agreement
- Step 1: Changing the product selection in the configuration file
- Enabling or disabling Action Center
- Enabling or Disabling Apps
- Enabling or Disabling AI Center
- Enabling or disabling Automation Hub
- Enabling or disabling Automation Ops
- Enabling or disabling Data Service
- Enabling or disabling Document Understanding
- Enabling or Disabling Insights
- Enabling or disabling Task Mining
- Enabling or Disabling Test Manager
- Step 2: Running the installer script to update the new product configuration