- Overview
- Requirements
- Pre-installation
- Preparing the installation
- Installing and configuring the service mesh
- Downloading the installation packages
- Configuring the OCI-compliant registry
- Granting installation permissions
- Installing and configuring the GitOps tool
- Deploying Redis through OperatorHub
- Applying miscellaneous configurations
- Running uipathctl
- Installation
- Post-installation
- Migration and upgrade
- Upgrading Automation Suite
- Migrating standalone products to Automation Suite
- Step 1: Restoring the standalone product database
- Step 2: Updating the schema of the restored product database
- Step 3: Moving the Identity organization data from standalone to Automation Suite
- Step 4: Backing up the platform database in Automation Suite
- Step 5: Merging organizations in Automation Suite
- Step 6: Updating the migrated product connection strings
- Step 7: Migrating standalone Orchestrator
- Step 8: Migrating standalone Insights
- Step 9: Deleting the default tenant
- Performing a single tenant migration
- Migrating between Automation Suite clusters
- Cluster administration
- Product-specific configuration
- Remapping the organization IDs
- Migrating Looker data to Automation Suite
- Performing Insights database maintenance
- Configuring the Insights custom security context
- Configuring log rules
- Troubleshooting

Automation Suite on OpenShift installation guide
Performing Insights database maintenance
dbo
and read
. Both tables store data from Orchestrator databases and pass data between each other. Over time, tables store a large amount
of data that can impact database performance. You can free up space by removing data that is not relevant or outdated.
The estimated threshold for storing historical data is set to
- Up to 100 million jobs
- Up to 1 billion job events
- Up to 100 million queue items
- Up to 500 million queue item events
- Up to 1 billion robot logs
[dbo].Jobs
, [dbo].JobEvents
, [dbo].QueueItems
, [dbo].QueueItemEvents
, [dbo].RobotLogs
tables, or just the [dbo].RobotLogs
.
PROCEDURE
[read].[Delete_Insights_Data_Read]
and PROCEDURE
[read].[Delete_Process_Logs_Read]
scripts, the Insights module needs to be
disabled.
- Scripts with the
_DBO
suffix are used to delete data from DBO schema tables, - Scripts with the
_Read
suffix are used to delete data from Read schema tables. - Large-sized batches might impact the speed of execution. Consider using the default batch size, which is set to 100000.
PROCEDURE [read].[Delete_Insights_Data_Read]
and
PROCEDURE [read].[Delete_Process_Logs_Read]
scripts, the Insights module
needs to be disabled.
Jobs
, JobEvents
, QueueItems
, QueueItemEvents
, RobotLogs
) for a cut-off timestamp.
SQL Stored Procedures | Parameter | Description |
---|---|---|
Delete_Insights_Data_DBO |
| Delete data from [dbo].[Jobs] ,
[dbo].[JobEvents] , [dbo].[QueueItems] , [dbo].[QueueItemEvents] , [dbo].[RobotLogs] tables based on a cut-off timestamp.
|
Delete_Insights_Data_Read | N/A | Truncate [read].Jobs , [read].JobEvents ,[read].QueueItems , [read].QueueItemEvents , [read].RobotLogs tables.
|
The Insights Dashboard widgets are available again after SQL Stored Procedure 1 has finished, and the data has been backfilled. Keep in mind that the backfill process may take hours to complete given that you have a large amount of data.
Delete_Insights_Data_DBO
SQL Script, it creates a QueueItems.IX_CreationTime
nonclustered index with the purpose of making future executions faster.
Generate and save this SQL Stored Procedure to delete RobotsLogs based on ProcessNames and Tenant ID.
SQL Stored Procedures | Parameter | Description |
---|---|---|
Delete_Process_Logs_DBO |
The Orchestrator process name.
Tenant ID is required as two tenants might run processes with identical names.
Delete data in batches of 100000. See Considerations. | Delete data from [dbo].[RobotLogs] table by ProcessName and Tenant ID.
|
Delete_Process_Logs_Read |
| Truncate [read].RobotLogs table.
|