automation-suite
2024.10
true
- Overview
- Requirements
- Pre-installation- Preparing the installation
- Installing and configuring the service mesh
- Downloading the installation packages
- Configuring the OCI-compliant registry
- Granting installation permissions
- Installing and configuring the GitOps tool
- Deploying Redis through OperatorHub
- Applying miscellaneous configurations
- Running uipathctl
 
- Installation
- Post-installation
- Migration and upgrade- Upgrading Automation Suite
- Migrating standalone products to Automation Suite
- Step 1: Restoring the standalone product database
- Step 2: Updating the schema of the restored product database
- Step 3: Moving the Identity organization data from standalone to Automation Suite
- Step 4: Backing up the platform database in Automation Suite
- Step 5: Merging organizations in Automation Suite
- Step 6: Updating the migrated product connection strings
- Step 7: Migrating standalone Orchestrator
- Step 8: Migrating standalone Insights
- Step 9: Deleting the default tenant
 
 
- Performing a single tenant migration
- Migrating between Automation Suite clusters
 
- Monitoring and alerting
- Cluster administration
- Product-specific configuration
- Troubleshooting- Log streaming does not work in proxy setups
- 500 errors and rate limiting on S3 requests in ODF
 
 
 

Automation Suite on OpenShift installation guide
Last updated Oct 8, 2025
Log forwarding does not work in Proxy setups because the proxy environment variables were not set in the logging pods.
- 
                        
                        Set thehttp_proxy,https_proxyandno_proxyenvironment variables. Example:export http_proxy=http://<proxy>:3128 export https_proxy=http://<proxy>:3128 export no_proxy=<fqdn>,.<fqdn>,10.0.0.0/8,kerberossql.autosuitead.local,kerberospostgres.AUTOSUITEAD.LOCAL,rook-ceph-rgw-rook-ceph.rook-ceph.svc.cluster.local,localhost,127.0.0.1,kubernetes,kubernetes.default,kubernetes.default.svc,kubernetes.default.svc.cluster,kubernetes.default.svc.cluster.local,.svc,.svc.cluster,.svc.cluster.local,.svc.cluster.local.,.uipath.svc.cluster.local,argocd-repo-server,istiod.istio-system.svc,logging-operator-logging-fluentd.logging.svc.cluster.local,.local,.cluster,ai-helper-svc,ai-pkgmanager-svc,ai-deployer-svc,ai-appmanager-svc,ai-trainer-svc,studioweb-backend,studioweb-frontend,studioweb-typecacheexport http_proxy=http://<proxy>:3128 export https_proxy=http://<proxy>:3128 export no_proxy=<fqdn>,.<fqdn>,10.0.0.0/8,kerberossql.autosuitead.local,kerberospostgres.AUTOSUITEAD.LOCAL,rook-ceph-rgw-rook-ceph.rook-ceph.svc.cluster.local,localhost,127.0.0.1,kubernetes,kubernetes.default,kubernetes.default.svc,kubernetes.default.svc.cluster,kubernetes.default.svc.cluster.local,.svc,.svc.cluster,.svc.cluster.local,.svc.cluster.local.,.uipath.svc.cluster.local,argocd-repo-server,istiod.istio-system.svc,logging-operator-logging-fluentd.logging.svc.cluster.local,.local,.cluster,ai-helper-svc,ai-pkgmanager-svc,ai-deployer-svc,ai-appmanager-svc,ai-trainer-svc,studioweb-backend,studioweb-frontend,studioweb-typecache
- 
                        
                        Execute the following script which ingests the environment variables into the logging pods and restarts them. #!/bin/bash set -euo pipefail APP_NAME="logging" NAMESPACE="argocd" HTTP_PROXY="${http_proxy:-}" HTTPS_PROXY="${https_proxy:-}" NO_PROXY="${no_proxy:-}" # Create a temporary JSON patch PATCH_FILE=$(mktemp) cat > "$PATCH_FILE" <<EOF { "spec": { "source": { "helm": { "parameters": [ { "name": "logging-operator.env[0].name", "value": "http_proxy" }, { "name": "logging-operator.env[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.env[1].name", "value": "https_proxy" }, { "name": "logging-operator.env[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.env[2].name", "value": "no_proxy" }, { "name": "logging-operator.env[2].value", "value": "${NO_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[0].name", "value": "http_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[1].name", "value": "https_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[2].name", "value": "no_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[2].value", "value": "${NO_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[0].name", "value": "http_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[1].name", "value": "https_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[2].name", "value": "no_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[2].value", "value": "${NO_PROXY}" } ] } } } } EOF # Patch the Argo CD Application kubectl patch application "$APP_NAME" -n "$NAMESPACE" --type merge --patch-file "$PATCH_FILE" # Cleanup rm -f "$PATCH_FILE" echo "Patched Argo CD Application '$APP_NAME' with proxy parameters." echo "Restarting logging pods..." kubectl rollout restart deploy/logging-logging-operator -n logging kubectl rollout restart sts/logging-fluentd -n logging kubectl rollout restart ds/logging-fluentbit -n logging echo "Rollout restart completed"#!/bin/bash set -euo pipefail APP_NAME="logging" NAMESPACE="argocd" HTTP_PROXY="${http_proxy:-}" HTTPS_PROXY="${https_proxy:-}" NO_PROXY="${no_proxy:-}" # Create a temporary JSON patch PATCH_FILE=$(mktemp) cat > "$PATCH_FILE" <<EOF { "spec": { "source": { "helm": { "parameters": [ { "name": "logging-operator.env[0].name", "value": "http_proxy" }, { "name": "logging-operator.env[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.env[1].name", "value": "https_proxy" }, { "name": "logging-operator.env[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.env[2].name", "value": "no_proxy" }, { "name": "logging-operator.env[2].value", "value": "${NO_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[0].name", "value": "http_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[1].name", "value": "https_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.logging.fluentd.envVars[2].name", "value": "no_proxy" }, { "name": "logging-operator.logging.fluentd.envVars[2].value", "value": "${NO_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[0].name", "value": "http_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[0].value", "value": "${HTTP_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[1].name", "value": "https_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[1].value", "value": "${HTTPS_PROXY}" }, { "name": "logging-operator.logging.fluentbit.envVars[2].name", "value": "no_proxy" }, { "name": "logging-operator.logging.fluentbit.envVars[2].value", "value": "${NO_PROXY}" } ] } } } } EOF # Patch the Argo CD Application kubectl patch application "$APP_NAME" -n "$NAMESPACE" --type merge --patch-file "$PATCH_FILE" # Cleanup rm -f "$PATCH_FILE" echo "Patched Argo CD Application '$APP_NAME' with proxy parameters." echo "Restarting logging pods..." kubectl rollout restart deploy/logging-logging-operator -n logging kubectl rollout restart sts/logging-fluentd -n logging kubectl rollout restart ds/logging-fluentbit -n logging echo "Rollout restart completed"