test-manager
latest
false
- Getting started
- Project management
- Documents
- Working with Change Impact Analysis
- Create test cases
- Assigning test cases to requirements
- Cloning test cases
- Exporting test cases
- Linking test cases in Studio to Test Manager
- Delete test cases
- Manual test cases
- Importing manual test cases
- Document test cases with Task Capture
- Parameters
- Enabling governance at project level
- Disabling governance at project level
- Enabling governance at test-case level
- Disabling governance at test-case level
- Managing approvers for governed test cases
- Managing governed test cases in the In Work state
- Managing governeed test cases in the In Review state
- Managing governed objects in the Signed state
- Managing comments for governed test cases
- Applying filters and views
- Importing Orchestrator test sets
- Creating test sets
- Adding test cases to a test set
- Assigning default users in test set execution
- Enabling activity coverage
- Enabling Healing Agent
- Configuring test sets for specific execution folders and robots
- Overriding parameters
- Cloning test sets
- Exporting test sets
- Applying filters and views
- Creating automated tests
- Step 1: Complete prerequisites for performance scenarios
- Step 2: Add performance scenarios
- Step 3: Add test cases to load groups
- Step 4: Configure load group settings
- Executing performance scenarios
- Known limitations for performance testing
- Best practices for performance testing
- Troubleshooting performance testing
- Accessibility testing for Test Cloud
- Searching with Autopilot
- Project operations and utilities
- Test Manager settings
- ALM tool integration
- API integration
- Troubleshooting
Test Manager user guide
Last updated May 7, 2026
When configuring a load group, you define how the selected test case will be executed at scale. Each parameter allows you to model realistic user behavior and enforce performance expectations.
- Log in to Test Manager.
- Go to Performance Scenarios, open a scenario, and expand a load group.
- Under Virtual users, specify the number of concurrent simulated users assigned to the test case. Each group can support up to 500 virtual users, with a total of 2,000 across a scenario. Check the number of VU units and machines consumed after a Dry Run, based on your selection.
- Define performance expectations. Breaching these thresholds will highlight warnings or errors in the results dashboard, allowing you to identify unstable scenarios.
-
- In the Max. response time threshold field, specify the upper time limit (in ms) that responses should not exceed (e.g., 5,000 ms for Group 1).
-
- In the Average error rate threshold, specify the acceptable failure rate (%) across all executions (e.g., 1%).
-
- In the Data source field, link a Data Fabric entity (also known as a Data Service entity) as the data source to feed parameterized inputs such as user credentials or account details. This ensures that each virtual user can run with unique data, avoiding collisions and making execution closer to real-world usage.
Tip:
To use Data Fabric for Performance Testing, ensure that your test case supports test data consumption from Data Fabric. For information on using Data Fabric test data in your tests, refer to Data-driven testing. When this test case is included in a Load Group, you can configure a specific data source at the group level. This allows you to define and use a dedicated Data Fabric source for your performance test executions.
- Decide whether you want to keep Studio parameters as is or override them. To override Studio parameters, under Parameters, select Override. In the Override parameters dialog, perform the required action.
- To override Studio parameters individually, toggle Override. For String types, enter a new value. For Boolean types, select True or False. When finished, select Save.
- To override all Studio parameters, select Override all. Enter or select the new values. When finished, select Save.
- To remove all Studio parameter overrides, select Remove overrides, and then select Save.
- Specify how the load is applied over time. Together, these settings allow you to model realistic traffic patterns such as gradual adoption, steady-state usage, and controlled shutdown.
- The Delay field shows the wait time before the load group starts execution (e.g., 0 minutes).
- The Ramp up field shows the period over which the number of concurrent users is gradually increased until the maximum number of users is reached (e.g., 2 minutes).
- The Peak field shows the duration during which the maximum number of users remain active (e.g., 26 minutes).
- The Ramp down field shows the period over which the number of users gradually decreases to zero (e.g., 2 minutes).