ai-center
2023.4
false
- Release Notes
- Before you begin
- Getting started
- Installing AI Center
- Migration and upgrade
- Projects
- Datasets
- Data Labeling
- ML packages
- Out of the box packages
- Pipelines- Training pipelines
- Evaluation pipelines
- Full pipelines
 
- Managing pipelines
- Closing the loop
 
- ML Skills
- ML Logs
- Document UnderstandingTM in AI Center
- AI Center API
- How to
- Licensing
- Basic Troubleshooting Guide

AI Center User Guide
Last updated Apr 15, 2025
An Evaluation Pipeline is used to evaluate a trained machine learning model. To use this pipeline, the package must contain
               code to evaluate a model (the 
            
         evaluate() function in the train.py file). This code, together with a dataset or sub-folder within a dataset, produce a score (the return of the evaluate() function) and any arbitrary outputs the user would like to persist in addition to the score.
            Create a new evaluation pipeline as described here. Make sure to provide the following evaluation pipeline specific information:
- In the Pipeline type field, select Evaluation run.
- In the Choose evaluation dataset field, select a dataset or folder from which you want to import data for evaluation. All files in this dataset/folder should
                     be available locally during the runtime of the pipeline, being passed to the argument to your evaluate()function.
- In the Enter parameters section, enter the environment variables defined and used by your pipeline, if any. The environment variables are:
                     - artifacts_directory, with default value artifacts: This defines the path to a directory that will be persisted as ancillary data related to this pipeline. Most, if not all users, will never have the need to override this through the UI. Anything can be saved during pipeline execution including images, pdfs, and subfolders. Concretely, any data your code writes in the directory specified by the path- os.environ['artifacts_directory']will be uploaded at the end of the pipeline run and will be viewable from the Pipeline details page.
- save_test_data, with default value false: If set to true,- data_directoryfolder will be uploaded at the end of the pipeline run as an output of the pipeline under directory- data_directory.Note: The pipeline execution might take some time. Check back to it after a while to see its status.After the pipeline was executed, in the Pipelines page, the pipeline's status changed to Successful. The Pipeline Details page displays the arbitrary files and folders related to the pipeline run. In our example, the run created a file called- my-evaluate-artifact.txt.
 
This example is a conceptually analogous execution of an
                  evaluation pipeline on some package, for example version 1.1, the output of a training
                  pipeline on version 1.0.
                  
               
               
               Note: This is a simplified example. Its
                     purpose is to illustrate how datasets and packages interact in an evaluation
                     pipeline. The steps are merely conceptual and do not represent how the platform
                     works.
                  
                  
               The 
               
               _results.json file contains a summary of the pipeline run execution,
                  exposing all inputs/outputs and execution times for an evaluation pipeline.
               {
    "parameters": {
        "pipeline": "< Pipeline_name >",
        "inputs": {
            "package": "<Package_name>",
            "version": "<version_number>",
            "evaluation_data": "<storage_directory>",
            "gpu": "True/False"
        },
        "env": {
            "key": "value",
            ...
        }
    },
    "run_summary": {
     "execution_time": <time>, #in seconds 
     "start_at": <timestamp>, #in seconds 
     "end_at": <timestamp>, #in seconds 
     "outputs": {
        "score": <score>, #float
        "train_data": "<test_storage_directory>", 
        "evaluation_data": "<test_storage_directory>/None", 
        "artifacts_data": "<artifacts_storage_directory>",
         }
    }
}{
    "parameters": {
        "pipeline": "< Pipeline_name >",
        "inputs": {
            "package": "<Package_name>",
            "version": "<version_number>",
            "evaluation_data": "<storage_directory>",
            "gpu": "True/False"
        },
        "env": {
            "key": "value",
            ...
        }
    },
    "run_summary": {
     "execution_time": <time>, #in seconds 
     "start_at": <timestamp>, #in seconds 
     "end_at": <timestamp>, #in seconds 
     "outputs": {
        "score": <score>, #float
        "train_data": "<test_storage_directory>", 
        "evaluation_data": "<test_storage_directory>/None", 
        "artifacts_data": "<artifacts_storage_directory>",
         }
    }
}Artifacts folder, visible only if not empty, is a folder regrouping all the
                  artifacts generated by the pipeline and saved under the
                  
               
               
               artifacts_directory folder.
               Dataset folder, existing only if 
               
               
            save_data was set to the default
                  true value, is a copy of the evaluation dataset folder.
               As in training pipelines, a user can set the
                  parameter 
               
               
            save_test_data = true to snapshot data
                  passed in for evaluation.