This topic describes the terms related to management, AI development, and modules in Machine Learning Platform for AI (PAI).
Terms related to management
Term | Description |
workspace | Workspace is a key concept in PAI. Workspaces allow your organization or team to manage computing resources and user permissions in a centralized manner and provide tools and features for collaboration at every stage of AI development. PAI runs on top of DataWorks. Therefore, workspaces in PAI are mapped to workspaces in DataWorks. Default Workspace: The default workspace contains commonly used pay-as-you-go resources. You need to activate the default workspace before you can use these resources. The default workspace can help first-time users quickly get started with model development and training without the need to understand concepts such as resource groups. |
Deep Learning Containers (DLC) | DLC is a cloud-native platform for basic AI computing jobs. DLC provides an elastic, stable, easy-to-use, and high-performance environment for model training. DLC provides multiple algorithm frameworks, allows you to run a large number of deep learning jobs in a distributed manner, and supports custom algorithm frameworks. DLC supports the following types of clusters:
|
resource group |
|
member | Members are Alibaba Cloud accounts or Resource Access Management (RAM) users that join workspaces. Members in the same workspace can assume different roles to collaborate throughout the AI development pipeline. Only the owner and administrators of a workspace can modify the members in the workspace. |
role | Roles are mappings between members and permissions. You can use the system roles or create custom roles. The system roles include:
|
dependency | PAI relies on other Alibaba Cloud services. To use the features provided by PAI, you must first activate these Alibaba Cloud services and complete RAM authorization by using an Alibaba Cloud account or the resource administrator role. The Alibaba Cloud services that you need to activate include Object Storage Service (OSS), File Storage NAS (NAS), Log Service, Container Registry, and API Gateway. |
Terms related to AI development
Term | Description |
dataset | Datasets are used in labeling, model training, and model evaluation. You can create datasets to use structured data or unstructured data or mount directories in datastores such as OSS, NAS, and MaxCompute. In addition, you can centrally manage the storage, versions, and schemas of datasets in PAI. |
pipeline | Pipelines are directed acyclic graphs (DAGs) that consist of upstream components and downstream components connected based on logical scheduling. You can submit multiple runs for a pipeline to generate PipelineRuns. |
PipelineDraft | PipelineDrafts are configurable pipeline objects on the canvas of Designer. You can edit a PipelineDraft to generate multiple pipelines. You can submit runs for PipelineDrafts to generate PipelineRuns. |
component | Components are the smallest configurable units in pipelines and PipelineDrafts, and the smallest executable units in PipelineRuns. Components consist of the following types:
|
node | Nodes represent components that are dragged and dropped to the canvas to form a pipeline. |
snapshot | Each time the system runs a single node in a PipelineDraft, multiple nodes in the PipelineDraft, or the entire PipelineDraft, the system creates a snapshot for the configurations of the PipelineDraft. The snapshot includes the node configurations, runtime parameters, and execution mode. Snapshots can be used in PipelineDraft versioning and configuration rollbacks. |
PipelineRun | A PipelineRun represents a single run of a pipeline. After you use Designer to submit a run for a PipelineDraft or use the SDK to submit a run for a pipeline, a PipelineRun is generated. |
job | Jobs, such as DLC jobs, use computing resources. The resource environment in which jobs run belongs to the user. |
run | A run represents a single execution of a pipeline. Run is equivalent to the same concept in MLflow. All runs must belong to an experiment. You can use runs to track the training jobs that you submitted in Machine Learning Platform for AI. You can also use the MLflow client to submit runs from an on-premises machine. A run can contain multiple jobs. |
model | Models are generated by training jobs based on datasets, algorithms, and code. You can use models to make predictions. |
Processor | A processor is a package of online prediction logic, including the logic for loading models and handling requests. In most cases, processors are deployed together with model files to provision services. Processors consist of the following types:
|
service | You can deploy model files together with the online prediction logic as services. EAS allows you to create, update, start, stop, scale out, and scale in services. |
image | PAI allows you to manage the following Docker images as AI assets:
You can use images in pipelines to create custom components to complete specific jobs. Images can also provide runtime environments for DSW instances and training jobs. |
instance | Instances are the smallest units for provisioning compute resources. Instances consist of the following types:
|
Terms related to modules
Term | Description |
iTAG | iTAG is a dataset labeling tool that is developed with black box models to help improve the quality and efficiency of dataset labeling. |
Designer | Designer is intended for pipeline design in the AI sector. Designer provides a variety of built-in machine learning algorithm components. You can drag and drop these components to train models without coding. |
Data Science Workshop (DSW) | DSW is an integrated development environment (IDE) intended for interactive AI development in the cloud. DSW consists of Notebook, VS Code, and Terminal. You can use images to deploy DSW instances that use NAS as the storage. |
Deep Learning Containers (DLC) | You can submit training jobs to computing resource groups, such as DLC clusters, in the current workspace. After you submit training jobs, you can view the details about the jobs in the Jobs module of the PAI console. |
Elastic Algorithm Service (EAS) | EAS allows you to deploy complex models as services on a large scale with a few clicks. EAS supports real-time scaling and provides a sophisticated monitoring and maintenance system. |
AI Asset Management | This module allows you to manage key AI assets, including datasets, models, and source code repositories. |
Scenario-based Solution | A collection of solutions provided by PAI to help you resolve issues in vertical markets. |