All Products
Search
Document Center

Platform For AI:Service architecture

Last Updated:Feb 26, 2026

This topic describes the architectural design and components of Platform for AI (PAI).

a98777f2a4ad660860b3da00b5e6a5b9.png

PAI uses a four-layer architecture that provides end-to-end capabilities for AI development and deployment:

  • Infrastructure layer (Computing resources and infrastructure)

    This layer provides the foundational computing and networking resources:

    • Infrastructure: Includes CPUs, GPUs, high-speed Remote Direct Memory Access (RDMA) networks, and Container Service for Kubernetes (ACK) for containerized workload orchestration.

    • Computing resources: Offers cloud-native computing resources (Lingjun specialized resources and general-purpose computing resources) alongside big data processing engines (MaxCompute and Flink).

  • Platform and tools layer (AI frameworks and acceleration)

    This layer delivers core AI development capabilities through frameworks, optimization tools, and end-to-end machine learning workflows:

  • Application layer (Model services and platforms)

    This layer integrates PAI with model ecosystems and application platforms, enabling seamless access to pre-trained models and AI services through the ModelScope community, PAI-DashScope API service, third-party Model-as-a-Service (MaaS) platforms, and Alibaba Cloud Model Studio.

  • Business layer (Industry solutions)

    This layer delivers domain-specific AI solutions for vertical industries including autonomous driving, AI for Science (AI4Science), financial risk management, and intelligent recommendation systems. For example, internal Alibaba Group systems leverage PAI for large-scale data mining tasks in search, recommendation, and financial service applications.