All Products
Search
Document Center

Container Service for Kubernetes:Create and manage a node pool

Last Updated:Jan 21, 2025

Container Service for Kubernetes (ACK) provides node pools for you to manage nodes in groups. A node pool is a group of nodes that have the same configurations, such as instance specifications, operating system, labels, and taints. You can create one or more node pools of different types and configurations in an ACK cluster. The configurations of a node pool consist of node attributes, such as the instance specifications, zones, operating system, labels, and taints. After you create a node pool, you can manage nodes in the node pool in a centralized manner.

Before you create a node pool, we recommend that you read the Node pool overview to familiarize yourself with the basic information, use scenarios, relevant features, and billing rules of node pools.

Console operations

On the Node Pools page of the cluster that you want to manage in the ACK console, you can create, edit, or delete a node pool. You can also view the details of a node pool.

  1. Log on to the ACK console. In the left-side navigation pane, click Clusters.

  2. On the Clusters page, find the cluster that you want to manage and click its name. In the left-side navigation pane, choose Nodes > Node Pools.

Create a node pool

When you create a node pool in the ACK console, you can configure the basic, network, and storage configurations. Some node pool parameters, especially the parameters related to the node pool availability and network, cannot be modified after the node pool is created. The following tables describe these parameters. Creating a node pool in a cluster does not affect the nodes and applications deployed in other node pools of the cluster.

Note

You can also create a node pool by calling the ACK API or by using Terraform. For more information, see CreateClusterNodePool or Use Terraform to create a node pool that has auto scaling enabled.

  1. On the Node Pools page, click Create Node Pool. In the Create Node Pool dialog box, configure the node pool parameters.

    After you create the node pool, you can modify the node pool parameters on the Edit Node Pool page. The Modifiable column in the following tables indicates whether the corresponding parameter can be modified after the node pool is created. 错 indicates that the parameter cannot be modified. 对 indicates that the parameter can be modified.

    Basic configurations

    Parameter

    Description

    Modifiable

    Node Pool Name

    Specify a node pool name.

    对

    Region

    By default, the region in which the cluster resides is selected. You cannot change the region.

    错

    Confidential Computing

    Note
    • To use confidential computing, submit a ticket to apply to be added to the whitelist.

    • This parameter is available only when you select containerd for the Container Runtime parameter.

    Specify whether to enable confidential computing. ACK provides an all-in-one cloud-native confidential computing solution based on hardware encryption technologies. Confidential computing ensures data security, integrity, and confidentiality. It simplifies the development and delivery of trusted or confidential applications to reduce costs. For more information, see TEE-based confidential computing.

    错

    Container Runtime

    Specify the container runtime based on the Kubernetes version. For more information about how to select a container runtime, see Comparison among Docker, containerd, and Sandboxed-Container.

    • containerd: containerd is recommended for all Kubernetes versions.

    • Sandboxed-Container: supports Kubernetes 1.31 and earlier.

    • Docker (deprecated): supports Kubernetes 1.22 and earlier.

    错

    Managed node pool settings

    Managed Node Pool

    Managed node pools provided by ACK support auto repair and auto CVE patching. This significantly reduces your O&M workload and improves node security. You can click Set to configure the detailed maintenance policies.

    对

    Auto Recovery Rule

    This parameter is available after you select Enable for the managed node pool feature.

    After you select Restart Faulty Node, the system automatically restarts relevant components to repair nodes in the NotReady state and drains the nodes before the nodes are restarted.

    对

    Auto Update Rule

    This parameter is available after you select Enable for the managed node pool feature.

    After you select Automatically Update Kubelet and Containerd, the system automatically updates the kubelet when a new version is available. For more information, see Update a node pool.

    对

    Auto CVE Patching (OS)

    This parameter is available after you select Enable for the managed node pool feature.

    You can configure ACK to automatically patch high-risk, medium-risk, and low-risk vulnerabilities. For more information, see Enable auto repair for nodes and Patch OS CVE vulnerabilities for node pools.

    Some patches take effect only after you restart the ECS instances. After you enable Restart Nodes if Necessary to Patch CVE Vulnerabilities, ACK automatically restarts nodes on demand. If you do not select this option, you must manually restart nodes.

    对

    Maintenance Window

    This parameter is available after you select Enable for the managed node pool feature.

    Image updates, runtime updates, and Kubernetes version updates are automatically performed during the maintenance window.

    Click Set. In the Maintenance Window dialog box, set the Cycle, Started At, and Duration parameters and click OK.

    对

    Scaling Mode

    Manual and Auto scalings are supported. Computing resources are automatically adjusted based on your business requirements and policies to reduce cluster costs.

    • Manual: ACK adjusts the number of nodes in the node pool based on the value of the Expected Nodes parameter. The number of nodes is always the same as the value of the Expected Nodes parameter. For more information, see Manually scale a node pool.

    • Auto: When the capacity planning of the cluster cannot meet the requirements of pod scheduling, ACK automatically scales out nodes based on the configured minimum and maximum number of instances. By default, node instant scaling is enabled for clusters that run Kubernetes 1.24 and later. By default, node auto scaling is enabled for clusters that run Kubernetes 1.24 and later. For more information, see Overview of node scaling.

    对

    Network configurations

    Parameter

    Description

    Modifiable

    Network Settings

    VPC

    By default, the virtual private cloud (VPC) in which the cluster resides is selected. You cannot change the VPC.

    错

    vSwitch

    When the node pool is being scaled out, new nodes are created in the zones of the selected vSwitches based on the policy that you select for the Scaling Policy parameter. You can select vSwitches in the zones that you want to use.

    If no vSwitch is available, click Create vSwitch to create one. For more information, see Create and manage a vSwitch.

    错

    Instance and Image

    Parameter

    Description

    Modifiable

    Billing Method

    The default billing method used when ECS instances are scaled in a node pool. You can select Pay-As-You-Go, Subscription, or Preemptible Instance.

    • If you select the Subscription billing method, you must configure the Duration parameter and choose whether to enable Auto Renewal.

    • Preemptible Instance: ACK supports only Preemptible Instance with a protection period. You must also configure the Upper Price Limit of Current Instance Spec parameter.

      If the real-time market price of an instance type that you select is lower than the value of this parameter, a preemptible instance of this instance type is created. After the protection period (1 hour) ends, the system checks the spot price and resource availability of the instance type every 5 minutes. If the real-time market price exceeds your bid price or if the resource inventory is insufficient, the preemptible instance is released. For more information, see Best practices for preemptible instance-based node pools.

    To ensure that all nodes in a node pool use the same billing method, ACK does not allow you to change the billing method of a node pool from pay-as-you-go or subscription to preemptible instances. For example, you cannot switch the billing method of a node pool between pay-as-you-go or subscription and preemptible instances.

    Important
    • If you change the billing method of a node pool, the change takes effect only on newly added nodes. The existing nodes in the node pool still use the original billing method. For more information about how to change the billing method of existing nodes in a node pool, see Change the billing method of an instance from pay-as-you-go to subscription.

    • To ensure that all nodes use the same billing method, ACK does not allow you to change the billing method of a node pool from pay-as-you-go or subscription to preemptible instances, or change the billing method of a node pool from preemptible instances to pay-as-you-go or subscription.

    对

    Instance-related parameters

    Select the ECS instances used by the worker node pool based on instance types or attributes. You can filter instance families by attributes such as vCPU, memory, instance family, and architecture. For more information about how to configure nodes, see Suggestions on choosing ECS specifications for ACK clusters.

    When the node pool is scaled out, ECS instances of the selected instance types are created. The scaling policy of the node pool determines which instance types are used to create new nodes during scale-out activities. Select multiple instance types to improve the success rate of node pool scale-out operations.

    If the node pool fails to be scaled out because the instance types are unavailable or the instances are out of stock, you can specify more instance types for the node pool. The ACK console automatically evaluates the scalability of the node pool. You can check the scalability of the node pool when you create the node pool or after you create the node pool.

    If you select only GPU-accelerated instances, you can select Enable GPU Sharing on demand. For more information, see cGPU overview.

    对

    Operating System

    Alibaba Cloud Marketplace images is in canary release.
    Note
    • After you change the OS image of the node pool, the change takes effect only on newly added node. The existing nodes in the node pool still use the original OS image. For more information about how to update the OS images of existing nodes, see Node pool updates.

    • To ensure that all nodes in the node pool use the same OS image, ACK allows you to only update the node OS image to the latest version. ACK does not allow you to change the type of OS image.

    对

    Security Hardening

    Enable security hardening for the cluster. You cannot modify this parameter after the cluster is created.

    • Disable: disables security hardening for ECS instances.

    • MLPS Security Hardening: Alibaba Cloud provides baselines and the baseline check feature to help you check the compliance of Alibaba Cloud Linux 2 images and Alibaba Cloud Linux 3 images with the level 3 standards of Multi-Level Protection Scheme (MLPS) 2.0. MLPS Security Hardening enhances the security of OS images to meet the requirements of GB/T 22239-2019 Information Security Technology - Baseline for Classified Protection of Cybersecurity without compromising the compatibility and performance of the OS images. For more information, see ACK security hardening based on MLPS.

      Important

      After you enable MLPS Security Hardening, remote logons through SSH are prohibited for root users. You can use Virtual Network Computing (VNC) to log on to the OS from the ECS console and create regular users that are allowed to log on through SSH. For more information, see Connect to an instance by using VNC.

    • OS Security Hardening: You can enable Alibaba Cloud Linux Security Hardening only when the system image is an Alibaba Cloud Linux 2 or Alibaba Cloud Linux 3 image.

    错

    Logon Type

    If you select MLPS Security Hardening, only the Password option is supported.

    Valid values: Key Pair, Password, and Later.

    • Configure the logon type when you create the node pool:

      • Key Pair: Alibaba Cloud SSH key pairs provide a secure and convenient method to log on to ECS instances. An SSH key pair consists of a public key and a private key. SSH key pairs support only Linux instances.

        Configure the Username (select root or ecs-user as the username) and the Key Pair parameters.

      • Password: The password must be 8 to 30 characters in length, and can contain letters, digits, and special characters.

        Configure the Username (select root or ecs-user as the username) and the Password parameters.

    • Later: Bind a key pair or reset the password after the instance is created. For more information, see Bind an SSH key pair to an instance and Reset the logon password of an instance.

    对

    Username

    If you select Key Pair or Password for Logon Type, you must select root or ecs-user as the username.

    对

    Storage configurations

    Parameter

    Description

    Modifiable

    System Disk

    ESSD AutoPL, Enterprise SSD (ESSD), ESSD Entry, Standard SSD, and Ultra Disk are supported. The types of system disks that you can select vary based on the instance families that you select. Disk types that are not displayed in the drop-down list are not supported by the instance types that you select.

    ESSD custom performance and encryption

    • If you select Enterprise SSD (ESSD), you can set a custom performance level. You can select higher performance levels (PLs) for ESSDs with larger storage capacities. For example, you can select PL 2 for an ESSD with a storage capacity of more than 460 GiB. You can select PL 3 for an ESSD with a storage capacity of more than 1,260 GiB. For more information, see Capacity and PLs.

    • You can select Encryption only if you set the system disk type to Enterprise SSD (ESSD). By default, the default service CMK is used to encrypt the system disk. You can also use an existing CMK generated by using Bring Your Own Key (BYOK) in KMS.

    You can select More System Disk Types and select a disk type other than the current one in the System Disk section to improve the success rate of system disk creation. The system will attempt to create a system disk based on the specified disk types in sequence.

    对

    Data Disk

    ESSD AutoPL, Enterprise SSD (ESSD), ESSD Entry, SSD, and Ultra Disk are supported. The data disk types that you can select vary based on the instance families that you select. Disk types that are not displayed in the drop-down list are not supported by the instance types that you select.

    ESSD AutoPL Disk

    • Performance provision: The performance provision feature allows you to configure provisioned performance settings for ESSD AutoPL disks to meet storage requirements that exceed the baseline performance without the need to extend the disks.

    • Performance burst: The performance burst feature allows ESSD AutoPL disks to burst their performance when spikes in read/write workloads occur and reduce the performance to the baseline level at the end of workload spikes.

    Enterprise SSD (ESSD)

    Configure a custom performance level. You can select higher PLs for ESSDs with larger storage capacities. For example, you can select PL 2 for an ESSD with a storage capacity of more than 460 GiB. You can select PL 3 for an ESSD with a storage capacity of more than 1,260 GiB. For more information, see Capacity and PLs.

    • You can select Encryption for all disk types when you specify the type of data disk. By default, the default service CMK is used to encrypt the data disk. You can also use an existing CMK generated by using BYOK in KMS.

    • You can also use snapshots to create data disks in scenarios where container image acceleration and fast loading of large language models (LLMs) are required. This improves the system response speed and enhances the processing capability.

    • Make sure that a data disk is mounted to /var/lib/container on each node, and /var/lib/kubelet and /var/lib/containerd are mounted to the /var/lib/container. For other data disks on the node, you can perform the initialization operation and customize their mount directories. For more information, see Can I mount a data disk to a custom directory in an ACK node pool?

    Note

    You can attach up to 64 data disks to an ECS instance. The maximum number of disks that can be attached to an ECS instance varies based on the instance type. To query the maximum number of disks that you can attach to an ECS instance of a specific instance type, call the DescribeInstanceTypes operation and check the DiskQuantity parameter in the response.

    对

    Instance quantity

    Parameter

    Description

    Modifiable

    Expected Nodes

    The expected number of nodes in the node pool. We recommend that you configure at least two nodes to ensure that cluster components run as expected. You can configure the Expected Nodes parameter to adjust the number of nodes in the node pool. For more information, see Scale a node pool.

    If you do not want to create nodes in the node pool, set this parameter to 0. You can manually modify this parameter to add nodes later.

    对

    Advanced configurations

    Click Advanced Options (Optional) and configure the node scaling policy, resource group, ECS tags, and taints.

    Advanced configurations

    Parameter

    Description

    Modifiable

    Resource Group

    The resource group to which the cluster belongs. Each resource can belong to only one resource group. You can regard a resource group as a project, an application, or an organization based on your business scenarios.

    对

    Scaling Mode

    You must select Auto for the Scaling Mode.
    • Standard mode: Auto scaling is implemented by creating and releasing ECS instances.

    • Swift mode: Auto scaling is implemented by creating, stopping, and starting ECS instances. ECS instances in the Stopped state can be directly restarted to accelerate scaling activities.

      When a node in swift mode is reclaimed, only disk fees are charged for the node. No computing fee is charged. This rule does not apply to instance families that use local disks, such as big data and local SSDs instance families. For more information about the billing rules and limits of the economical mode, see Economical mode.

    错

    Scaling Policy

    • Priority: The system scales the node pool based on the priorities of the vSwitches that you select for the node pool. The vSwitches that you select are displayed in descending order of priority. If Auto Scaling fails to create ECS instances in the zone of the vSwitch with the highest priority, Auto Scaling attempts to create ECS instances in the zone of the vSwitch with a lower priority.

    • Cost Optimization: The system creates instances based on the vCPU unit prices in ascending order.

      If the Billing Method of the node pool is set to Preemptible Instance, preemptible instances are preferentially created. You can also set the Percentage of Pay-as-you-go Instances parameter. If preemptible instances cannot be created due to reasons such as insufficient stocks, pay-as-you-go instances are automatically created as supplement.

    • Distribution Balancing: The even distribution policy takes effect only when you select multiple vSwitches. This policy ensures that ECS instances are evenly distributed among the zones (the vSwitches) of the scaling group. If ECS instances are unevenly distributed across the zones due to reasons such as insufficient stocks, you can perform a rebalancing operation.

    对

    Use Pay-as-you-go Instances When Preemptible Instances Are Insufficient

    You must set the Billing Method parameter to Preemptible Instance.

    After this feature is enabled, if enough preemptible instances cannot be created because of price or inventory constraints, ACK automatically creates pay-as-you-go instances to meet the required number of ECS instances.

    对

    Enable Supplemental Preemptible Instances

    You must set the Billing Method parameter to Preemptible Instance.

    After this feature is enabled, when a system message that indicates preemptible instances are reclaimed is received, the node pool with auto scaling enabled attempts to create new instance to replace the reclaimed the preemptible instances.

    对

    ECS Tags

    Add tags to the ECS instances that are automatically added during auto scaling. Tag keys must be unique. A key cannot exceed 128 characters in length. Keys and values cannot start with aliyun or acs:. Keys and values cannot contain https:// or http://.

    An ECS instance can have up to 20 tags. To increase the quota limit, submit an application in the Quota Center console. The following tags are automatically added to an ECS node by ACK and Auto Scaling. Therefore, you can add at most 17 tags to an ECS node.

    • The following two ECS tags are added by ACK:

      • ack.aliyun.com:<Cluster ID>

      • ack.alibabacloud.com/nodepool-id:<Node pool ID>

    • The following label is added by Auto Scaling: acs:autoscaling:scalingGroupId:<Scaling group ID>.

    Note
    • After you enable auto scaling, the following ECS tags are added to the node pool by default: k8s.io/cluster-autoscaler:true and k8s.aliyun.com:true.

    • The auto scaling component simulates scale-out activities based on node labels and taints. To meet this purpose, the format of node labels is changed to k8s.io/cluster-autoscaler/node-template/label/Label key:Label value and the format of taints is changed to k8s.io/cluster-autoscaler/node-template/taint/Taint key/Taint value:Taint effect.

    对

    Taints

    Add taints to nodes. A taint consists of a key, a value, and an effect. A taint key can be prefixed. If you want to specify a prefixed taint key, add a forward slash (/) between the prefix and the remaining content of the key. For more information, see Taints and tolerations. The following limits apply to taints:

    • Key: A key must be 1 to 63 characters in length, and can contain letters, digits, hyphens (-), underscores (_), and periods (.). A key must start and end with a letter or digit.

      If you want to specify a prefixed key, the prefix must be a subdomain name. A subdomain name consists of DNS labels that are separated by periods (.), and cannot exceed 253 characters in length. It must end with a forward slash (/). For more information about subdomain names, see DNS subdomain names.

    • Value: A value cannot exceed 63 characters in length, and can contain letters, digits, hyphens (-), underscores (_), and periods (.). A value must start and end with a letter or digit. You can also leave a value empty.

    • You can specify the following effects for a taint: NoSchedule, NoExecute, and PreferNoSchedule.

      • NoSchedule: If a node has a taint whose effect is NoSchedule, the system does not schedule pods to the node.

      • NoExecute: Pods that do not tolerate this taint are evicted after this taint is added to a node. Pods that tolerate this taint are not evicted after this taint is added to a node.

      • PreferNoSchedule: The system attempts to avoid scheduling pods to nodes with taints that are not tolerated by the pods.

    对

    Node Labels

    Add labels to nodes. A label is a key-value pair. A label key can be prefixed. If you want to specify a prefixed label key, add a forward slash (/) between the prefix and the remaining content of the key. The following limits apply to labels:

    • Key: The name must be 1 to 63 characters in length, and can contain letters, digits, hyphens (-), underscores (_), and periods (.). It must start and end with a letter or a digit.

      If you want to specify a prefixed label key, the prefix must be a subdomain name. A subdomain name consists of DNS labels that are separated by periods (.), and cannot exceed 253 characters in length. It must end with a forward slash (/).

      The following prefixes are used by key Kubernetes components and cannot be used in node labels:

      • kubernetes.io/

      • k8s.io/

      • Prefixes that end with kubernetes.io/ or k8s.io/. Example: test.kubernetes.io/.

        However, you can still use the following prefixes:

        • kubelet.kubernetes.io/

        • node.kubernetes.io

        • Prefixes that end with kubelet.kubernetes.io/.

        • Prefixes that end with node.kubernetes.io.

    • Value: A value cannot exceed 63 characters in length, and can contain letters, digits, hyphens (-), underscores (_), and periods (.). A value must start and end with a letter or digit. You can also leave a value empty.

    对

    Container Image Acceleration

    Only clusters that use containerd versions 1.6.34 and later supports this configuration. This feature is in canary release.

    After you select this option, new nodes automatically detect whether container images support on-demand loading. If supported, container startups will be accelerated by using on-demand loading.

    对

    CPU Policy

    The CPU management policy for kubelet nodes.

    • None: The default CPU management policy.

    • Static: This policy allows pods with specific resource characteristics on the node to be granted enhanced CPU affinity and exclusivity.

    对

    Custom Node Name

    Specify whether to use a custom node name. If you choose to use a custom node name, the name of the node, name of the ECS instance, and hostname of the ECS instance are changed.

    Note

    If a Windows instance uses a custom node name, the hostname of the instance is fixed to an IP address. You need to use hyphens (-) to replace the periods (.) in the IP address. In addition, no prefix or suffix is allowed in the IP address.

    A custom node name consists of a prefix, an IP substring, and a suffix.

    • A custom node name must be 2 to 64 characters in length. The name must start and end with a lowercase letter or digit.

    • The prefix and suffix can contain letters, digits, hyphens (-), and periods (.). The prefix and suffix must start with a letter and cannot end with a hyphen (-) or period (.). The prefix and suffix cannot contain consecutive hyphens (-) or periods (.).

    • The prefix is required due to ECS limits and the suffix is optional.

    For example, the node IP address is 192.XX.YY.55, the prefix is aliyun.com, and the suffix is test.

    • If the node is a Linux node, the node name, ECS instance name, and ECS instance hostname are aliyun.com192.XX.YY.55test.

    • If the node is a Windows node, the ECS instance hostname is 192-XX-YY-55 and the node name and ECS instance name are aliyun.com192.XX.YY.55test.

    错

    Worker RAM Role

    ACK managed clusters that run Kubernetes 1.22 or later are supported.

    You can assign a worker Resource Access Management (RAM) role to a node pool to reduce the potential risk of sharing a worker RAM role among all nodes in the cluster.

    • Default Role: The node pool uses the default worker RAM role created by the cluster.

    • Custom: The node pool uses the specified role as the worker RAM role. The default role is used when this parameter is left empty. For more information, see Use custom worker RAM roles.

    对

    Pre-defined Custom Data

    To use this feature, submit an application in the Quota Center console.

    Nodes automatically run predefined scripts before they are added to the cluster. For more information about user-data scripts, see User-data scripts.

    For example, if you enter echo "hello world", a node runs the following script:

    #!/bin/bash
    echo "hello world"
    [Node initialization script]

    对

    User Data

    Nodes automatically run user-data scripts after they are added to the cluster. For more information about user-data scripts, see User-data scripts.

    For example, if you enter echo "hello world", a node runs the following script:

    #!/bin/bash
    [Node initialization script]
    echo "hello world"
    Note

    After you create a cluster or add nodes, the execution of the user-data script on a node may fail. We recommend that you log on to a node and run the grep cloud-init/var/log/messages command to view the execution log and check whether the execution succeeds or fails on the node.

    对

    CloudMonitor Agent

    After you install CloudMonitor, you can view the monitoring information about the nodes in the CloudMonitor console.

    This parameter takes effect only on newly added nodes and does not take effect on existing nodes. If you want to install the CloudMonitor agent on an existing ECS node, go to the CloudMonitor console.

    对

    Public IP

    Specify whether to assign an IPv4 address to each node. If you clear the check box, no public IP address is allocated. If you select the check box, you must configure the Bandwidth Billing Method and Peak Bandwidth parameters.

    This parameter takes effect only on newly added nodes and does not take effect on existing nodes. If you want to enable an existing node to access the Internet, you must create an EIP and associate the EIP with the node. For more information, see Associate an EIP with an ECS instance.

    对

    Custom Security Group

    To use custom security groups, you must apply to be added to the whitelist in the Quota Center console.

    You can select Basic Security Group or Advanced Security Group, but you can select only one security group type. You cannot modify the security groups of node pools or change the type of security group. For more information about security groups, see Overview.

    Important
    • Each ECS instance supports up to five security groups. Make sure that the quota of security groups for your ECS instance is sufficient. For more information about security group limits and how to increase the quota limit of security groups for your ECS instance, see Security group limits.

    • If you select an existing security group, the system does not automatically configure security group rules. This may cause errors when you access the nodes in the cluster. You must manually configure security group rules. For more information about how to manage security group rules, see Configure security group rules to enforce access control on ACK clusters.

    错

    RDS Whitelist

    Add node IP addresses to the whitelist of an ApsaraDB RDS instance.

    对

    Deployment Set

    Only users in the whitelist can use this feature. To apply to be added to the whitelist, log on to the Quota Center console and submit an application.

    You must first create a deployment set in the ECS console and then specify the deployment set when you create a node pool in the ACK console. The deployment set cannot be modified after the cluster is created.

    You can use a deployment set to distribute your ECS instances to different physical servers to ensure high service availability and implement underlying disaster recovery. If you specify a deployment set when you create ECS instances, the instances are created and distributed based on the deployment strategy that you preset for the deployment set within the specified region. For more information, see Best practices for associating deployment sets with node pools.

    Important

    After you select a deployment set, the maximum number of nodes that can be created in the node pool is limited. By default, the maximum number of nodes supported by a deployment set is 20 × Number of zones. The number of zones depends on the number of vSwitches. Exercise caution when you select the deployment set. To avoid node creation failures, make sure that the ECS quota of the deployment set that you select is sufficient.

    错

    Private Pool Type

    Valid values: Open, Do Not Use, and Specified.

    • Open: The system automatically matches an open private pool. If no matching is found, resources in the public pool are used.

    • Do Not Use: No private pool is used. Only resources in the public pool are used.

    • Specified: Specify a private pool by ID. If the specified private pool is unavailable, ECS instances fail to start up.

    For more information, see Private pools.

    对

  2. Click Confirm Order.

    In the node pool list, if the Status column of the node pool displays Initializing, the node pool is being created. After the node pool is created, the Status column of the node pool displays Active.

Modify a node pool

After you create a node pool, you can modify the configurations of the node pool in the ACK console. For example, you can change the billing method, vSwitches, instance specifications, and system disks that are used by the node pool. You can also enable or disable auto scaling for the node pool. For more information about the modifiable parameters, see Create a node pool.

Important
  • Modifying a node pool does not affect the nodes and applications deployed in other node pools of the cluster.

  • In most scenarios, after you modify a node pool, the modified configurations apply only to newly added nodes. In specific scenarios, such as when you update the ECS tags or labels and taints of existing nodes, the modified configurations also apply to existing nodes in the node pool.

  • After you update the configurations of a node pool, nodes that are subsequently added to the node pool use the modified configurations.

  • To modify the node pool configuration, refer to the following steps. If you have made changes to the nodes through other methods, these changes will be overwritten when the node pool is updated.

  1. On the Node Pools page, find the node pool that you want to modify and click Edit in the Actions column.

  2. In the dialog box that appears, modify the parameters of the node pool based on the on-screen instructions.

    On the Node Pools page, if the Status column of the node pool displays Updatng, the node pool is being modified. After the node pool is updated, the Status column displays Activated.

View a node pool

You can view the basic information, monitoring data, node information, and scaling events of a node pool in the ACK console.

Click the name of the node pool that you want to manage to view the following information on the details page of the node pool:

  • Click the Overview tab to view the cluster information, node pool information, and node configurations. If the cluster has auto scaling enabled, you can also view the auto scaling configurations.

  • Click the Monitor tab to view the node monitoring information provided by Managed Service for Prometheus. The monitoring information includes the resource watermarks in the node pool, such as CPU usage, memory usage, disk usage, and average CPU or memory utilization per node.

  • Click the Nodes tab to view the list of nodes in the node pool. You can drain a node, configure the scheduling settings of a node, or perform O&M operations on a node. You can also remove a node from the node pool. You can click Export to export the details of the nodes to a comma-separated values (CSV) file.

  • Click the Scaling Activities tab to view the latest scaling events of the node pool. Each event record provides a description of the scaling activity and the number of ECS instances after the scaling activity is performed. You can also view the reasons for scaling failures. For more information about the common error codes for scaling failures, see Manually scale a node pool.

Delete a node pool

The release rules of an ECS instance vary based on the billing method of the instance. When you remove a node from a node pool, we recommend that you perform the operations described in the following table. Before you delete a node pool, check whether the Expected Nodes parameter is configured for the node pool. This parameter may affect the node release process.

Node pool

Release rule

Node pool that is configured with the Expected Nodes parameter

  • Pay-as-you-go nodes: released after the node pool is deleted.

  • Subscription nodes: retained after the node pool is deleted.

    If you want to release a subscription node, refer to Change the billing method of an instance from subscription to pay-as-you-go to change the billing method of the instance to pay-as-you-go and then log on to the ECS console and release the instance.

  • After you delete a node pool, all nodes in the node pool are removed from the API server of the cluster.

Node pool that is not configured with the Expected Nodes parameter

  • Nodes that are manually or automatically added to the node pool and subscription nodes are not released. Other nodes are released when you delete the node pool.

    If you want to release a subscription node, refer to Change the billing method of an instance from subscription to pay-as-you-go to change the billing method of the instance to pay-as-you-go and then log on to the ECS console and release the instance.

  • Released nodes are removed from the API server of the cluster. Retained nodes are not removed from the API server of the cluster.

  1. Optional: Click the name of the node pool that you want to manage. On the Overview tab, you can check whether the Expected Nodes parameter is configured. If a hyphen (-) is displayed, the Expected Nodes parameter is not configured.

  2. Select the node pool that you want to delete and choose More > Delete in the Actions. Read and confirm the information in the dialog box and then click OK.

What to do next

After the node pool is created, you can click an action in the Actions column of the node pool or click More in the Actions column to perform the following operations.

Action

Description

References

Sync Node Pool

If the node information is abnormal, you can synchronize the node pool.

None

Details

View the details of the node pool.

None

Edit

Modify the configurations of the node pool. For example, you can modify the vSwitches, managed node pool settings, billing method, and instance specifications. You can also enable or disable auto scaling for the node pool.

Modify a node pool

Monitor

View the node monitoring information provided by Managed Service for Prometheus. The monitoring information includes the resource watermarks in the node pool, such as CPU usage, memory usage, disk usage, and average CPU or memory utilization per node.

View a node pool

Scale

Manual and Auto scalings are supported. Computing resources are automatically adjusted based on your business requirements and policies to reduce cluster costs.

  • Manual: ACK adjusts the number of nodes in the node pool based on the value of the Expected Nodes parameter. The number of nodes is always the same as the value of the Expected Nodes parameter. For more information, see Manually scale a node pool.

  • Auto: When the capacity planning of the cluster cannot meet the requirements of pod scheduling, ACK automatically scales out nodes based on the configured minimum and maximum number of instances. By default, node instant scaling is enabled for clusters that run Kubernetes 1.24 and later. By default, node auto scaling is enabled for clusters that run Kubernetes 1.24 and later. For more information, see Overview of node scaling.

Logon Mode

Configure the logon type of nodes. You can specify a key pair or password.

Instance and Image

Enable Managed Node Pool

Automate node O&M for the node pool. After you enable the managed node pool feature, O&M tasks such as node repair, kubelet updates, runtime updates, and OS Common Vulnerabilities and Exposures (CVE) patching are automatically performed.

Basic configurations

Add Existing Node

Add an existing ECS instance to the cluster as a worker node. You can perform this operation to add a worker node that you have previously removed from the cluster. Specific limits and usage notes apply to this operation. Refer to the ACK documentation for details.

Add existing ECS instances to an ACK cluster

Clone

Clone a node pool that contains the expected number of nodes based on the current node pool configurations.

None

Node Repair

When exceptions occur on a node in a managed node pool, ACK automatically repairs the node. However, you may still need to manually fix some complex node exceptions. For more information about the check items of node status and repair solutions for node exceptions, refer to the relevant topics in the ACK documentation.

Enable auto repair for nodes

CVE Patching (OS)

Patch CVE vulnerabilities in the node pool in batches to improve the stability, security, and compliance of the cluster. ACK may need to restart nodes to patch specific vulnerabilities. For more information about CVE patching and the usage notes for CVE patching, refer to the relevant topics in the ACK documentation.

Patch OS CVE vulnerabilities for node pools

Kubelet Update

Customize the kubelet parameters for all nodes in the node pool to manage the behavior of the nodes. For example, you can customize the kubelet parameters if you need to modify resource reservations to adjust the resource usage. We recommend that you do not use the CLI to customize kubelet parameters that are unavailable in the ACK console.

Customize the kubelet parameters of a node pool

OS Configuration

Customize the OS parameters of all nodes in the node pool to improve OS performance. We recommend that you do not use the CLI to customize OS parameters that are unavailable in the ACK console.

Customize the OS parameters of a node pool

Change Operating System

Change or update the operating system of the node pool.

None

Kubelet Update

Update the kubelet and containerd for all nodes in the node pool.

Update a node pool

Delete

Delete the node pool if the node pool is no longer in use to save costs. The release rules of nodes in a node pool vary based on the billing method of the nodes and whether the Expected Nodes parameter is configured for the node pool.

Delete a node pool

FAQ

How do I create a custom image from an ECS instance and use the image to create a node?

After you create an ECS instance, you can customize the instance by performing operations such as installing software and deploying application environments. Then, you can create a custom image from the instance. Instances created from the custom image contain all of the customized items, which eliminates the need to configure these items for each new instance.

  1. Log on to the ECS instance and run the following command to delete the specified files. For more information about how to log on to an ECS instance, see Use Workbench to connect to a Linux instance over SSH.

    chattr -i /etc/acknode/nodeconfig-*
    rm -rf /etc/acknode
    systemctl disable ack-reconfig
    rm -rf /etc/systemd/system/ack-reconfig.service
    rm -rf /usr/local/bin/reconfig.sh
    rm -rf /var/lib/cloud
  2. Create a custom image from the ECS instance. For more information about the procedure and usage notes, see Create a custom image from an instance.

  3. Create a node pool. When you create the node pool, select Custom Image for the Operating System parameter. Configure other parameters based on the description previously mentioned in this topic.

Important
  • Create a custom image based on the operating system supported by the ACK cluster. For more information, see OS images.

  • Do not build custom images on running ECS instances in an ACK cluster. To do this, you must first remove the ECS instances from the cluster. For more information, see Remove a node.

  • The predefined behavior logic in a custom image may affect operations such as cluster node initialization, container launching, node updates, and automatic recovery of nodes in a managed node pool. Before you use it in a production environment, ensure that the custom image has been tested and validated.

References

  • If a node is no longer in use, you can remove the node. For more information, see Remove nodes.

  • ACK reserves a certain amount of node resources to run Kubernetes components and system processes. For more information, see Resource reservation policy.

  • When the resource capacity of the cluster cannot meet the requirements for pod scheduling, you can enable node auto scaling. For more information, see Overview of node scaling.

  • The maximum number of pods on a worker node varies based on the network plug-in and cannot be adjusted in most cases. To increase the maximum number of pods in a cluster, you can scale out the node pools in the cluster, upgrade the instance specifications used by the cluster, and reset the pod CIDR block. For more information, see Increase the maximum number of pods in a cluster.