Usage Instructions for Logical Queues and Physical Queues

CCE CCE

  • Function Release Records
  • Common Tools
    • Command Line Scenario Examples
  • API Reference
    • Overview
    • Common Headers and Error Responses
    • General Description
  • Product Announcement
    • Announcement on the Discontinuation of CCE Standalone Clusters
    • CCE New Cluster Management Release Announcement
    • Upgrade Announcement for CCE Cluster Audit Component kube-external-auditor
    • CCE Console Upgrade Announcement
    • Announcement on Management Fees for CCE Managed Clusters
    • Container Runtime Version Release Notes
    • Announcement on the Decommissioning of CCE Image Repository
    • Kubernetes Version Release Notes
      • CCE Release of Kubernetes v1_26 History
      • CCE Kubernetes Version Update Notes
      • CCE Release of Kubernetes v1_24 History
      • CCE Release of Kubernetes v1_30 History
      • CCE Release of Kubernetes v1_22 History
      • CCE Release of Kubernetes v1_18 History
      • CCE Release of Kubernetes v1_20 History
      • CCE Release of Kubernetes v1_28 History
      • Release Notes for CCE Kubernetes 1_31 Version
      • Kubernetes Version Overview and Mechanism
    • Security Vulnerability Fix Announcement
      • Vulnerability CVE-2019-5736 Fix Announcement
      • Vulnerability CVE-2021-30465 Fix Announcement
      • CVE-2025-1097, CVE-2025-1098, and Other Vulnerabilities Fix Announcement
      • CVE-2020-14386 Vulnerability Fix Announcement
      • Impact Statement on runc Security Issue (CVE-2024-21626)
  • Service Level Agreement (SLA)
    • CCE Service Level Agreement SLA (V1_0)
  • Typical Practices
    • Pod Anomaly Troubleshooting
    • Adding CGroup V2 Node
    • Common Linux System Configuration Parameters Description
    • Encrypting etcd Data Using KMS
    • Configuring Container Network Parameters Using CNI
    • CCE - Public Network Access Practice
    • Practice of using private images in CCE clusters
    • Unified Access for Virtual Machines and Container Services via CCE Ingress
    • User Guide for Custom CNI Plugins
    • CCE Cluster Network Description and Planning
    • Cross-Cloud Application Migration to Baidu CCE Using Velero
    • CCE Resource Recommender User Documentation
    • Continuous Deployment with Jenkins in CCE Cluster
    • CCE Best Practice-Guestbook Setup
    • CCE Best Practice-Container Network Mode Selection
    • CCE Usage Checklist
    • VPC-ENI Mode Cluster Public Network Access Practice
    • CCE Container Runtime Selection
    • Cloud-native AI
      • Elastic and Fault-Tolerant Training Using CCE AITraining Operator
      • Deploy the TensorFlow Serving inference service
      • Best Practice for GPU Virtualization with Optimal Isolation
  • FAQs
    • How do business applications use load balancer
    • Using kubectl on Windows
    • Cluster management FAQs
    • Common Questions Overview
    • Auto scaling FAQs
    • Create a simple service via kubectl
  • Operation guide
    • Prerequisites for use
    • Identity and access management
    • Permission Management
      • Configure IAM Tag Permission Policy
      • Permission Overview
      • Configure IAM Custom Permission Policy
      • Configure Predefined RBAC Permission Policy
      • Configure IAM Predefined Permission Policy
      • Configure Cluster OIDC Authentication
    • Configuration Management
      • Configmap Management
      • Secret Management
    • Traffic access
      • BLB ingress annotation description
      • Use K8S_Service via CCE
      • Use K8S_Ingress via CCE
      • Implement Canary Release with CCE Based on Nginx-Ingress
      • Create CCE_Ingress via YAML
      • LoadBalancer Service Annotation Description
      • Service Reuses Existing Load Balancer BLB
      • Use Direct Pod Mode LoadBalancer Service
      • NGINX Ingress Configuration Reference
      • Create LoadBalancer_Service via YAML
      • Use NGINX Ingress
    • Virtual Node
      • Configuring BCIPod
      • Configuring bci-profile
      • Managing virtual nodes
    • Node management
      • Add a node
      • Managing Taints
      • Setting Node Blocking
      • Setting GPU Memory Sharing
      • Remove a node
      • Customizing Kubelet Parameters
      • Kubelet Container Monitor Read-Only Port Risk Warning
      • Managing Node Tag
      • Drain node
    • Component Management
      • CCE CSI CDS Plugin Description
      • CCE Fluid Description
      • CCE CSI PFS L2 Plugin
      • CCE Calico Felix Description
      • CCE Ingress Controller Description
      • CCE QoS Agent Description
      • CCE GPU Manager Description
      • CCE Ingress NGINX Controller Description
      • CCE P2P Accelerator Description
      • CCE Virtual Kubelet Component
      • CoreDNS Description
      • CCE Log Operator Description
      • CCE Node Remedier Description
      • CCE Descheduler Description
      • CCE Dynamic Scheduling Plugin Description
      • Kube Scheduler Documentation
      • CCE NPU Manager Description
      • CCE CronHPA Controller Description
      • CCE LB Controller Description
      • Kube ApiServer Description
      • CCE Backup Controller Description
      • CCE Network Plugin Description
      • CCE CSI PFS Plugin Description
      • CCE Credential Controller Description
      • CCE Deep Learning Frameworks Operator Description
      • Component Overview
      • CCE Image Accelerate Description
      • CCE CSI BOS Plugin Description
      • CCE Onepilot Description
      • Description of Kube Controller Manager
      • CCE_Hybrid_Manager Description
      • CCE NodeLocal DNSCache Description
      • CCE Node Problem Detector Description
      • CCE Ascend Mindx DL Description
      • CCE RDMA Device Plugin Description
      • CCE AI Job Scheduler Description
    • Image registry
      • Image Registry Basic Operations
      • Using Container Image to Build Services
    • Helm Management
      • Helm Template
      • Helm Instance
    • Cluster management
      • Upgrade Cluster Kubernetes Version
      • CCE Node CDS Dilatation
      • Managed Cluster Usage Instructions
      • Create cluster
      • CCE Supports GPUSharing Cluster
      • View Cluster
      • Connect to Cluster via kubectl
      • CCE Security Group
      • CCE Node Resource Reservation Instructions
      • Operate Cluster
      • Cluster Snapshot
    • Serverless Cluster
      • Product overview
      • Using Service in Serverless Cluster
      • Creating a Serverless Cluster
    • Storage Management
      • Using Cloud File System
      • Overview
      • Using Parallel File System PFS
      • Using RapidFS
      • Using Object Storage BOS
      • Using Parallel File System PFS L2
      • Using Local Storage
      • Using Cloud Disk CDS
    • Inspection and Diagnosis
      • Cluster Inspection
      • GPU Runtime Environment Check
      • Fault Diagnosis
    • Cloud-native AI
      • Cloud-Native AI Overview
      • AI Monitoring Dashboard
        • Connecting to a Prometheus Instance and Starting a Job
        • NVIDIA Chip Resource Observation
          • AI Job Scheduler component
          • GPU node resources
          • GPU workload resources
          • GPUManager component
          • GPU resource pool overview
        • Ascend Chip Resource Observation
          • Ascend resource pool overview
          • Ascend node resource
          • Ascend workload resource
      • Task Management
        • View Task Information
        • Create TensorFlow Task
        • Example of RDMA Distributed Training Based on NCCL
        • Create PaddlePaddle Task
        • Create AI Training Task
        • Delete task
        • Create PyTorch Task
        • Create Mxnet Task
      • Queue Management
        • Modify Queue
        • Create Queue
        • Usage Instructions for Logical Queues and Physical Queues
        • Queue deletion
      • Dataset Management
        • Create Dataset
        • Delete dataset
        • View Dataset
        • Operate Dataset
      • AI Acceleration Kit
        • AIAK Introduction
        • Using AIAK-Training PyTorch Edition
        • Deploying Distributed Training Tasks Using AIAK-Training
        • Accelerating Inference Business Using AIAK-Inference
      • GPU Virtualization
        • GPU Exclusive and Shared Usage Instructions
        • Image Build Precautions in Shared GPU Scenarios
        • Instructions for Multi-GPU Usage in Single-GPU Containers
        • GPU Virtualization Adaptation Table
        • GPU Online and Offline Mixed Usage Instructions
        • MPS Best Practices & Precautions
        • Precautions for Disabling Node Video Memory Sharing
    • Elastic Scaling
      • Container Timing Horizontal Scaling (CronHPA)
      • Container Horizontal Scaling (HPA)
      • Implementing Second-Level Elastic Scaling with cce-autoscaling-placeholder
      • CCE Cluster Node Auto-Scaling
    • Network Management
      • How to Continue Dilatation When Container Network Segment Space Is Exhausted (VPC-ENI Mode)
      • Container Access to External Services in CCE Clusters
      • CCE supports dual-stack networks of IPv4 and IPv6
      • Using NetworkPolicy Network Policy
      • Traffic Forwarding Configuration for Containers in Peering Connections Scenarios
      • CCE IP Masquerade Agent User Guide
      • Creating VPC-ENI Mode Cluster
      • How to Continue Dilatation When Container Network Segment Space Is Exhausted (VPC Network Mode)
      • Using NetworkPolicy in CCE Clusters
      • Network Orchestration
        • Container Network QoS Management
        • VPC-ENI Specified Subnet IP Allocation (Container Network v2)
        • Cluster Pod Subnet Topology Distribution (Container Network v2)
      • Network Connectivity
        • Container network accesses the public network via NAT gateway
      • Network Maintenance
        • Common Error Code Table for CCE Container Network
      • DNS
        • CoreDNS Component Manual Dilatation Guide
        • DNS Troubleshooting Guide
        • DNS Principle Overview
    • Namespace Management
      • Set Limit Range
      • Set Resource Quota
      • Basic Namespace Operations
    • Workload
      • CronJob Management
      • Set Workload Auto-Scaling
      • Deployment Management
      • Job Management
      • View the Pod
      • StatefulSet Management
      • Password-Free Pull of Container Image
      • Create Workload Using Private Image
      • DaemonSet Management
    • Monitor Logs
      • Monitor Cluster with Prometheus
      • CCE Event Center
      • Cluster Service Profiling
      • CCE Cluster Anomaly Event Alerts
      • Java Application Monitor
      • Cluster Audit Dashboard
      • Logging
      • Cluster Audit
      • Log Center
        • Configure Collection Rules Using CRD
        • View Cluster Control Plane Logs
        • View Business Logs
        • Log Overview
        • Configure Collection Rules in Cloud Container Engine Console
    • Application management
      • Overview
      • Secret
      • Configuration dictionary
      • Deployment
      • Service
      • Pod
    • NodeGroup Management
      • NodeGroup Management
      • NodeGroup Node Fault Detection and Self-Healing
      • Configuring Scaling Policies
      • NodeGroup Introduction
      • Adding Existing External Nodes
      • Custom NodeGroup Kubelet Configuration
      • Adding Alternative Models
      • Dilatation NodeGroup
    • Backup Center
      • Restore Management
      • Backup Overview
      • Backup Management
      • Backup repository
  • Quick Start
    • Quick Deployment of Nginx Application
    • CCE Container Engine Usage Process Overview
  • Product pricing
    • Product pricing
  • Product Description
    • Application scenarios
    • Introduction
    • Usage restrictions
    • Features
    • Advantages
    • Core concepts
  • Solution-Fabric
    • Fabric Solution
  • Development Guide
    • EFK Log Collection System Deployment Guide
    • Using Network Policy in CCE Cluster
    • Creating a LoadBalancer-Type Service
    • Prometheus Monitoring System Deployment Guide
    • kubectl Management Configuration
  • API_V2 Reference
    • Overview
    • Common Headers and Error Responses
    • Cluster Related Interfaces
    • Instance Related Interfaces
    • Service domain
    • General Description
    • Kubeconfig Related Interfaces
    • RBAC Related Interfaces
    • Autoscaler Related Interfaces
    • Network Related Interfaces
    • InstanceGroup Related Interfaces
    • Appendix
    • Component management-related APIs
    • Package adaptation-related APIs
    • Task Related Interfaces
  • Solution-Xchain
    • Hyperchain Solution
  • SDK
    • Go-SDK
      • Overview
      • NodeGroup Management
      • Initialization
      • Install the SDK Package
      • Cluster management
      • Node management
All documents
menu
No results found, please re-enter

CCE CCE

  • Function Release Records
  • Common Tools
    • Command Line Scenario Examples
  • API Reference
    • Overview
    • Common Headers and Error Responses
    • General Description
  • Product Announcement
    • Announcement on the Discontinuation of CCE Standalone Clusters
    • CCE New Cluster Management Release Announcement
    • Upgrade Announcement for CCE Cluster Audit Component kube-external-auditor
    • CCE Console Upgrade Announcement
    • Announcement on Management Fees for CCE Managed Clusters
    • Container Runtime Version Release Notes
    • Announcement on the Decommissioning of CCE Image Repository
    • Kubernetes Version Release Notes
      • CCE Release of Kubernetes v1_26 History
      • CCE Kubernetes Version Update Notes
      • CCE Release of Kubernetes v1_24 History
      • CCE Release of Kubernetes v1_30 History
      • CCE Release of Kubernetes v1_22 History
      • CCE Release of Kubernetes v1_18 History
      • CCE Release of Kubernetes v1_20 History
      • CCE Release of Kubernetes v1_28 History
      • Release Notes for CCE Kubernetes 1_31 Version
      • Kubernetes Version Overview and Mechanism
    • Security Vulnerability Fix Announcement
      • Vulnerability CVE-2019-5736 Fix Announcement
      • Vulnerability CVE-2021-30465 Fix Announcement
      • CVE-2025-1097, CVE-2025-1098, and Other Vulnerabilities Fix Announcement
      • CVE-2020-14386 Vulnerability Fix Announcement
      • Impact Statement on runc Security Issue (CVE-2024-21626)
  • Service Level Agreement (SLA)
    • CCE Service Level Agreement SLA (V1_0)
  • Typical Practices
    • Pod Anomaly Troubleshooting
    • Adding CGroup V2 Node
    • Common Linux System Configuration Parameters Description
    • Encrypting etcd Data Using KMS
    • Configuring Container Network Parameters Using CNI
    • CCE - Public Network Access Practice
    • Practice of using private images in CCE clusters
    • Unified Access for Virtual Machines and Container Services via CCE Ingress
    • User Guide for Custom CNI Plugins
    • CCE Cluster Network Description and Planning
    • Cross-Cloud Application Migration to Baidu CCE Using Velero
    • CCE Resource Recommender User Documentation
    • Continuous Deployment with Jenkins in CCE Cluster
    • CCE Best Practice-Guestbook Setup
    • CCE Best Practice-Container Network Mode Selection
    • CCE Usage Checklist
    • VPC-ENI Mode Cluster Public Network Access Practice
    • CCE Container Runtime Selection
    • Cloud-native AI
      • Elastic and Fault-Tolerant Training Using CCE AITraining Operator
      • Deploy the TensorFlow Serving inference service
      • Best Practice for GPU Virtualization with Optimal Isolation
  • FAQs
    • How do business applications use load balancer
    • Using kubectl on Windows
    • Cluster management FAQs
    • Common Questions Overview
    • Auto scaling FAQs
    • Create a simple service via kubectl
  • Operation guide
    • Prerequisites for use
    • Identity and access management
    • Permission Management
      • Configure IAM Tag Permission Policy
      • Permission Overview
      • Configure IAM Custom Permission Policy
      • Configure Predefined RBAC Permission Policy
      • Configure IAM Predefined Permission Policy
      • Configure Cluster OIDC Authentication
    • Configuration Management
      • Configmap Management
      • Secret Management
    • Traffic access
      • BLB ingress annotation description
      • Use K8S_Service via CCE
      • Use K8S_Ingress via CCE
      • Implement Canary Release with CCE Based on Nginx-Ingress
      • Create CCE_Ingress via YAML
      • LoadBalancer Service Annotation Description
      • Service Reuses Existing Load Balancer BLB
      • Use Direct Pod Mode LoadBalancer Service
      • NGINX Ingress Configuration Reference
      • Create LoadBalancer_Service via YAML
      • Use NGINX Ingress
    • Virtual Node
      • Configuring BCIPod
      • Configuring bci-profile
      • Managing virtual nodes
    • Node management
      • Add a node
      • Managing Taints
      • Setting Node Blocking
      • Setting GPU Memory Sharing
      • Remove a node
      • Customizing Kubelet Parameters
      • Kubelet Container Monitor Read-Only Port Risk Warning
      • Managing Node Tag
      • Drain node
    • Component Management
      • CCE CSI CDS Plugin Description
      • CCE Fluid Description
      • CCE CSI PFS L2 Plugin
      • CCE Calico Felix Description
      • CCE Ingress Controller Description
      • CCE QoS Agent Description
      • CCE GPU Manager Description
      • CCE Ingress NGINX Controller Description
      • CCE P2P Accelerator Description
      • CCE Virtual Kubelet Component
      • CoreDNS Description
      • CCE Log Operator Description
      • CCE Node Remedier Description
      • CCE Descheduler Description
      • CCE Dynamic Scheduling Plugin Description
      • Kube Scheduler Documentation
      • CCE NPU Manager Description
      • CCE CronHPA Controller Description
      • CCE LB Controller Description
      • Kube ApiServer Description
      • CCE Backup Controller Description
      • CCE Network Plugin Description
      • CCE CSI PFS Plugin Description
      • CCE Credential Controller Description
      • CCE Deep Learning Frameworks Operator Description
      • Component Overview
      • CCE Image Accelerate Description
      • CCE CSI BOS Plugin Description
      • CCE Onepilot Description
      • Description of Kube Controller Manager
      • CCE_Hybrid_Manager Description
      • CCE NodeLocal DNSCache Description
      • CCE Node Problem Detector Description
      • CCE Ascend Mindx DL Description
      • CCE RDMA Device Plugin Description
      • CCE AI Job Scheduler Description
    • Image registry
      • Image Registry Basic Operations
      • Using Container Image to Build Services
    • Helm Management
      • Helm Template
      • Helm Instance
    • Cluster management
      • Upgrade Cluster Kubernetes Version
      • CCE Node CDS Dilatation
      • Managed Cluster Usage Instructions
      • Create cluster
      • CCE Supports GPUSharing Cluster
      • View Cluster
      • Connect to Cluster via kubectl
      • CCE Security Group
      • CCE Node Resource Reservation Instructions
      • Operate Cluster
      • Cluster Snapshot
    • Serverless Cluster
      • Product overview
      • Using Service in Serverless Cluster
      • Creating a Serverless Cluster
    • Storage Management
      • Using Cloud File System
      • Overview
      • Using Parallel File System PFS
      • Using RapidFS
      • Using Object Storage BOS
      • Using Parallel File System PFS L2
      • Using Local Storage
      • Using Cloud Disk CDS
    • Inspection and Diagnosis
      • Cluster Inspection
      • GPU Runtime Environment Check
      • Fault Diagnosis
    • Cloud-native AI
      • Cloud-Native AI Overview
      • AI Monitoring Dashboard
        • Connecting to a Prometheus Instance and Starting a Job
        • NVIDIA Chip Resource Observation
          • AI Job Scheduler component
          • GPU node resources
          • GPU workload resources
          • GPUManager component
          • GPU resource pool overview
        • Ascend Chip Resource Observation
          • Ascend resource pool overview
          • Ascend node resource
          • Ascend workload resource
      • Task Management
        • View Task Information
        • Create TensorFlow Task
        • Example of RDMA Distributed Training Based on NCCL
        • Create PaddlePaddle Task
        • Create AI Training Task
        • Delete task
        • Create PyTorch Task
        • Create Mxnet Task
      • Queue Management
        • Modify Queue
        • Create Queue
        • Usage Instructions for Logical Queues and Physical Queues
        • Queue deletion
      • Dataset Management
        • Create Dataset
        • Delete dataset
        • View Dataset
        • Operate Dataset
      • AI Acceleration Kit
        • AIAK Introduction
        • Using AIAK-Training PyTorch Edition
        • Deploying Distributed Training Tasks Using AIAK-Training
        • Accelerating Inference Business Using AIAK-Inference
      • GPU Virtualization
        • GPU Exclusive and Shared Usage Instructions
        • Image Build Precautions in Shared GPU Scenarios
        • Instructions for Multi-GPU Usage in Single-GPU Containers
        • GPU Virtualization Adaptation Table
        • GPU Online and Offline Mixed Usage Instructions
        • MPS Best Practices & Precautions
        • Precautions for Disabling Node Video Memory Sharing
    • Elastic Scaling
      • Container Timing Horizontal Scaling (CronHPA)
      • Container Horizontal Scaling (HPA)
      • Implementing Second-Level Elastic Scaling with cce-autoscaling-placeholder
      • CCE Cluster Node Auto-Scaling
    • Network Management
      • How to Continue Dilatation When Container Network Segment Space Is Exhausted (VPC-ENI Mode)
      • Container Access to External Services in CCE Clusters
      • CCE supports dual-stack networks of IPv4 and IPv6
      • Using NetworkPolicy Network Policy
      • Traffic Forwarding Configuration for Containers in Peering Connections Scenarios
      • CCE IP Masquerade Agent User Guide
      • Creating VPC-ENI Mode Cluster
      • How to Continue Dilatation When Container Network Segment Space Is Exhausted (VPC Network Mode)
      • Using NetworkPolicy in CCE Clusters
      • Network Orchestration
        • Container Network QoS Management
        • VPC-ENI Specified Subnet IP Allocation (Container Network v2)
        • Cluster Pod Subnet Topology Distribution (Container Network v2)
      • Network Connectivity
        • Container network accesses the public network via NAT gateway
      • Network Maintenance
        • Common Error Code Table for CCE Container Network
      • DNS
        • CoreDNS Component Manual Dilatation Guide
        • DNS Troubleshooting Guide
        • DNS Principle Overview
    • Namespace Management
      • Set Limit Range
      • Set Resource Quota
      • Basic Namespace Operations
    • Workload
      • CronJob Management
      • Set Workload Auto-Scaling
      • Deployment Management
      • Job Management
      • View the Pod
      • StatefulSet Management
      • Password-Free Pull of Container Image
      • Create Workload Using Private Image
      • DaemonSet Management
    • Monitor Logs
      • Monitor Cluster with Prometheus
      • CCE Event Center
      • Cluster Service Profiling
      • CCE Cluster Anomaly Event Alerts
      • Java Application Monitor
      • Cluster Audit Dashboard
      • Logging
      • Cluster Audit
      • Log Center
        • Configure Collection Rules Using CRD
        • View Cluster Control Plane Logs
        • View Business Logs
        • Log Overview
        • Configure Collection Rules in Cloud Container Engine Console
    • Application management
      • Overview
      • Secret
      • Configuration dictionary
      • Deployment
      • Service
      • Pod
    • NodeGroup Management
      • NodeGroup Management
      • NodeGroup Node Fault Detection and Self-Healing
      • Configuring Scaling Policies
      • NodeGroup Introduction
      • Adding Existing External Nodes
      • Custom NodeGroup Kubelet Configuration
      • Adding Alternative Models
      • Dilatation NodeGroup
    • Backup Center
      • Restore Management
      • Backup Overview
      • Backup Management
      • Backup repository
  • Quick Start
    • Quick Deployment of Nginx Application
    • CCE Container Engine Usage Process Overview
  • Product pricing
    • Product pricing
  • Product Description
    • Application scenarios
    • Introduction
    • Usage restrictions
    • Features
    • Advantages
    • Core concepts
  • Solution-Fabric
    • Fabric Solution
  • Development Guide
    • EFK Log Collection System Deployment Guide
    • Using Network Policy in CCE Cluster
    • Creating a LoadBalancer-Type Service
    • Prometheus Monitoring System Deployment Guide
    • kubectl Management Configuration
  • API_V2 Reference
    • Overview
    • Common Headers and Error Responses
    • Cluster Related Interfaces
    • Instance Related Interfaces
    • Service domain
    • General Description
    • Kubeconfig Related Interfaces
    • RBAC Related Interfaces
    • Autoscaler Related Interfaces
    • Network Related Interfaces
    • InstanceGroup Related Interfaces
    • Appendix
    • Component management-related APIs
    • Package adaptation-related APIs
    • Task Related Interfaces
  • Solution-Xchain
    • Hyperchain Solution
  • SDK
    • Go-SDK
      • Overview
      • NodeGroup Management
      • Initialization
      • Install the SDK Package
      • Cluster management
      • Node management
  • Document center
  • arrow
  • CCECCE
  • arrow
  • Operation guide
  • arrow
  • Cloud-native AI
  • arrow
  • Queue Management
  • arrow
  • Usage Instructions for Logical Queues and Physical Queues
Table of contents on this page
  • Prerequisites
  • Logical queue
  • Ordinary queue
  • Elastic queue
  • Hierarchical queues
  • Physical queue
  • Instructions for quota setting
  • Queue parameter description
  • Queue notes
  • Create a queue
  • Queue deletion
  • Queue update
  • Logical queue usage
  • Create elastic queues and hierarchical queues
  • Update logical queue
  • Submit tasks to logical queues
  • Physical queue usage
  • Node partitioning
  • Create physical queue
  • Create logical queues under a queue (optional)
  • Modify physical queue

Usage Instructions for Logical Queues and Physical Queues

Updated at:2025-10-27

overview of queue functions

Queues, as subsets of resource pools, are used to handle workloads such as training tasks and model services. Users can partition resource pools into multiple independent queues (logical or physical queues) to execute different workloads. A default queue is automatically created once a resource pool is successfully set up. This document provides guidance on creating, updating, and utilizing queues through YAML templates.

Prerequisites

On the CCE Component Management page, navigate to the Cloud Native AI section and upgrade the CCE AI Job Scheduler component to version 1.7.14 or above to access queue-related features, as shown in the figure below.

image.png

Logical queue

Ordinary queue

Standard queues are the most common way to allocate resource quotas among various application workloads. They assign fixed-quantity quota resources to different departments to enable basic resource allocation functionalities.

Elastic queue

Elastic queues are designed to help platform-based customers allocate resource quotas more dynamically and efficiently across different application departments. These queues enable borrowing idle resource quotas from one queue to another, supporting features such as resource lending/reclaiming and task oversell/preemption, thus enhancing GPU/NPU resource utilization. Detailed task oversell and resource preemption strategies can be found in the appendix.

Hierarchical queues

Hierarchical queues cater to medium and large-scale customers with complex organizational frameworks or scenarios requiring multi-level resource quota management. They enable resource quota management across multiple levels and queues.

Physical queue

Utilize physical queues to ensure GPU/NPU resources are physically isolated, guaranteeing resource availability for critical applications while preventing resource preemption. Additionally, multiple sub-logical queues can be created within a physical queue.

Description of queue functions

Instructions for quota setting

For elastic queues, configure three types of quotas: guaranteed quota, deserved quota, and capability quota. This classification divides computing resources into exclusive, shared, and preemptive categories, optimizing costs and enhancing GPU/NPU resource utilization.

  • Guaranteed quota: The portion of resources exclusively reserved for the queue, which will never be borrowed by other queues.
  • Deserved quota: The portion of resources requested by the queue based on estimated confirmed usage. If inter-queue preemption is enabled, other queues may use idle shared resources allocated to this queue.
  • Capability quota: The maximum resources available to the queue.

image.png

For a better understanding of the above quota concepts, refer to the following examples:

image.png

Queue parameter description

Parameters Parameter type Description
name Metadata Queue name: The name of the new queue; “root” is not allowed
labels Metadata The name of the parent queue to which this queue belongs. When creating a queue, the default queue and the second-level queues of hierarchical queues are prohibited from being set as parent queues
Guarantee Spec Guaranteed resources: Reserved resources for the queue that cannot be borrowed by others, even if unallocated. The summed guaranteed resources of sub-queues should not exceed that of the parent queue, which in turn should not exceed cluster resources. By default, this is not configured; idle cluster resources are usable.
Deserved Spec Deserved resources: Resources requested by the queue. The summed deserved resources of sub-queues should not exceed that of the parent queue, which in turn should not exceed cluster resources. By default, this is not configured; idle cluster resources are usable.
Capability Spec Capability resources: The maximum permissible resources for the queue. The capability quota of sub-queues should not exceed that of the parent queue, which itself should not exceed cluster resources. By default, this is not configured.
Reclaimable Spec Enable inter-queue preemption and allow borrowed resources to be preempted by default.
State Status Queue status: Defaults to Open; supports setting to Close to disable the queue. Once closed, the queue will reject new task submissions
baidu.com/elastic-queue-is-parent Annotations Is parent queue: Defaults to No. Direct task submission to parent queues is not supported, and sub-queues cannot be created if the parent queue has running tasks
baidu.com/elastic-queue-parent Annotations The name of the parent queue can be specified; defaults to the cluster root queue (root). Cannot be specified as the default queue, second-level queues in hierarchical queues, or queues with running tasks

Queue notes

Create a queue

  • If the system root queue is not initialized, queue creation/modification is not allowed
  • Creating queues with duplicate names is not allowed
  • If no parent queue is specified, the root queue is set as the parent queue by default
  • Queue quota is subject to the following restrictions:

    • Applying for negative quotas is not allowed
    • Queues must meet the configuration rule: capability ≥ deserved ≥ guarantee
    • The capability quota of a queue cannot exceed that of its parent queue
    • The sum of deserved or guarantee quotas of queues at the same level in the cluster cannot exceed those of their parent queue
  • Hierarchical queues are subject to the following restrictions:

    • The specified parent queue must exist
    • Specify the default queue as a parent queue
    • Specify a Level 2 queue as a parent queue
    • Adding sub-queues to queues with running tasks is not allowed
  • Physical queue is subject to the following restrictions:

    • Configuration of capability, deserved and guarantee must be the same

Queue deletion

  • In hierarchical queues, directly deleting a parent queue is not allowed. It is required to delete the sub-queues before deleting the parent queue;
  • Queues with active tasks cannot be deleted.

Queue update

  • Physical queues cannot be converted to ordinary queues
  • The physical queue binding label of logical queues under a physical queue cannot be modified
  • The capability quota cannot be less than the allocated (already allocated) resources
  • After modifying the parent queue

    • The capability quota of the queue cannot be less than the sum of capability quotas of its sub-queues
    • The parent queue’s deserved or guarantee quota cannot be less than the sum of those of its sub-queues
  • Violations of the queue creation rules are not allowed

YAML template operation steps

Logical queue usage

Create elastic queues and hierarchical queues

  1. Prepare the YAML file for queue creation (see the example below)
Plain Text
1apiVersion: scheduling.volcano.sh/v1beta1
2kind: Queue
3metadata:
4  name: demo-queue
5  labels:
6    baidu.com/queue-parent: xxxx
7spec:
8  capability:
9    baidu.com/a800_80g_cgpu: "6"
10  guarantee:
11    resource:
12      baidu.com/a800_80g_cgpu: "1"
13  deserved:
14    baidu.com/a800_80g_cgpu: "2"
15  reclaimable: false
16status:
17  allocated:
18    cpu: "0"
19    memory: "0"
20  reservation: {}
21  state: Open
  1. Submit the YAML file to create the queue in the cluster

kubectl apply -f demo-queue.yaml

Update logical queue

Run the following command to modify the queue configuration:

kubectl edit queue demo-queue

Submit tasks to logical queues

Submit tasks using YAML to test and verify queue functionalities.

Plain Text
1apiVersion: "kubeflow.org/v1"
2kind: PyTorchJob
3metadata:
4  name: training-task
5spec:
6  runPolicy:
7    schedulingPolicy:
8 queue: "queue-name" // Specify the queue
9      priorityClass: "normal"
10  pytorchReplicaSpecs:
11    Worker:
12      replicas: 2
13      template:
14        spec:
15          tolerations:
16          - key: "kwok.x-k8s.io/node"
17            operator: "Equal"
18            value: "fake"
19            effect: "NoSchedule"
20          containers:
21          - name: pytorch
22            image: demo-image:1.0
23            imagePullPolicy: IfNotPresent
24            resources:
25              limits:
26                nvidia.com/gpu: 4
27          schedulerName: volcano
Parameters Parameter type Description
name Meta Task name
runPolicy.schedulingPolicy.queue Spec Submit tasks to the specified queue
runPolicy.schedulingPolicy.priorityClass Spec Task priority: Supports three levels: high, medium, low
pytorchReplicaSpecs.Worker.replicas Spec Number of task replicas

Physical queue usage

Node partitioning

Include target nodes in the physical queue. It's advisable to evacuate tasks from the relevant nodes beforehand; nodes with uncompleted tasks can also be directly added to the physical queue, but their GPU resources will only be released once the tasks are finalized.

Note: The resource application for the physical queue must match the resources of the incorporated nodes precisely. For example, if a physical queue requests 16 GPU cards, two nodes with 8 GPU cards each can be included. Over-partitioning wastes resources, while under-partitioning might cause scheduling issues.

Partitioning is divided into two steps:

  • Add taints: This will only affect the scheduling of newly submitted tasks and will not impact tasks already in progress.

kubectl taint nodes targetNode aihc.baidu.com/dedicated-pool=queue-physical-name:NoSchedule

  • Add tag

kubectl label nodes targetNode aihc.baidu.com/dedicated-pool=queue-physical-name

Plain Text
1Note: When adding labels and taints, replace targetNode with the node IP. Separate multiple IPs with spaces.

Create physical queue

Define a physical queue using the YAML template below

Plain Text
1apiVersion: scheduling.volcano.sh/v1beta1
2kind: Queue
3metadata:
4  labels:
5    baidu.com/queue-is-physical-queue: "true"
6  name: queue-physical-name
7spec:
8  capability:
9    nvidia.com/gpu: "16"
10  deserved:
11    nvidia.com/gpu: "16"
12  guarantee:
13    resource:
14      nvidia.com/gpu: "16"
15  reclaimable: false
16  weight: 1
Plain Text
1Note: Only modify 4 parameters: name (queue name), capability, deserved, and guarantee. The three quota-related parameters (capability, deserved, guarantee) must be consistent and correspond to the number of GPU cards applied for by the physical queue. Keep other parameters unchanged.

Create logical queues under a queue (optional)

To allocate more granular resources within a physical queue, you can establish logical sub-queues under it. Unlike physical queues, sub-queues are not tied to specific nodes and only ensure compliance with resource quotas.

Plain Text
1apiVersion: scheduling.volcano.sh/v1beta1
2kind: Queue
3metadata:
4  labels:
5    baidu.com/queue-parent: queue-physical-name
6  name: queue-logical-name
7spec:
8  deserved:
9    nvidia.com/gpu: "8"
10  guarantee:{}
11  reclaimable: false
12  weight: 1

To request a fixed-quota logical queue, refer to the example above and set the desired quota. Ensure that the sum of desired quotas across all sub-queues does not exceed the total quota assigned to the physical queue.

Plain Text
1Note: The following actions are invalid:
2 * Adding sub-queues under a physical queue with running tasks
3 * Submitting tasks to a physical queue that already has sub-queues

Modify physical queue

(1) Update physical queue quota

Use the command kubectl edit queue queue-physical-name to modify the physical queue quota. Set the quota to the total quota after incorporating new nodes (2) Incorporate new nodes Similar to the node partitioning steps, perform the following two operations on the new nodes

  • Add taints

kubectl taint nodes targetNode aihc.baidu.com/dedicated-pool=queue-physical-name:NoSchedule

  • Add tag

kubectl label nodes targetNode aihc.baidu.com/dedicated-pool=queue-physical-name

Previous
Create Queue
Next
Queue deletion