Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions docs/queues/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,6 +2,8 @@

Scheduling queues are the core resource management primitive in KAI Scheduler, providing hierarchical resource allocation with quota guarantees and priority-based distribution.

Currently only queues that have parent queues can be scheduled to and top level queues cannot be used by jobs. This is a limitation that will be fixed in the future [#621](github.com/NVIDIA/KAI-scheduler/issues/621).

## Table of Contents
- [Queue Attributes](#queue-attributes)
- [API Reference](#api-reference)
Expand Down
2 changes: 2 additions & 0 deletions docs/quickstart/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,6 +11,8 @@ After installing KAI Scheduler, a **two-level queue hierarchy** is automatically
* `default-parent-queue` – Top-level (parent) queue. By default, this queue has no reserved resource quotas, allowing governance of resource distribution for its leaf queues.
* `default-queue` – Leaf (child) queue under the `default-parent-queue` top-level queue. Workloads should reference this queue.

The expected default queues are in [default-queues.yaml](default-queues.yaml).

No manual queue setup is required. Both queues will exist immediately after installation, allowing you to start submitting workloads right away.
To customize scheduling, you can create additional queues or modify existing ones to set quotas, priorities, and hierarchies.

Expand Down
4 changes: 2 additions & 2 deletions docs/quickstart/pods/cpu-only-pod.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ kind: Pod
metadata:
name: cpu-only-pod
labels:
kai.scheduler/queue: test
kai.scheduler/queue: default-queue
spec:
schedulerName: kai-scheduler
containers:
Expand All @@ -16,4 +16,4 @@ spec:
resources:
requests:
cpu: 100m
memory: 250M
memory: 250M
4 changes: 2 additions & 2 deletions docs/quickstart/pods/gpu-pod.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@ kind: Pod
metadata:
name: gpu-pod
labels:
kai.scheduler/queue: test
kai.scheduler/queue: default-queue
spec:
schedulerName: kai-scheduler
containers:
Expand All @@ -16,4 +16,4 @@ spec:
args: ["nvidia-smi; trap 'exit 0' TERM; sleep infinity & wait"]
resources:
limits:
nvidia.com/gpu: "1"
nvidia.com/gpu: "1"
Loading