You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/gpu-sharing/mps/README.md
+7-7Lines changed: 7 additions & 7 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -1,15 +1,15 @@
1
1
# GPU Sharing with MPS
2
-
The KAI Scheduler supports GPU sharing by efficiently allocating a single GPU device to multiple pods.
2
+
KAI Scheduler supports GPU sharing by efficiently scheduling multiple pods to a single GPU device.
3
3
4
-
The Multi-Process Service (MPS) is an alternative, binary-compatible implementation of the CUDA Application Programming Interface (API). The MPS runtime architecture is designed to transparently enable co-operative multi-process CUDA applications, typically MPI jobs, to utilize Hyper-Q capabilities on the latest NVIDIA (Kepler-based) Tesla and Quadro GPUs.
4
+
Multi-Process Service (MPS) is an alternative, binary-compatible implementation of the CUDA Application Programming Interface (API). MPS runtime architecture is designed to transparently enable co-operative multi-process CUDA applications, typically MPI jobs, to utilize Hyper-Q capabilities on the latest NVIDIA GPUs.
5
5
See the [NVIDIA MPS documentation](https://docs.nvidia.com/deploy/mps/index.html) for more details.
6
6
7
-
There are multiple ways to enable MPS in a Kubernetes cluster. This README focuses on how to use MPS with the KAI Scheduler.
7
+
There are multiple ways to enable MPS in a Kubernetes cluster. This README focuses on how to use MPS with KAI Scheduler.
8
8
9
9
### Prerequisites
10
10
To use GPU sharing, ensure the following requirements are met:
11
11
* KAI Scheduler is installed and running in your cluster, with gpu-sharing feature enabled.
12
-
2.The MPS server is running on all GPU-enabled hosts (`nvidia-cuda-mps-control`), with the `CUDA_MPS_PIPE_DIRECTORY` environment variable set to `/tmp/nvidia-mps`.
12
+
2. MPS server is running on all GPU-enabled hosts (`nvidia-cuda-mps-control`), with the `CUDA_MPS_PIPE_DIRECTORY` environment variable set to `/tmp/nvidia-mps`.
13
13
14
14
### MPS Enabled PODs
15
15
To submit a pod that can share a GPU device and connect to the MPS server, run the following command:
@@ -24,8 +24,8 @@ If the MPS server on the host is configured with a custom `CUDA_MPS_PIPE_DIRECTO
24
24
25
25
For additional MPS-related environment variables, refer to the [NVIDIA MPS documentation](https://docs.nvidia.com/deploy/mps/index.html#environment-variables).
26
26
27
-
### Running MPS Server as Pods in the Cluster
28
-
If you're running the MPS server as pods on GPU nodes, you must ensure that workload pods are scheduled to the same nodes.
27
+
### Running MPS Server as a Pod in the Cluster
28
+
If you're running the MPS server as a pod on a GPU node, you must ensure that the workload pods are scheduled to the same nodes.
29
29
To achieve this, label the relevant nodes and apply node affinity or a node selector to the workload pods.
0 commit comments