Skip to content

Commit c7c3262

Browse files
authored
Update README.md (#175)
Small text changes
1 parent c1df2a9 commit c7c3262

File tree

1 file changed

+7
-7
lines changed

1 file changed

+7
-7
lines changed

docs/gpu-sharing/mps/README.md

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,15 @@
11
# GPU Sharing with MPS
2-
The KAI Scheduler supports GPU sharing by efficiently allocating a single GPU device to multiple pods.
2+
KAI Scheduler supports GPU sharing by efficiently scheduling multiple pods to a single GPU device.
33

4-
The Multi-Process Service (MPS) is an alternative, binary-compatible implementation of the CUDA Application Programming Interface (API). The MPS runtime architecture is designed to transparently enable co-operative multi-process CUDA applications, typically MPI jobs, to utilize Hyper-Q capabilities on the latest NVIDIA (Kepler-based) Tesla and Quadro GPUs.
4+
Multi-Process Service (MPS) is an alternative, binary-compatible implementation of the CUDA Application Programming Interface (API). MPS runtime architecture is designed to transparently enable co-operative multi-process CUDA applications, typically MPI jobs, to utilize Hyper-Q capabilities on the latest NVIDIA GPUs.
55
See the [NVIDIA MPS documentation](https://docs.nvidia.com/deploy/mps/index.html) for more details.
66

7-
There are multiple ways to enable MPS in a Kubernetes cluster. This README focuses on how to use MPS with the KAI Scheduler.
7+
There are multiple ways to enable MPS in a Kubernetes cluster. This README focuses on how to use MPS with KAI Scheduler.
88

99
### Prerequisites
1010
To use GPU sharing, ensure the following requirements are met:
1111
* KAI Scheduler is installed and running in your cluster, with gpu-sharing feature enabled.
12-
2. The MPS server is running on all GPU-enabled hosts (`nvidia-cuda-mps-control`), with the `CUDA_MPS_PIPE_DIRECTORY` environment variable set to `/tmp/nvidia-mps`.
12+
2. MPS server is running on all GPU-enabled hosts (`nvidia-cuda-mps-control`), with the `CUDA_MPS_PIPE_DIRECTORY` environment variable set to `/tmp/nvidia-mps`.
1313

1414
### MPS Enabled PODs
1515
To submit a pod that can share a GPU device and connect to the MPS server, run the following command:
@@ -24,8 +24,8 @@ If the MPS server on the host is configured with a custom `CUDA_MPS_PIPE_DIRECTO
2424

2525
For additional MPS-related environment variables, refer to the [NVIDIA MPS documentation](https://docs.nvidia.com/deploy/mps/index.html#environment-variables).
2626

27-
### Running MPS Server as Pods in the Cluster
28-
If you're running the MPS server as pods on GPU nodes, you must ensure that workload pods are scheduled to the same nodes.
27+
### Running MPS Server as a Pod in the Cluster
28+
If you're running the MPS server as a pod on a GPU node, you must ensure that the workload pods are scheduled to the same nodes.
2929
To achieve this, label the relevant nodes and apply node affinity or a node selector to the workload pods.
3030

3131
For example:
@@ -34,4 +34,4 @@ For example:
3434
```
3535
nodeSelector:
3636
nvidia.mps/enabled: true
37-
```
37+
```

0 commit comments

Comments
 (0)