WARNING: Using conda in DCS images is no longer supported starting Databricks Runtime 9.0. We highly recommend users to extend cuda-11.8 examples.
We no longer support cuda-10.1 and cuda-11.0 compatibility with latest databricks runtime.
WARNING: DCS images which extends venv need to be paired with DBR 14.x when creating a
cluster. In order for REPL to launch successfully, the python packages used in the image HAVE to match with python packges used in the paried DBR version. venv uses version set from DBR 14.0
There are three variations of GPU containers that can be used depending upon the CUDA version you wish to use:
cuda-11.8contains the layers which install CUDA 11.8cuda-11.0(Deprecated) contains the layers which install CUDA 11.0cuda-10.1(Deprecated) contains the layers which install CUDA 10.1
Example base layers to build your own container:
gpu-baseextends the official NVIDIA CUDA container with Databricks Container Service minimal requirements.gpu-venvextendsgpu-baseby installing cuda dependencies and commmon Databricks python dependencies in venv.
Example containers for common GPU use cases:
gpu-tensorflowextendsgpu-venvby creating a conda environment that contains TensorFlow.gpu-pytorchextendsgpu-venvby creating a conda environment that contains PyTorch.
- After the cluster is ready, you can run
%sh nvidia-smito view GPU devices and confirm that they are available.
-
You can modify the
gpu-baseDockerfile and add additional system packages and NVIDIA libraries, for example, TensorRT (libnvinfer). -
You cannot change the NVIDIA driver version, because it must match the driver version on the host machine, which is 450.80.
-
The
gpu-tensorflowandgpu-pytorchDockerfiles provide examples to create the root conda environment from an environment.yml file. These packages are required for Python notebooks and PySpark to work: python, ipython, numpy, pandas, pyarrow, six, and ipykernel.