Go to file
2015-12-11 14:32:39 -08:00
centos-7/cuda Split cuda image into runtime and devel 2015-11-20 18:08:10 -08:00
samples Split cuda image into runtime and devel 2015-11-20 18:08:10 -08:00
ubuntu-14.04/cuda Use new deb package for installing cuDNN v4 2015-12-11 14:32:39 -08:00
centos Add CUDA 7.0 and CUDA 7.5 Dockerfiles for CentOS 7.1 2015-11-17 16:56:18 -08:00
CLA Update CLA 2015-11-13 10:05:15 -08:00
LICENSE Initial commit 2015-11-04 13:00:25 -08:00
Makefile Use new deb package for installing cuDNN v4 2015-12-11 14:32:39 -08:00
nvidia-docker Use the docker command from environment if provided 2015-11-17 15:03:20 -08:00
README.md Split cuda image into runtime and devel 2015-11-20 18:08:10 -08:00
ubuntu Initial commit 2015-11-04 13:00:25 -08:00

NVIDIA Docker

This repository includes utilities to build and run NVIDIA Docker images.
Please be aware that this project is currently experimental.

docker
Example of how CUDA integrates with Docker

Benefits of GPU containerization

  • Reproducible builds
  • Ease of deployment
  • Isolation of individual devices
  • Run across heterogeneous driver/toolkit environments
  • Requires only the NVIDIA driver

Building images

Images can be built on any machine running Docker, it doesn't require a NVIDIA GPU nor any driver installation.

CUDA

Each CUDA image comes in two flavors:

  • runtime: a lightweight image containing the bare minimum to deploy a pre-built application which uses CUDA.
  • devel: extends the runtime image by adding the compiler toolchain, the debugging tools and the development files for the standard CUDA libraries. Use this image to compile a CUDA application from source.

The devel image must be built after the runtime image:

# Building a CUDA 7.5 development image based on Ubuntu
docker build -t cuda:7.5-runtime ubuntu/cuda/7.5/runtime
docker build -t cuda:7.5-devel ubuntu/cuda/7.5/devel
docker tag cuda:7.5-devel cuda
# Building a CUDA 7.5 development image based on CentOS
docker build -t cuda:7.5-runtime centos/cuda/7.5/runtime
docker build -t cuda:7.5-devel centos/cuda/7.5/devel
docker tag cuda:7.5-devel cuda

Alternatively, one can build an image directly from this repository:

# Building a CUDA 7.5 development image based on Ubuntu
docker build -t cuda:7.5-runtime github.com/NVIDIA/nvidia-docker#:ubuntu/cuda/7.5/runtime
docker build -t cuda:7.5-devel github.com/NVIDIA/nvidia-docker#:ubuntu/cuda/7.5/devel
docker tag cuda:7.5-devel cuda

cuDNN

Currently, only cuDNN v2 based on CUDA 7.0 is supported, this image also comes in two flavors: cudnn-runtime and cudnn-devel.

# Building a cuDNN image based on CUDA 7.0 runtime
docker build -t cuda:7.0-runtime ubuntu/cuda/7.0/runtime
docker build -t cuda:7.0-cudnn2-runtime ubuntu/cuda/7.0/runtime/cudnn2
docker tag cuda:7.0-cudnn2-runtime cuda:cudnn-runtime

Building the development image requires the cuda:7.0-devel image:

# Building a cuDNN image based on CUDA 7.0 development
docker build -t cuda:7.0-runtime ubuntu/cuda/7.0/runtime
docker build -t cuda:7.0-devel ubuntu/cuda/7.0/devel
docker build -t cuda:7.0-cudnn2-devel ubuntu/cuda/7.0/devel/cudnn2
docker tag cuda:7.0-cudnn2-devel cuda:cudnn-devel

NVIDIA Docker wrapper

The nvidia-docker script is a drop-in replacement for docker CLI. In addition, it takes care of setting up the NVIDIA host driver environment inside Docker containers for proper execution.

GPUs are exported through a list of comma-separated IDs using the environment variable GPU. The numbering is the same as reported by nvidia-smi or when running CUDA code with CUDA_DEVICE_ORDER=PCI_BUS_ID, it is however different from the default CUDA ordering.

GPU=0,1 ./nvidia-docker <docker-options> <docker-command> <docker-args>

CUDA requirements

Running a CUDA container requires a machine with at least one CUDA-capable GPU and a driver compatible with the CUDA toolkit version you are using.
The machine running the CUDA container only requires the NVIDIA driver, the CUDA toolkit doesn't have to be installed.

NVIDIA drivers are backward-compatible with CUDA toolkits versions

CUDA toolkit version Minimum driver version Minimum GPU architecture
6.5 >= 340.29 >= 2.0 (Fermi)
7.0 >= 346.46 >= 2.0 (Fermi)
7.5 >= 352.39 >= 2.0 (Fermi)

Samples

Once you have built the required images, a few examples are provided in the folder samples.
The following assumes that you have an Ubuntu-based CUDA image in your repository (see samples/ubuntu/deviceQuery/Dockerfile):

# Run deviceQuery with one selected GPU
$ docker build -t device_query samples/ubuntu/deviceQuery
$ GPU=0 ./nvidia-docker run device_query

[ NVIDIA ] =INFO= Driver version: 352.39
[ NVIDIA ] =INFO= CUDA image version: 7.5

./deviceQuery Starting...

 CUDA Device Query (Runtime API) version (CUDART static linking)

Detected 1 CUDA Capable device(s)

Device 0: "GeForce GTX 980"
  [...]

deviceQuery, CUDA Driver = CUDART, CUDA Driver Version = 7.5, CUDA Runtime Version = 7.5, NumDevs = 1, Device0 = GeForce GTX 980
Result = PASS

Issues and Contributing