From 714bb486f555535dbcaf1a2aeed0cbc263dc1cca Mon Sep 17 00:00:00 2001 From: William Buchwalter Date: Fri, 8 Jun 2018 16:53:37 -0400 Subject: [PATCH] Update NVIDIA drivers to 396 (#3224) --- docs/kubernetes/gpu.md | 2 +- pkg/acsengine/engine.go | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/docs/kubernetes/gpu.md b/docs/kubernetes/gpu.md index 870289520..003e0701b 100644 --- a/docs/kubernetes/gpu.md +++ b/docs/kubernetes/gpu.md @@ -1,7 +1,7 @@ # Microsoft Azure Container Service Engine - Using GPUs with Kubernetes If you created a Kubernetes cluster with one or multiple agent pool(s) whose VM size is `Standard_NC*` or `Standard_NV*` you can schedule GPU workload on your cluster. -The NVIDIA drivers are automatically installed on every GPU agent in your cluster, so you don't need to do that manually, unless you require a specific version of the drivers. Currently, the installed driver is version 390.30. +The NVIDIA drivers are automatically installed on every GPU agent in your cluster, so you don't need to do that manually, unless you require a specific version of the drivers. Currently, the installed driver is version 396.26. To make sure everything is fine, run `kubectl describe node `. You should see the correct number of GPU reported (in this example shows 2 GPU for a NC12 VM): diff --git a/pkg/acsengine/engine.go b/pkg/acsengine/engine.go index 293008b92..8c3996b45 100644 --- a/pkg/acsengine/engine.go +++ b/pkg/acsengine/engine.go @@ -445,7 +445,7 @@ func isCustomVNET(a []*api.AgentPoolProfile) bool { func getGPUDriversInstallScript(profile *api.AgentPoolProfile) string { // latest version of the drivers. Later this parameter could be bubbled up so that users can choose specific driver versions. - dv := "390.30" + dv := "396.26" dest := "/usr/local/nvidia" nvidiaDockerVersion := "2.0.3" dockerVersion := "1.13.1-1"