Test tensorflow-gpu failed with Status: CUDA driver version is insufficient for CUDA runtime version (which...
Around usage results
Circuit to "zoom in" on mV fluctuations of a DC signal?
Most bit efficient text communication method?
2001: A Space Odyssey's use of the song "Daisy Bell" (Bicycle Built for Two); life imitates art or vice-versa?
How can I use the Python library networkx from Mathematica?
What is the meaning of the simile “quick as silk”?
Did MS DOS itself ever use blinking text?
また usage in a dictionary
Is there such thing as an Availability Group failover trigger?
What do you call the main part of a joke?
Would "destroying" Wurmcoil Engine prevent its tokens from being created?
Significance of Cersei's obsession with elephants?
Trademark violation for app?
How do I make this wiring inside cabinet safer? (Pic)
Where are Serre’s lectures at Collège de France to be found?
Is grep documentation wrong?
Crossing US/Canada Border for less than 24 hours
Why aren't air breathing engines used as small first stages
An adverb for when you're not exaggerating
Amount of permutations on an NxNxN Rubik's Cube
When the Haste spell ends on a creature, do attackers have advantage against that creature?
If my PI received research grants from a company to be able to pay my postdoc salary, did I have a potential conflict interest too?
Delete nth line from bottom
Do wooden building fires get hotter than 600°C?
Test tensorflow-gpu failed with Status: CUDA driver version is insufficient for CUDA runtime version (which is not true)
Announcing the arrival of Valued Associate #679: Cesar Manara
Planned maintenance scheduled April 17/18, 2019 at 00:00UTC (8:00pm US/Eastern)How to configure X-Windows for CUDA-GDM with 1 GPU card and 1 motherboard adapterCould not find a configuration file for package “ECM” that is compatible with requested version 1.5.0CentOS 6.6 nVidia driver and CUDA 6.5 are in conflict for system with GTX980Nvidia Graphics Driver for system with Dual GPU, GTX670MX and Intel Onchip Graphics Installation for Linux Mint (aka: Hybrid Graphics)<math.h> fails on Intel Centos 6.6 clusterTwo systems with identical GPUs, have very different performances when running Tensorflow script on GPUAny reason for pci-stub not working but vfio-pci working with QEMU/KVM GPU passthrough?Getting the error “DLL load failed: The specified module could not be found.” while trying to import tensorflow for Windows in Anaconda using PyCharmTensorflow gpu was working but is not working anymore error- Failed to get convolution algorithm. This is probably because cuDNN failed to initializeinstall nvidia-driver418 and cuda9.2.-->CUDA driver version is insufficient for CUDA runtime version
.everyoneloves__top-leaderboard:empty,.everyoneloves__mid-leaderboard:empty,.everyoneloves__bot-mid-leaderboard:empty{ height:90px;width:728px;box-sizing:border-box;
}
I have the following configuration:
- SUSE Linux Enterprise Server 12 SP3 (x86_64)
- CUDA Toolkit: CUDA 9.2 (9.2.148 Update 1)
- CUDA Driver Version: 396.37
According to NVIDIA just right (https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html#major-components).
I set up a new environment with Anaconda and installed tensorflow-gpu in it:
conda create -n keras python=3.6.8 anaconda
conda install -c anaconda tensorflow-gpu
But if I then want to check the installation via python console:
import tensorflow as tf
sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))
I get the following error:
2019-04-17 15:23:45.753926: I
tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports
instructions that this TensorFlow binary was not compiled to use:
SSE4.1 SSE4.2 AVX AVX2 FMA
2019-04-17 15:23:45.793109: I
tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency:
2600180000 Hz
2019-04-17 15:23:45.798218: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f42601240 executing computations on platform Host. Devices:
2019-04-17 15:23:45.798258: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): ,
2019-04-17 15:23:45.981727: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f426ad9b0 executing computations on platform CUDA. Devices:
2019-04-17 15:23:45.981777: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): Tesla K40c, Compute Capability 3.5
2019-04-17 15:23:45.982175: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1433] Found device 0
with properties:
name: Tesla K40c major: 3 minor: 5 memoryClockRate(GHz): 0.745
pciBusID: 0000:06:00.0 totalMemory: 11.17GiB freeMemory: 11.09GiB
2019-04-17 15:23:45.982206: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible
gpu devices: 0
Traceback (most recent call last): File "", line 1, in
File
"/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 1551, in init
super(Session, self).init(target, graph, config=config) File "/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 676, in init
self._session = tf_session.TF_NewSessionRef(self._graph._c_graph, opts) tensorflow.python.framework.errors_impl.InternalError:
cudaGetDevice() failed. Status: CUDA driver version is insufficient
for CUDA runtime version
I've been looking for solutions from others with this problem, but for most of them it was because the CUDA Toolkit and Driver version didn't match. Which is not the case with me.
I'd really appreciate the help.
linux drivers python gpu cuda
New contributor
add a comment |
I have the following configuration:
- SUSE Linux Enterprise Server 12 SP3 (x86_64)
- CUDA Toolkit: CUDA 9.2 (9.2.148 Update 1)
- CUDA Driver Version: 396.37
According to NVIDIA just right (https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html#major-components).
I set up a new environment with Anaconda and installed tensorflow-gpu in it:
conda create -n keras python=3.6.8 anaconda
conda install -c anaconda tensorflow-gpu
But if I then want to check the installation via python console:
import tensorflow as tf
sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))
I get the following error:
2019-04-17 15:23:45.753926: I
tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports
instructions that this TensorFlow binary was not compiled to use:
SSE4.1 SSE4.2 AVX AVX2 FMA
2019-04-17 15:23:45.793109: I
tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency:
2600180000 Hz
2019-04-17 15:23:45.798218: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f42601240 executing computations on platform Host. Devices:
2019-04-17 15:23:45.798258: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): ,
2019-04-17 15:23:45.981727: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f426ad9b0 executing computations on platform CUDA. Devices:
2019-04-17 15:23:45.981777: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): Tesla K40c, Compute Capability 3.5
2019-04-17 15:23:45.982175: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1433] Found device 0
with properties:
name: Tesla K40c major: 3 minor: 5 memoryClockRate(GHz): 0.745
pciBusID: 0000:06:00.0 totalMemory: 11.17GiB freeMemory: 11.09GiB
2019-04-17 15:23:45.982206: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible
gpu devices: 0
Traceback (most recent call last): File "", line 1, in
File
"/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 1551, in init
super(Session, self).init(target, graph, config=config) File "/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 676, in init
self._session = tf_session.TF_NewSessionRef(self._graph._c_graph, opts) tensorflow.python.framework.errors_impl.InternalError:
cudaGetDevice() failed. Status: CUDA driver version is insufficient
for CUDA runtime version
I've been looking for solutions from others with this problem, but for most of them it was because the CUDA Toolkit and Driver version didn't match. Which is not the case with me.
I'd really appreciate the help.
linux drivers python gpu cuda
New contributor
add a comment |
I have the following configuration:
- SUSE Linux Enterprise Server 12 SP3 (x86_64)
- CUDA Toolkit: CUDA 9.2 (9.2.148 Update 1)
- CUDA Driver Version: 396.37
According to NVIDIA just right (https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html#major-components).
I set up a new environment with Anaconda and installed tensorflow-gpu in it:
conda create -n keras python=3.6.8 anaconda
conda install -c anaconda tensorflow-gpu
But if I then want to check the installation via python console:
import tensorflow as tf
sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))
I get the following error:
2019-04-17 15:23:45.753926: I
tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports
instructions that this TensorFlow binary was not compiled to use:
SSE4.1 SSE4.2 AVX AVX2 FMA
2019-04-17 15:23:45.793109: I
tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency:
2600180000 Hz
2019-04-17 15:23:45.798218: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f42601240 executing computations on platform Host. Devices:
2019-04-17 15:23:45.798258: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): ,
2019-04-17 15:23:45.981727: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f426ad9b0 executing computations on platform CUDA. Devices:
2019-04-17 15:23:45.981777: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): Tesla K40c, Compute Capability 3.5
2019-04-17 15:23:45.982175: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1433] Found device 0
with properties:
name: Tesla K40c major: 3 minor: 5 memoryClockRate(GHz): 0.745
pciBusID: 0000:06:00.0 totalMemory: 11.17GiB freeMemory: 11.09GiB
2019-04-17 15:23:45.982206: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible
gpu devices: 0
Traceback (most recent call last): File "", line 1, in
File
"/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 1551, in init
super(Session, self).init(target, graph, config=config) File "/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 676, in init
self._session = tf_session.TF_NewSessionRef(self._graph._c_graph, opts) tensorflow.python.framework.errors_impl.InternalError:
cudaGetDevice() failed. Status: CUDA driver version is insufficient
for CUDA runtime version
I've been looking for solutions from others with this problem, but for most of them it was because the CUDA Toolkit and Driver version didn't match. Which is not the case with me.
I'd really appreciate the help.
linux drivers python gpu cuda
New contributor
I have the following configuration:
- SUSE Linux Enterprise Server 12 SP3 (x86_64)
- CUDA Toolkit: CUDA 9.2 (9.2.148 Update 1)
- CUDA Driver Version: 396.37
According to NVIDIA just right (https://docs.nvidia.com/cuda/cuda-toolkit-release-notes/index.html#major-components).
I set up a new environment with Anaconda and installed tensorflow-gpu in it:
conda create -n keras python=3.6.8 anaconda
conda install -c anaconda tensorflow-gpu
But if I then want to check the installation via python console:
import tensorflow as tf
sess = tf.Session(config=tf.ConfigProto(log_device_placement=True))
I get the following error:
2019-04-17 15:23:45.753926: I
tensorflow/core/platform/cpu_feature_guard.cc:141] Your CPU supports
instructions that this TensorFlow binary was not compiled to use:
SSE4.1 SSE4.2 AVX AVX2 FMA
2019-04-17 15:23:45.793109: I
tensorflow/core/platform/profile_utils/cpu_utils.cc:94] CPU Frequency:
2600180000 Hz
2019-04-17 15:23:45.798218: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f42601240 executing computations on platform Host. Devices:
2019-04-17 15:23:45.798258: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): ,
2019-04-17 15:23:45.981727: I
tensorflow/compiler/xla/service/service.cc:150] XLA service
0x561f426ad9b0 executing computations on platform CUDA. Devices:
2019-04-17 15:23:45.981777: I
tensorflow/compiler/xla/service/service.cc:158] StreamExecutor
device (0): Tesla K40c, Compute Capability 3.5
2019-04-17 15:23:45.982175: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1433] Found device 0
with properties:
name: Tesla K40c major: 3 minor: 5 memoryClockRate(GHz): 0.745
pciBusID: 0000:06:00.0 totalMemory: 11.17GiB freeMemory: 11.09GiB
2019-04-17 15:23:45.982206: I
tensorflow/core/common_runtime/gpu/gpu_device.cc:1512] Adding visible
gpu devices: 0
Traceback (most recent call last): File "", line 1, in
File
"/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 1551, in init
super(Session, self).init(target, graph, config=config) File "/home/fuchs/.conda/envs/keras/lib/python3.6/site-packages/tensorflow/python/client/session.py",
line 676, in init
self._session = tf_session.TF_NewSessionRef(self._graph._c_graph, opts) tensorflow.python.framework.errors_impl.InternalError:
cudaGetDevice() failed. Status: CUDA driver version is insufficient
for CUDA runtime version
I've been looking for solutions from others with this problem, but for most of them it was because the CUDA Toolkit and Driver version didn't match. Which is not the case with me.
I'd really appreciate the help.
linux drivers python gpu cuda
linux drivers python gpu cuda
New contributor
New contributor
New contributor
asked 14 hours ago
Stefan RenardStefan Renard
11
11
New contributor
New contributor
add a comment |
add a comment |
1 Answer
1
active
oldest
votes
I'm sorry, I've already found the solution myself. My mistake was that I created the Anaconda environment with a python version >3.5. Under these circumstances tensorflow-gpu=1.13 will be installed, if you execute the following command:
conda install -c anaconda tensorflow-gpu
However, if you create an environment with python=3.5, tensorflow-gpu=1.10 will be installed, which works for this CUDA version.
New contributor
add a comment |
Your Answer
StackExchange.ready(function() {
var channelOptions = {
tags: "".split(" "),
id: "3"
};
initTagRenderer("".split(" "), "".split(" "), channelOptions);
StackExchange.using("externalEditor", function() {
// Have to fire editor after snippets, if snippets enabled
if (StackExchange.settings.snippets.snippetsEnabled) {
StackExchange.using("snippets", function() {
createEditor();
});
}
else {
createEditor();
}
});
function createEditor() {
StackExchange.prepareEditor({
heartbeatType: 'answer',
autoActivateHeartbeat: false,
convertImagesToLinks: true,
noModals: true,
showLowRepImageUploadWarning: true,
reputationToPostImages: 10,
bindNavPrevention: true,
postfix: "",
imageUploader: {
brandingHtml: "Powered by u003ca class="icon-imgur-white" href="https://imgur.com/"u003eu003c/au003e",
contentPolicyHtml: "User contributions licensed under u003ca href="https://creativecommons.org/licenses/by-sa/3.0/"u003ecc by-sa 3.0 with attribution requiredu003c/au003e u003ca href="https://stackoverflow.com/legal/content-policy"u003e(content policy)u003c/au003e",
allowUrls: true
},
onDemand: true,
discardSelector: ".discard-answer"
,immediatelyShowMarkdownHelp:true
});
}
});
Stefan Renard is a new contributor. Be nice, and check out our Code of Conduct.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fsuperuser.com%2fquestions%2f1426527%2ftest-tensorflow-gpu-failed-with-status-cuda-driver-version-is-insufficient-for%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
1 Answer
1
active
oldest
votes
1 Answer
1
active
oldest
votes
active
oldest
votes
active
oldest
votes
I'm sorry, I've already found the solution myself. My mistake was that I created the Anaconda environment with a python version >3.5. Under these circumstances tensorflow-gpu=1.13 will be installed, if you execute the following command:
conda install -c anaconda tensorflow-gpu
However, if you create an environment with python=3.5, tensorflow-gpu=1.10 will be installed, which works for this CUDA version.
New contributor
add a comment |
I'm sorry, I've already found the solution myself. My mistake was that I created the Anaconda environment with a python version >3.5. Under these circumstances tensorflow-gpu=1.13 will be installed, if you execute the following command:
conda install -c anaconda tensorflow-gpu
However, if you create an environment with python=3.5, tensorflow-gpu=1.10 will be installed, which works for this CUDA version.
New contributor
add a comment |
I'm sorry, I've already found the solution myself. My mistake was that I created the Anaconda environment with a python version >3.5. Under these circumstances tensorflow-gpu=1.13 will be installed, if you execute the following command:
conda install -c anaconda tensorflow-gpu
However, if you create an environment with python=3.5, tensorflow-gpu=1.10 will be installed, which works for this CUDA version.
New contributor
I'm sorry, I've already found the solution myself. My mistake was that I created the Anaconda environment with a python version >3.5. Under these circumstances tensorflow-gpu=1.13 will be installed, if you execute the following command:
conda install -c anaconda tensorflow-gpu
However, if you create an environment with python=3.5, tensorflow-gpu=1.10 will be installed, which works for this CUDA version.
New contributor
New contributor
answered 14 hours ago
Stefan RenardStefan Renard
11
11
New contributor
New contributor
add a comment |
add a comment |
Stefan Renard is a new contributor. Be nice, and check out our Code of Conduct.
Stefan Renard is a new contributor. Be nice, and check out our Code of Conduct.
Stefan Renard is a new contributor. Be nice, and check out our Code of Conduct.
Stefan Renard is a new contributor. Be nice, and check out our Code of Conduct.
Thanks for contributing an answer to Super User!
- Please be sure to answer the question. Provide details and share your research!
But avoid …
- Asking for help, clarification, or responding to other answers.
- Making statements based on opinion; back them up with references or personal experience.
To learn more, see our tips on writing great answers.
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
StackExchange.ready(
function () {
StackExchange.openid.initPostLogin('.new-post-login', 'https%3a%2f%2fsuperuser.com%2fquestions%2f1426527%2ftest-tensorflow-gpu-failed-with-status-cuda-driver-version-is-insufficient-for%23new-answer', 'question_page');
}
);
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Sign up or log in
StackExchange.ready(function () {
StackExchange.helpers.onClickDraftSave('#login-link');
});
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Sign up using Google
Sign up using Facebook
Sign up using Email and Password
Post as a guest
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown
Required, but never shown