Questions tagged [nvidia]

25 questions
13
votes
1 answer

How to make my Neural Netwok run on GPU instead of CPU

I have installed Anaconda3 and have installed latest versions of Keras and Tensorflow. Running this command : from tensorflow.python.client import device_lib print(device_lib.list_local_devices()) I find the Notebook is running in CPU: [name:…
Deni Avinash
  • 133
  • 1
  • 1
  • 5
10
votes
2 answers

interpret results of nvidia-smi

Every 1.0s: nvidia-smi Tue Feb 20 12:49:34 2018 Tue Feb 20 12:49:34…
kRazzy R
  • 203
  • 2
  • 7
4
votes
2 answers

Validatioin loss zigzagging

I'm training a speech recognition model using the Nvidia Nemo framework. Just results with the small fastconformer model and two dozen iterations are pretty good; for my data I would say they are quite amazing. However, I have noticed something…
comodoro
  • 143
  • 3
3
votes
1 answer

Is Nvidia Jetson product family also suitable for machine learing model training?

I recently came accross these products (Nvidia Jetson) and they are all tagged as "edge", so i think they are designed only for machine learning inference and not model training. They are quite interesting for their low power consumpion and price…
Skary
  • 135
  • 5
3
votes
2 answers

What does images per second mean when benchmarking Deep Learning GPU?

I've been reviewing performance of several NVIDIA GPU's and I see that typically results are presented in terms of "images per second" that can be processed. Experiments are typically being performed on classical network architectures such as Alex…
Javierfdr
  • 1,500
  • 13
  • 14
2
votes
1 answer

What are "belief maps" and "affinity maps"?

When evaluating Nvidia-Deep Object Pose Estimation, I came across these terms, belief maps and affinity maps. I haven't been able to find a satisfying answer online on what these terms mean.
rajput
  • 31
  • 3
2
votes
1 answer

How can I monitor the usage of NVLink connections?

If I'm running a Keras model on some Nvidia GPUs which are connected via NVLink, how can I monitor the usage of the NVLink connections? I want to be sure the NVLink connections are being used and see at what rate.
user3731622
  • 121
  • 1
  • 3
2
votes
1 answer

"model.to('cuda:6')" becomes (nvidia-smi) GPU 4, same with any other "cuda:MY_GPU", only "cuda:0" becomes GPU 0. How do I get rid of this mapping?

Strange mapping: example In the following example, the first column is chosen in the code, second column is the one that does the work instead: 0:0 1234 MiB 1:2 1234 MiB 2:7 1234 MiB 3:5 2341 MiB 4:1 3412 MiB 5:3 3412 MiB 6:4 3412 MiB 7:6 3412…
questionto42
  • 215
  • 1
  • 10
2
votes
0 answers

Is attention cache useful during transformer pretraining?

I am looking at the MegatronLM implementation, and the only thing that is cached are the results of xK and xV computation: https://github.com/NVIDIA/Megatron-LM/blob/b44dca25727c294a7f825e74a3c4a53744cc8404/megatron/model/transformer.py#L339 Which…
LOST
  • 131
  • 1
1
vote
1 answer

DIGITS Docker container not picking up GPU

I am running DIGITS Docker container but for some reason it fails to recognize host's GPU: it does not report any GPUs (where I expect 1 to be reported) so in the upper right corner of the DIGITS home page there is no indication of any GPUs and also…
1
vote
1 answer

Not able to connect to GPU on Google Colab

I'm trying to use tensorflow with a GPU on Google Colab. I followed the steps listed at https://www.tensorflow.org/install/gpu I confirmed that gpu is visible and CUDA is installed with the commands - !nvcc --version !nvidia-smi This works as…
anirudh
  • 143
  • 1
  • 1
  • 5
1
vote
0 answers

Why doesn't this CNN model need fetures for reducing overfitting?

I found this CNN model by Nvidia end-to-end-deeplearning and with training this model, I'm wondering why this model doesn't need to have dropout layers to reduce overfitting. Neither, this doesn't have activation function. I know we can tune the…
Yuki.U
  • 83
  • 5
1
vote
1 answer

Two different GPUs for Keras (Python)?

One question guys, someone knows if it should be ok to get one more GPU of type Nvidia Geforce GTX 1070 (gaming version), given that now I have GTX 1070 Titanium? They don't have another Titanium card available here, so I have to get a different…
user68541
1
vote
1 answer

How to setup my Ubuntu PC for these Deep Learning frameworks?

I need to use both NVIDIA Digits and tensorflow Object Detection API for different deep learning purposes. I am a bit lost with the compatibility issues, especially for what concerns CUDA. So which version should I choose for DIGITS, Tensorflow-gpu,…
firion
  • 199
  • 4
1
vote
1 answer

Does it make sense to parallelize machine learning algorithms as part of PhD research?

I'm developing machine learning algorithms to aid in the diagnosis and prognosis of various cancers for my PhD. My lab is an Nvidia teaching center (CUDA). My supervisor thinks that I need to also optimize ML by parallelizing it in CUDA. However,…
Sterls
  • 160
  • 1
  • 7
1
2