question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

CuDNN error: CUDNN_STATUS_INTERNAL_ERROR

See original GitHub issue

hi, i use it to train for my own data set, but it run error, can u give me some suggestion? (my segmentation class img is gray image).

/opt/conda/conda-bld/pytorch_1532581333611/work/aten/src/THCUNN/SpatialClassNLLCriterion.cu:99: void cunn_SpatialClassNLLCriterion_updateOutput_kernel(T *, T *, T *, long *, T *, int, int, int, int, int, long) [with T = float, AccumT = float]: block: [3,0,0], thread: [190,0,0] Assertion t >= 0 && t < n_classes failed. /opt/conda/conda-bld/pytorch_1532581333611/work/aten/src/THCUNN/SpatialClassNLLCriterion.cu:99: void cunn_SpatialClassNLLCriterion_updateOutput_kernel(T *, T *, T *, long *, T *, int, int, int, int, int, long) [with T = float, AccumT = float]: block: [3,0,0], thread: [191,0,0] Assertion t >= 0 && t < n_classes failed. Traceback (most recent call last): File “train.py”, line 234, in <module> main() File “train.py”, line 215, in main loss.backward() File “/home/zhangjunyi/anaconda3/envs/pytorch/lib/python3.6/site-packages/torch/tensor.py”, line 93, in backward torch.autograd.backward(self, gradient, retain_graph, create_graph) File “/home/zhangjunyi/anaconda3/envs/pytorch/lib/python3.6/site-packages/torch/autograd/init.py”, line 90, in backward allow_unreachable=True) # allow_unreachable flag RuntimeError: CuDNN error: CUDNN_STATUS_INTERNAL_ERROR

Issue Analytics

  • State:open
  • Created 5 years ago
  • Comments:5 (1 by maintainers)

github_iconTop GitHub Comments

1reaction
J-zincommented, Oct 31, 2018

i got it, the gray num must be lower than the class nums!!!

0reactions
Lixy1997commented, Jul 23, 2020

@speedinghzl I use VOC dataset, NUM_CLASSED=21, i also get the same problem

Read more comments on GitHub >

github_iconTop Results From Across the Web

PyTorch Model Training: RuntimeError: cuDNN error
Error occurs after about 583 episodes (8 hours) or 1,150,000 steps, where each step involves a forward pass through the LSTM model. My...
Read more >
RuntimeError: CUDNN_STATUS_INTERNAL_ERROR
I got my problem solved by applying the solution: https://stackoverflow.com/questions/45810356/runtimeerror-cudnn-status-internal-error.
Read more >
CUDNN_STATUS_INTERNAL_E...
I receive RuntimeError: cuDNN error: CUDNN_STATUS_INTERNAL_ERROR error running on a multi-gpus server via NGC PyTorch docker image, ...
Read more >
cuDNN fails with CUDNN_STATUS_INTERNAL_ERROR on ...
My System: OS: Ubuntu 16.04. GPU: GTX 1080. CUDA: 8.0.61 cuDNN: 6.0.21. I've installed CUDA / cuDNN in the following installation routine:
Read more >
Pytorch CUDA报错,RuntimeError: cuDNN error - CSDN博客
RuntimeError: cuDNN error: CUDNN_STATUS_INTERNAL_ERROR. 在代码中设置指定GPU,代码仍旧运行不成功. device = torch.device(“cuda:3” if ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found