RuntimeError: "bernoulli_scalar_cuda_" not implemented for 'torch.cuda.HalfTensor'
See original GitHub issueHi, I’ve been found that latest PyTorch and latest apex conflict with this error:
RuntimeError: "bernoulli_scalar_cuda_" not implemented for 'torch.cuda.HalfTensor'
and it requires some remapping the tables in amp
model. Thank you!
Issue Analytics
- State:
- Created 5 years ago
- Comments:5 (3 by maintainers)
Top Results From Across the Web
RuntimeError: "exp" not implemented for 'torch.LongTensor'
I happened to follow this tutorial too. For me I just got the torch.arange to generate float type tensor. from position = torch.arange(0, ......
Read more >RuntimeError: "lu_cuda" not implemented for 'Half'
Hi All, I've been starting to run my code on a GPU and started to change the default dtype via torch.set_default_type(torch.half).
Read more >View not implemented for type torch.HalfTensor
Trying to use views with halfTensors, I get this error: >>> import torch >>> torch.__version__ '1.0.1.post2' >>> t = torch.tensor([1,2,3,4.]
Read more >RuntimeError: arange_out not supported on CPUType for Half
HalfTensor ' and device(type='cuda', index=1) so everything should be in GPU memory … I dont understand what is the CPU data causing the...
Read more >"addcmul_cuda" not implemented for 'ComplexFloat' - complex
The input, output and trainable weights in this module are with dtype=torch.complex64. When training this module with Adam optimizer on GPU ...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
It still might be a good idea to use inplace=False, because that will trigger fused kernels for dropout, whereas inplace=True will fall back to a slower implementation. Difference in memory use is not that much.
Oh I didn’t know that, thanks @ngimel !