Masking not working in training, thanks
See original GitHub issueHi, I have tried to train the model on GPU with masking enabled. The line 94
t = torch.flip(t, dims = (2,))
reports an error: RuntimeError: “flip_cuda” not implemented for ‘Bool’, even though I have tried to move mask to CPU.
Any ideas to solve the problem? Thanks a lot.
Issue Analytics
- State:
- Created 2 years ago
- Reactions:1
- Comments:6 (3 by maintainers)
Top Results From Across the Web
Transformer masking during training or inference?
The trick is that you do not need masking at inference time. The purpose of masking is that you prevent the decoder state...
Read more >Dear Littler: How do we handle “mask shaming” when we ...
Integrate masking as a form of “bias” in your periodic implicit bias and/or anti-harassment training initiatives by specifically including ...
Read more >Draw mask not working correctly - Apple Support Communities
Draw mask not working correctly ... Thank you in advance ... it's a pretty good mask wipe tutorial from the Ripple Training guys....
Read more >Image masks not working? · Issue #1460 · colmap ... - GitHub
Am I doing anything wrong, or am I wrong in thinking that colmap uses these keypoints? Would appreciate any advice! Thanks!
Read more >Masking up while working out - University of Washington
I vividly remember my first time running a trail in a face mask. ... especially during intense training or aerobic exercise.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Thanks a lot for the info. I will check that. If not working, I will post an issue again. Let’s close this issue and thanks again for your help.
Dear @lucidrains I can confirm that the new change fixed the problem of masking, my train can go without problems. Thanks a lot for your efforts.
However, there is another issue when I train the model on multi-GPUs.
\site-packages\rotary_embedding_torch\rotary_embedding_torch.py", line 45, in apply_rotary_emb t = (t * freqs.cos()) + (rotate_half(t) * freqs.sin()) RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:1 and cuda:0!
I am not sure if this is a known issue. But at least I can train it on a single GPU. So we can close this issue. Thank you very much.