Training cannot start
See original GitHub issueHi,
Good job! I tried to used it as
device=torch.device('cuda:0' if torch.cuda.is_available() else 'cpu')
model = Model(...)
model = nn.DataParallel(model, device_ids=[0, 1])
model = convert_model(model).to(device)
However, it stucked and training couldn 't start. Have you seem similar problems before ?
Issue Analytics
- State:
- Created 3 years ago
- Comments:7 (1 by maintainers)
Top Results From Across the Web
Can't start training · Issue #1301 · ultralytics/yolov5 - GitHub
Question Hi, when I try to train ,the progress bar does not move.I don't know how to solve it.Hope your reply.
Read more >Cannot start Training on Google Colab - Usage & Issues
Hi. I am trying to use Google Colab to train DLC. I am getting the error: “NameError: name 'deeplabcut' is not defined” I...
Read more >Cannot start training because another fit() call is ongoing ...
It means that the second call has to start only when the first one has completed. Using async and await will prevent your...
Read more >Can't Start Training? : r/apexlegends - Reddit
Been being booted to menu screen every time I try to join into the training mode so I can finally test some stuff....
Read more >Can't start Training mission. :: Cold Waters General Discussions
Now I wanted to start and wanted to get going with the first Training mission, but I can't get it started. After the...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
I wrote a test script which solved the hanging problem. I guess there is something wrong with reduce function.
Please refer to the script: https://gist.github.com/shuuchen/7463009370e9ddf77e649f3fec259024
You can adapt the code with your own task easily.
See also https://github.com/vacancy/Synchronized-BatchNorm-PyTorch/issues/44#issuecomment-815135207