Incorrect use of torchaudio's rnnt_loss
See original GitHub issueIt is suprising that no one has noticed that the current code is feeding the output of log_softmax to torchaudio’s rnnt_loss, which is not correct.
torchaudio’s rnnt_loss accepts only logits
as input.
I am wondering whether anyone has succeeded in training a transducer model with the current usage of torchaudio’s rnnt_loss using speechbrain.
The relevant code is listed below:
Issue Analytics
- State:
- Created a year ago
- Reactions:1
- Comments:6
Top Results From Across the Web
torchaudio.transformes.RNNTLoss · Issue #2196 · pytorch/audio
Describe the bug. If I don't use the max seq length for logit length, it will cause run time error, input length mismatch....
Read more >torchaudio.functional.rnnt_loss - PyTorch
The RNN Transducer loss extends the CTC loss by defining a distribution over output sequences of all lengths, and by jointly modelling both...
Read more >Somshubra Majumdar on Twitter: "All this to say, now we have ...
Recently Torchaudio 0.10 added RNNT loss too, so now its a lot more available to users. The days of being limited by the...
Read more >torchaudio Changelog - pyup.io
Removed invalid token blanking logic from RNN-T decoder (2180) ... Moved TorchAudio conda package to use pytorch-mutex (1904)
Read more >Speech Command Recognition with torchaudio
We use torchaudio to download and represent the dataset. ... 1 pin_memory = True else: num_workers = 0 pin_memory = False train_loader =...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
It’s strange but I didn’t notice any improvements during the training process, after fixing this bug it produces even worse results during the validation… I’ll make the update once I decode it on the test set
Hello,
This issue has been solved with #1368, so I’m closing this issue. Thanks again for reporting this problem! 😃