Bad inference result.
See original GitHub issueHello, I’m trying to reproduce this issue #308 using the same audio but I’m still receiving Gibberish (ish) inferences.
Since I could not find any information on which model and which command they were using in the issue, I’m posting here the info I’m using:
# Download test audio and resample (sampling rate) to 16k
deepspeech.pytorch# wget https://dare.wiscweb.wisc.edu/wp-content/uploads/sites/1051/2008/04/Arthur.mp3
deepspeech.pytorch# sox Arthur.mp3 -c 1 -r 16000 arthur_clip.wav trim 0 15
# Running the inference on the audio clip
deepspeech.pytorch# python transcribe.py --model-path librispeech_pretrained_v2.pth --audio-path arthur_clip.wav --lm-path 3-gram.pruned.3e-7.arpa --alpha 1.65 --beta 0.35
>>>
{
"output": [
{
"transcription": "THE STARY OF OWT OF THE WRAPTH ONCE UPON A TIME THERE WAS A YOUNG RAG AND CUTD IN MYE GUFF ERS MOINE WHENEVER THE HAD THE RIHT SAYES HIM IF HE WOULD LIKE TO COME OUT HUNTING BOT THEM HE WHEN ANSWER IN A HORSE"
}
],
"_meta": {
"acoustic_model": {
"name": "librispeech_pretrained_v2.pth"
},
"language_model": {
"name": "3-gram.pruned.3e-7.arpa"
},
"decoder": {
"lm": true,
"alpha": 1.65,
"beta": 0.35,
"type": "greedy"
}
}
}
I’m using the latest release (v2) as well as its respective commit.
As you can see, those are fairly different results from the one @ryanleary got in aforementioned comment
I tested several different configurations with the different models and both 3-gram.pruned.3e-7.arpa
and 3-gram.3e-7.arpa
as arguments for the transcribe.py
script but in every case I got weird results with those uppercase characters and random words.
Am I doing something wrong here?
Issue Analytics
- State:
- Created 4 years ago
- Comments:12 (3 by maintainers)
Top Results From Across the Web
Bad Inferences – Fallacies and Biases
Overview: What better way to avoid relying on weak inferences than knowing what they look like? This page lists common fallacies and biases....
Read more >Bad Inference result during testing even when using training ...
Hi, I have encounter a weird behavior from my model, during training phase, I printed out some of my validation result from my...
Read more >MobileNet and MobileNetV2: Bad Inference Results
After loading a model of MobileNetV2, I am exectuing a classification on random images from ImageNet or Google Images. Almost always the top-1 ......
Read more >(PDF) Does bad inference drive out good? - ResearchGate
This paper joins the debate on the (mis)use of statistics in the medical literature. Even though the validation process of a statistical result...
Read more >Statistical fallacies and how to avoid them | Geckoboard
The practice of selecting results that fit your claim and excluding those that don't. The worst and most harmful example of being dishonest...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
My bad, aggressive stale bot… will try to get time to reproduce/investigate
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.