`activation_dropout` in OPT is never used
See original GitHub issueSystem Info
main
Who can help?
@patil-suraj, @patrickvonplaten, @LysandreJik
Information
- The official example scripts
- My own modified scripts
Tasks
- An officially supported task in the
examples
folder (such as GLUE/SQuAD, …) - My own task or dataset (give details below)
Reproduction
activation_dropout
in modeling_opt.py
is never used. It would not behave as expected if one initial a model randomly while setting it to non-zero.
Expected behavior
activation_dropout
is used or removed.
Issue Analytics
- State:
- Created a year ago
- Comments:8 (7 by maintainers)
Top Results From Across the Web
Dropout Regularization in Deep Learning Models with Keras
Dropout is only used during the training of a model and is not used when evaluating the skill of the model.
Read more >Where should I place dropout layers in a neural network?
Dropout was used after the activation function of each convolutional layer: CONV->RELU->DROP. So should they be placed after all layers, or only the...
Read more >Dropout behavior in Keras with rate=1 (dropping all input units ...
The Dropout layer simply doesn't do anything when rate is set to 1 (or 0, see here). I guess it's because the scaling...
Read more >InvalidArgumentError: No OpKernel was registered to ... - GitHub
I am user macintosh code `import tensorflow as tf from tensorflow.keras.models import Sequential from tensorflow.keras.layers import Dense, Dropout, ...
Read more >Dropout and Batch Normalization - Kaggle
It seems that batch normalization can be used at almost any point in a network. You can put it after a layer... layers.Dense(16,...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I’m happy to contribute if removing is what we want 😊
Gonna merge it to main 🥳