Not meant as an issue, but a little perplexed by feature_dim...
See original GitHub issueI noticed feature_dim
is set to 50
, which is quite a bottleneck from the encoding dim of 32 * 35 * 35
and the downstream hidden_dim
of 1024
. Very interesting. Do you think the bottleneck helps create some kind of better compression for learning?
Issue Analytics
- State:
- Created 2 years ago
- Comments:5 (3 by maintainers)
Top Results From Across the Web
Perplexed Definition & Meaning - Merriam-Webster
The meaning of PERPLEXED is filled with uncertainty : puzzled. How to use perplexed in a sentence.
Read more >Use perplexed in a sentence
Effie looked somewhat perplexed, but not satisfied. ... His dark features were twisted with a perplexed expression as he swung her around.
Read more >Perplexed definition and meaning | Collins English Dictionary
If you are perplexed, you feel confused and slightly worried by something because you do not understand it. She is perplexed about what...
Read more >Sentences (1/2) Flashcards - Quizlet
Study with Quizlet and memorize flashcards containing terms like abbreviate shorten, abate subside or moderate, abortive unsuccessful; fruitless and more.
Read more >Magical Skills and Goals - Features - Reverse Shot
Magical Skills and Goals Daniel Witkin on the politics and proliferation of the soccer highlight video. Perplexed followers of the World Cup ......
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Initially I used
256
hidden size after the encoder’s50
dimensional output, which also works fine. Over time, however, I noticed that wider value/policy heads are useful for increasing stability of training. I suspect this is because with a wider network you have a better chance to end up in a good init position in the parameter space, so the seed won’t collapse.Yeah, that makes sense. I hadn’t thought of it that way.