train model?
See original GitHub issueWould it be possible to train a model on our own data leveraging pretrained BERT models from transformers
library for instance?
I don’t see anything related to the training part (on the Pytorch part in particular)
Issue Analytics
- State:
- Created 4 years ago
- Comments:6 (2 by maintainers)
Top Results From Across the Web
Model Trains, Train Sets, & Railroad Accessories ...
ModelTrainStuff is your one-stop-shop for all your model train stuff - model trains, train sets, railroad models, and more! Explore our huge inventory...
Read more >TrainWorld: Model Train | Train Set | Model Railroad | Model ...
Lionel Model Train Sets At TrainWorld is your one-stop shop for all your model train stuff. Whether you are looking for Lionel Trains...
Read more >Lionel Trains: World's Best Model Trains & Railroad
Mickey Mouse Ready-to-Play Train Set. 7-11773. $114.99. Denver & Rio Grande LEGACY EM-1 #224. 2031110. $1699.99. Santa Fe LEGACY 4-8-4 #3765. 2031190.
Read more >Model Train - Amazon.com
Wood Trick Wooden Toy Train Set with Railway - 34x7″ - Locomotive Train Toy Mechanical Model Kit - 3D Wooden Puzzles for Adults...
Read more >Walthers | Model Railroading | Ho Scale Trains, Scenery ...
A lifetime of model railroading fun starts here, with complete starter sets in popular scales. Our huge selection of model trains, scale structure...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Thanks. I close the issue, however I think that sharing the training part source code would be really great (even if easy to adapt). Cho code is old and TF based, TF is changing all the time, it requires an analysis to be sure to have the same hyper parameters, etc.
We use reuse of raw logits as in Cho paper, and model code based on huggingface’s transformers library.
The training can be benefited from gradually increasing the number of positive examples to avoid overfitting (i forget the name of this technique off the top of my head…)