Support for pre-training the language model
See original GitHub issueIs your feature request related to a problem? Please describe. In order to use the classifier on different languages / specific domains it would be useful to be able to pretrain the language model.
Describe the solution you’d like
Calling .fit on a corpus (i.e.) no labels should train the language model.
model.fit(corpus)
Describe alternatives you’ve considered Use the original repo which doesn’t have a simple to use interface.
Issue Analytics
- State:
- Created 5 years ago
- Comments:11 (7 by maintainers)
Top Results From Across the Web
Why Do Pretrained Language Models Help in Downstream ...
Abstract: Pretrained language models have achieved state-of-the-art performance when adapted to a downstream NLP task.
Read more >Why Do Pretrained Language Models Help in ... - OpenReview
We propose an analysis framework that links the pretraining and downstream tasks with an underlying latent variable generative model of text — the...
Read more >Training a causal language model from scratch - Hugging Face
In this chapter, we'll take a different approach and train a completely new model from scratch. This is a good approach to take...
Read more >Pre-trained Language Models: Simplified | by Prakhar Ganesh
A model which trains only on the task-specific dataset needs to both understand the language and the task using a comparatively smaller dataset....
Read more >lyeoni/pretraining-for-language-understanding: Pre-training of ...
A language model would be trained on a massive corpus, and then we can use it as a component in other models that...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found

@xuy2 This code is merged into master now.
lt means the latter – randomly choosing 512 contiguous tokens from an article. A random slice of text.