Using a specific transformer encoder model for Text Classification Task
See original GitHub issueHi,
I would like to use a specific transformer encoder model such as roberta-large
instead of the default roberta
, which loads a roberta-base
. Is there any way to do so?
Issue Analytics
- State:
- Created a year ago
- Comments:5
Top Results From Across the Web
Transformers For Text Classification - Paperspace Blog
This new comprehensive guide will usher you through creating your own transformers nlp model for semantic analysis following two methodologies: from scratch ...
Read more >A Hands-On Guide To Text Classification With Transformer ...
A step-by-step tutorial on using Transformer Models for Text Classification tasks. Learn how to load, fine-tune, and evaluate text ...
Read more >Text classification with Transformer - Keras
Create classifier model using transformer layer. Transformer layer outputs one vector for each time step of our input sequence. Here, we take ...
Read more >Text Classification with Transformers | by Ashwin N - Medium
Text classification is one of the most common tasks in NLP; it can be used for a broad range of applications, such as...
Read more >Text classification - Hugging Face
Text classification is a common NLP task that assigns a label or class to text. ... from transformers import TFAutoModelForSequenceClassification >>> model ......
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hi,
You can use
auto_transformer
, i.e.:Related discussion: https://github.com/ludwig-ai/ludwig/discussions/2057
Documentation: https://ludwig-ai.github.io/ludwig-docs/0.5/configuration/features/text_features/#autotransformer
Hi @justinxzhao ! I recently came across a similar machine learning framework - Lightwood .
On comparison with Ludwig, they both have functionalities to handle features of multiple input types, pre-process them based on their input type, encode the inputs, TRAIN THE MODEL USING THE ENCODED INPUT (emphasizing on the model), decode the outputs from the trained model into output features.
In Lightwood, the possible models that could be trained are provided here - https://lightwood.io/mixer.html
In similar fashion, are there a list of possible models that could be trained using Ludwig? And if yes, how do we specify our preferred model to train (using the concated output from the combiner)?
Thanks!