`T5ForSequenceClassification`
See original GitHub issue🚀 Feature request
T5 to classify sequences by using only the encoder of T5 and a ClassificationHead
.
Motivation
This gives the benefits of fine-tuning a model with no maximum sequence length (useful for long sequence tasks) without having to load the decoder weights into memory/treat it as a generative task.
Your contribution
I already have working code for this, and saw some requests for it in other forums (slack, torch, huggingface) so if it’s a welcome addition I’d be happy to add it to the library.
Issue Analytics
- State:
- Created 2 years ago
- Reactions:5
- Comments:17 (2 by maintainers)
Top Results From Across the Web
T5 - Hugging Face
T5 is an encoder-decoder model and converts all NLP problems into a text-to-text format. It is trained using teacher forcing. This means that...
Read more >Classification: T5 - seekinginference
In this guide we use T5, a pre-trained and very large (e.g., roughly twice the size of BERT-base) encoder-decoder Transformer model for a ......
Read more >The Guide to Multi-Tasking with the T5 Transformer
The T5 Transformer can perform any NLP task. It can perform multiple tasks, at the same time, with the same model. Here's how!...
Read more >T5 for sequence classification - PyTorch Forums
T5 for sequence classification · Shivam_Marathe (Shivam Marathe) May 6, 2021, 12:32pm #1. is it possible to use t5 model to train which...
Read more >Fine tuning a T5 text-classification model on colab
One of the most interesting recent developments in natural language processing is the T5 family of language models.
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
This seems like a useful addition, especially considering the EncT5 paper
Hi @osainz59 I think one really interesting dataset would be the CoNLL-2003 (see https://huggingface.co/datasets/conll2003).
When testing the mT5 model series, the WikiANN (Rahimi splits from here: https://huggingface.co/datasets/wikiann) is also very interesting (train on English split only and test it on the other languages for comparisons with the mT5 paper) 😃