question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

`T5ForSequenceClassification`

See original GitHub issue

🚀 Feature request

T5 to classify sequences by using only the encoder of T5 and a ClassificationHead.

Motivation

This gives the benefits of fine-tuning a model with no maximum sequence length (useful for long sequence tasks) without having to load the decoder weights into memory/treat it as a generative task.

Your contribution

I already have working code for this, and saw some requests for it in other forums (slack, torch, huggingface) so if it’s a welcome addition I’d be happy to add it to the library.

Issue Analytics

  • State:open
  • Created 2 years ago
  • Reactions:5
  • Comments:17 (2 by maintainers)

github_iconTop GitHub Comments

2reactions
minmaxmecommented, Dec 10, 2021

This seems like a useful addition, especially considering the EncT5 paper

1reaction
stefan-itcommented, Dec 14, 2022

Hi @osainz59 I think one really interesting dataset would be the CoNLL-2003 (see https://huggingface.co/datasets/conll2003).

When testing the mT5 model series, the WikiANN (Rahimi splits from here: https://huggingface.co/datasets/wikiann) is also very interesting (train on English split only and test it on the other languages for comparisons with the mT5 paper) 😃

Read more comments on GitHub >

github_iconTop Results From Across the Web

T5 - Hugging Face
T5 is an encoder-decoder model and converts all NLP problems into a text-to-text format. It is trained using teacher forcing. This means that...
Read more >
Classification: T5 - seekinginference
In this guide we use T5, a pre-trained and very large (e.g., roughly twice the size of BERT-base) encoder-decoder Transformer model for a ......
Read more >
The Guide to Multi-Tasking with the T5 Transformer
The T5 Transformer can perform any NLP task. It can perform multiple tasks, at the same time, with the same model. Here's how!...
Read more >
T5 for sequence classification - PyTorch Forums
T5 for sequence classification · Shivam_Marathe (Shivam Marathe) May 6, 2021, 12:32pm #1. is it possible to use t5 model to train which...
Read more >
Fine tuning a T5 text-classification model on colab
One of the most interesting recent developments in natural language processing is the T5 family of language models.
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found