question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

T5 Conversion from Original Tensorflow Produce rubbish Text

See original GitHub issue

Environment info

  • transformers version: 3.0.2
  • Platform: Linux-4.19.112±x86_64-with-Ubuntu-18.04-bionic
  • Python version: 3.6.9
  • PyTorch version (GPU?): 1.6.0+cu101 (False)
  • Tensorflow version (GPU?): 2.3.0 (False)
  • Using GPU in script?: yes
  • Using distributed or parallel set-up in script?: no

Who can help

Text Generation: @TevenLeScao T5: @patrickvonplaten

Information

Model I am using (Bert, XLNet …): T5

The problem arises when using:

  • the official example scripts: (give details below)
  • my own modified scripts: (give details below)

The tasks I am working on is:

  • an official GLUE/SQUaD task: (give the name)
  • my own task or dataset: (give details below)

To reproduce

Steps to reproduce the behavior:

https://colab.research.google.com/drive/112Jt7VFwHHT-QmMxFPJ764GNJBn0d5eX?usp=sharing

Expected behavior

We have started a big project for source code tasks (generation, summarisation, documentation, etc.) using language models. Using T5 text to text library, the model can predict the input correctly, However, after we converted the Tensorflow checkpoint to huggingface the output text is rubbish. I am not sure if we are doing something wrong during conversion or there is a problem in loading and converting the weights from the original Tensorflow checkpoint to Pytorch.

The above Colab re-produce the issue. Important Note: We are using a copy of “adapt_t5_for_covid_19_3b” branch which should fix the conversion problem with only one small modification, setting is_tied to false.

Your help is highly appreciated.

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Comments:16 (16 by maintainers)

github_iconTop GitHub Comments

4reactions
patrickvonplatencommented, Nov 13, 2020

should be fixed now. Everything is explained in the PR.

4reactions
patrickvonplatencommented, Oct 15, 2020

Thanks, I hope to be able to take a look at this soon!

Read more comments on GitHub >

github_iconTop Results From Across the Web

A Full Guide to Finetuning T5 for Text2Text and Building ...
In this article, we see a complete example of fine-tuning of T5 for ... we have a text in input and we want...
Read more >
Text classification
Text classification is a common NLP task that assigns a label or class to text. Some of the largest companies run text classification...
Read more >
T5: Text-To-Text Transfer Transformer | by Rohan Jagtap
In this article, we'll discuss Google's state of the art, T5 — Text-to-Text Transfer Transformer Model which was proposed earlier this year ...
Read more >
Text Summarization Approaches for NLP - Practical Guide ...
To make things worse, junk food also clogs your arteries and increases the risk of a heart attack. Therefore, it must be avoided...
Read more >
Chapter 4. Model Explainability and Interpretability
See the original blog post for the Tensorflow version. ... In this section you create a selection of texts for the model to...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found