question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Guide HF transformer users to use the corresponding hub functions

See original GitHub issue

Is your feature request related to a problem? Please describe. We have push_to_hub_keras to push Keras models. However, HF transformer architectures have a more comprehensive model hub push in save_pretrained. It would be nice to nudge users there if they try to use push_to_hub_keras with an HF transformer.

Describe the solution you’d like A loud warning nudging users to the right place 😄

Describe alternatives you’ve considered An exception, but might be too disruptive.

Additional context As discussed here. I don’t know if there is a similar problem for PyTorch – if there is, this issue would also be applicable there.

cc @merveenoyan

Issue Analytics

  • State:open
  • Created a year ago
  • Comments:6 (6 by maintainers)

github_iconTop GitHub Comments

2reactions
adrinjalalicommented, Apr 21, 2022
1reaction
osansevierocommented, Apr 20, 2022

I think this might be a bit of an overthinking; push_to_hub_keras is not as widely know as transformers push_to_hub, and everywhere in transformers docs we show push_to_hub, so I don’t think people would try using push_to_hub_keras for Transformers models.

EDIT: I do see how the two different use of config can cause confusions, although we also do that with the PT mixin I think.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Contribute to 🤗 Transformers
Submit issues related to bugs or desired new features. Implement new models. Contribute to the examples or to the documentation.
Read more >
Transformers API
The BaseTransformerTrial calls many helper functions below that are also useful when subclassing BaseTransformerTrial or writing custom transformers trials for ...
Read more >
Using Hugging Face Integrations
Gradio has multiple features that make it extremely easy to leverage existing models and Spaces on the Hub. This guide walks through these...
Read more >
7 models on HuggingFace you probably didn't know existed
For most of the people, “using BERT” is synonymous to using the version with weights available in HF's transformers library.
Read more >
Support for Hugging Face Transformer Models
Use the SageMaker model parallelism library's tensor parallelism for training the Hugging Face Transformer models: GPT-2, GPT-J, BERT, and RoBERTa.
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found