Guide HF transformer users to use the corresponding hub functions
See original GitHub issueIs your feature request related to a problem? Please describe.
We have push_to_hub_keras
to push Keras models. However, HF transformer architectures have a more comprehensive model hub push in save_pretrained
. It would be nice to nudge users there if they try to use push_to_hub_keras
with an HF transformer.
Describe the solution you’d like A loud warning nudging users to the right place 😄
Describe alternatives you’ve considered An exception, but might be too disruptive.
Additional context As discussed here. I don’t know if there is a similar problem for PyTorch – if there is, this issue would also be applicable there.
cc @merveenoyan
Issue Analytics
- State:
- Created a year ago
- Comments:6 (6 by maintainers)
Top Results From Across the Web
Contribute to 🤗 Transformers
Submit issues related to bugs or desired new features. Implement new models. Contribute to the examples or to the documentation.
Read more >Transformers API
The BaseTransformerTrial calls many helper functions below that are also useful when subclassing BaseTransformerTrial or writing custom transformers trials for ...
Read more >Using Hugging Face Integrations
Gradio has multiple features that make it extremely easy to leverage existing models and Spaces on the Hub. This guide walks through these...
Read more >7 models on HuggingFace you probably didn't know existed
For most of the people, “using BERT” is synonymous to using the version with weights available in HF's transformers library.
Read more >Support for Hugging Face Transformer Models
Use the SageMaker model parallelism library's tensor parallelism for training the Hugging Face Transformer models: GPT-2, GPT-J, BERT, and RoBERTa.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I’d say it’s also related to https://github.com/huggingface/huggingface_hub/issues/824
I think this might be a bit of an overthinking;
push_to_hub_keras
is not as widely know astransformers
push_to_hub
, and everywhere intransformers
docs we showpush_to_hub
, so I don’t think people would try usingpush_to_hub_keras
for Transformers models.EDIT: I do see how the two different use of config can cause confusions, although we also do that with the PT mixin I think.