[FEATURE] Hub + Hugging Face Tutorial
See original GitHub issue🚨🚨 Feature Request
- Related to an existing Issue
- A new implementation (Improvement, Extension)
If your feature will improve HUB
Along with computer vision, there is significant interest in NLP from the Hub community, particularly around transformers like BERT.
Description of the possible solution
End to end tutorial with a small dataset such as CoLA.
Issue Analytics
- State:
- Created 3 years ago
- Comments:8 (8 by maintainers)
Top Results From Across the Web
Hugging Face Hub documentation
The Hugging Face Hub is a platform with over 60K models, 6K datasets, and 6K demos in which people can easily collaborate in...
Read more >Quick start - Hugging Face
Quick start. The Hugging Face Hub is the go-to place for sharing machine learning models, demos, datasets, and metrics. huggingface_hub library helps you ......
Read more >Share a model - Hugging Face
In this tutorial, you will learn two methods for sharing a trained or fine-tuned model on the Model Hub: Programmatically push your files...
Read more >Quick tour - Hugging Face
If you're a beginner, we recommend checking out our tutorials or course next for more in-depth explanations of the concepts introduced here.
Read more >Transformers, what can they do? - Hugging Face Course
The Transformers library provides the functionality to create and use those shared models. The Model Hub contains thousands of pretrained models that anyone ......
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
I think huggingface transformers now support special tokens because I tried passing some into the tokenizer and the tokenizer mapped them to token ids without error. On decoding those token ids I got the special characters back.
This Google Colab Notebook is my attempt at creating the tutorial using the CoLA dataset. Suggestions from the community are highly welcome