How to finetune ELECTRA on glue?
See original GitHub issueAfter pretraining my own electra model, I wanted to test it out in Glue using run_glue.py. However I got this:
ValueError: Unrecognized configuration class <class 'transformers.configuration_electra.ElectraConfig'> for this kind of AutoModel: AutoModelForSequenceClassification.
Model type should be one of DistilBertConfig, AlbertConfig, CamembertConfig, XLMRobertaConfig, BartConfig, RobertaConfig, BertConfig, XLNetConfig, FlaubertConfig, XLMConfig.
After taking a look at the source code, It seems like ElectraConfig isn’t available for sequence classification, is there a reason for that? Did anyone finetune electra on glue?
Issue Analytics
- State:
- Created 3 years ago
- Comments:8 (6 by maintainers)
Top Results From Across the Web
anirudh21/electra-base-discriminator-finetuned-rte
This model is a fine-tuned version of google/electra-base-discriminator on the glue dataset. It achieves the following results on the evaluation ...
Read more >Tutorial: How to pre-train ELECTRA for Spanish from Scratch
At a small scale, ELECTRA-small can be trained on one GPU for 4 days to outperform GPT on the GLUE benchmark. At a...
Read more >Solve GLUE tasks using BERT on TPU | Text - TensorFlow
Load a BERT model from TensorFlow Hub · Choose one of GLUE tasks and download the dataset · Preprocess the text · Fine-tune...
Read more >Pretrain & Finetune MLM - fastai dev - fast.ai Course Forums
Pretrain & Finetune MLM - 6: Reproduce GLUE finetuning results ; ELECTRA-Small, 54.6, 89.1 ; ELECTRA-Small (finetuned with fastai), 52.8, 89.8 ...
Read more >Understanding ELECTRA and Training an ELECTRA ...
All scores are GLUE benchmark scores given in the ELECTRA paper unless stated otherwise. Loss defined over all input tokens vs only masked ......
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
awesome, it works perfectly, thank you very much!
This problem happened again, when I use ELECTRA on question-answering pipeline. My Transformers version is 2.11.0.