question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Sharing Microsoft's DialogRPT (new dialog ranking model)

See original GitHub issue

🌟 New model addition

Model description

Thanks for the awesome work!

DialogRPT (Dialog Ranking Pretrained Transformers) is a set of GPT-2 based dialogue ranking models recently released with an EMNLP paper by Microsoft Research. It’s a follow-up work of DialoGPT (thanks for hosting it!) The architecture is pretty simple: a GPT2Model followed by a torch.nn.Linear(n_embd, 1, bias=False), and implemented based on a previous HuggingFace commit At first, I’m trying to create a model card for it, but then realized that it seems there’s no existing model architecture in HuggingFace is compatible with DialogRPT. I noticed a lot of BERT-based sequence classification models, but ours is GPT-2 based.

If there’s a simple fix (or I missed something) please let me know! If implementation in modeling_gpt2.py is necessary, I’m also glad to help!

Open source status

Issue Analytics

  • State:closed
  • Created 3 years ago
  • Reactions:2
  • Comments:12 (7 by maintainers)

github_iconTop GitHub Comments

2reactions
LysandreJikcommented, Oct 1, 2020

Hi @golsun! GPT2ForSequenceClassification has been implemented on #7501 and I verified that I obtain the same results as you do on your README using your examples.

You should only need to upload your models on the model hub now! Some helpers regarding the configuration:

  • You should upload a model configuration on the hub, for every model.
  • You can simply copy-paste the gpt2-medium configuration that you can find here.
  • You will need to add a num_labels=1 field to these configurations.
  • In the architectures field, you should put GPT2ForSequenceClassification
1reaction
golsuncommented, Oct 7, 2020

thank you @LysandreJik AutoModelForSequenceClassification works now. The inference webpage still gives the Unrecognized configuration class error but I guess it will sync with the latest code soon. I’m going to introduce model card in the original repo. Thanks again for the help!

Read more comments on GitHub >

github_iconTop Results From Across the Web

Dialogue Response Ranking Training with Large ... - Microsoft
We trained DialogRPT, a set of GPT-2 based models on 133M pairs of human feedback data and the resulting ranker outperformed several baselines....
Read more >
microsoft/DialogRPT-depth - Hugging Face
It is a set of dialog response ranking models proposed by Microsoft Research NLP Group trained on 100 + millions of human feedback...
Read more >
MSDialog - Center for Intelligent Information Retrieval
The MSDialog dataset is a labeled dialog dataset of question answering (QA) ... the data to produce MSDialog-ResponseRank for conversation response ranking.
Read more >
Results of the Multi-Domain Task-Completion Dialog Challenge
out the system to new domains with limited in-domain data. With the fast domain adaptation task ... model is shared among the implicit...
Read more >
Response Ranking with Deep Matching Networks and ... - arXiv
search models. To this end, we create a new information-seeking conversation data set MSDialog on technical support dialogs of. Microsoft ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found