Commands for Seeker Training on Dialog GPT2 or T5?
See original GitHub issueHey I was wondering what the set of commands are to get parlai to train t5 or GPT2 to train with seeker parameters where the search query is the copy task. There seems to be a lot of options, was hoping there would be a way to combine the huggingface parameters and the seeker training parameters. This is what i tried
parlai train_model -m hugging_face/dialogpt --add-special-tokens True --delimiter '\n' --add-start-token True --gpt2-size medium --t projects.seeker.tasks.knowledge,projects.seeker.tasks.dialogue,projects.seeker.tasks.search_query -bs 2 -mf microsoft/DialoGPT-medium
just wondering how it knows what to do in the search task and the number of documents it looks up
Issue Analytics
- State:
- Created a year ago
- Comments:10 (4 by maintainers)
Top Results From Across the Web
DialoGPT - Hugging Face
We present a large, tunable neural conversational response generation model, DialoGPT (dialogue generative pre-trained transformer).
Read more >Hugging Face — ParlAI Documentation - GPT2
To use GPT2, run your command with the flag: -m hugging_face/gpt2 . ... The T5 model in ParlAI is based on the T5ForConditionalGeneration...
Read more >Exploring Pre-trained Model Use Cases with GPT-2 and T5
What are pre-trained deep learning models, and why are they setting a new standard in computing? This article explores the possible use cases...
Read more >Top 291 resources for gpt models - NLP Hub - Metatext
This is a complete list of resources about Gpt Models for your next project in natural language processing. Found 291 Gpt. Let's get...
Read more >Releases · facebookresearch/ParlAI - GitHub
A framework for training and evaluating AI models on a variety of openly available dialogue datasets. - Releases · facebookresearch/ParlAI.
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found

sorry, the fix here is to edit
projects/seeker/tasks/__init__.pyto have the following line:Not necessary! To clarify:
projects.seeker.tasks.dialogueis actually a multi-task wrapper over several of the dialogue tasks. Because you want to multitask with dialogue and knowledge tasks, you can add special syntax (:mutators=my_mutator) to apply mutators (which are a way to “mutate” the data) to only certain tasks; since the dialogue and knowledge tasks require different mutators, you will need to specify each task manually (you can take a look at the respectiveDefaultTeachers in theprojects/seeker/tasks/*files to see which teachers are being multitasked).this is just to separate from the standard tasks which assume an encoder/decoder model; our BB3 3B model is a FiD-style model, which requires an encoder-decoder architecture, whereas in decoder-only models deal we treat retrieved documents as simply part of the context. The
DecoderOnlyjust means that the documents are pre-loaded into the context