LXMERT pre-training tasks
See original GitHub issue❓ Questions & Help
Hello, congrats to all contributors for the awesome work with LXMERT! It is exciting to see multimodal transformers coming to hugginface/transformers. Of course, I immediately tried it out and played with the demo.
LXMERT pre-trained model, trained on what exactly?
Question:
Does the line lxmert_base = LxmertForPreTraining.from_pretrained("unc-nlp/lxmert-base-uncased")
load an already pre-trained LXMERT model on the tasks enumerated in the original paper “(1) masked crossmodality language modeling, (2) masked object prediction via RoI-feature regression, (3) masked object prediction via detected-label classification, (4) cross-modality matching, and (5) image question answering.” (Tan & Bansal, 2019)? If the pre-training tasks are not all the ones from the paper, would that line load pre-trained weights at all and if yes, on what?
Thanks in advance! 🤗
A link to original question on the forum/Stack Overflow: Here is the link to the hugginface forum.
Issue Analytics
- State:
- Created 3 years ago
- Reactions:1
- Comments:23 (7 by maintainers)
Is there any entry-level example of Lxmert? Following example from Lxmert.
comes up
Hi, “unc-nlp/lxmert-base-uncased” was trained with all tasks specified in the paper (as aforementioned). We have benchmarked the pre-trained model to make sure it reaches the same performance on all QA tasks. If you do run into any troubles though, please let me know!