Why change sequence order of prev_output_tokens in BART?
See original GitHub issueHi,
In the extract_features
example of BART (link to code), both src_tokens
(for encoder) and prev_output_tokens
(for decoder) are fed into model, but prev_output_tokens
is just the src_tokens
with EOS being moved to the beginning. I think during training (teacher forcing) the prev_output_tokens
starts with BOS, doesn’t it?
So what’s the purpose of this reorder? To add some deliberate noise, or assuming there’s another sentence before it?
Thanks in advance.
Rui
Issue Analytics
- State:
- Created 4 years ago
- Comments:16 (8 by maintainers)
Top Results From Across the Web
BART schedule changes 9/12/22 with several improvements
On September 12, 2022, BART's schedule is changing to provide better spaced apart trains, add earlier train options, end scheduled single ...
Read more >System Facts - BART.gov
Additionally, there are over 80 bill-to-bill change machines located throughout the BART system. These machines break $10 and $20 bills into $5 denominations, ......
Read more >BART schedule change begins 2/14/22, extending service to ...
BART's schedule will change on Monday, February 14, 2022, with significant improvements to Sunday service.
Read more >Frequently Asked Questions (FAQs) - BART.gov
A: BART communicates cancelled trains in advance on our website and the official app in the station specific “real time departures,” “schedule by...
Read more >Using BART - BART.gov
Welcome to BART!Bay Area Rapid Transit (BART) connects the San Francisco Peninsula with communities in the East Bay and South Bay.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
https://github.com/pytorch/fairseq/blob/master/fairseq/data/language_pair_dataset.py#L63 is the fixed link. Ahh I see, so actually the input format to encoder for above case is actually:
and Decoder input:
And target:
We will look into updating those figures. Thanks for pointing it out.
@Colanim @villmow