What did you change for 'xlm-r-distilroberta-base-paraphrase-v1' training?
See original GitHub issueHi,
what did you change for xlm-r-distilroberta-base-paraphrase-v1
training compared to xlm-r-bert-base-nli-stsb-mean-tokens
?
My tests show that it performs better on the german translated stsb:
model = SentenceTransformer('xlm-r-bert-base-nli-stsb-mean-tokens') #Spearman: 0.8181
model = SentenceTransformer('xlm-r-distilroberta-base-paraphrase-v1') # Spearman: 0.8201
Issue Analytics
- State:
- Created 3 years ago
- Comments:11 (11 by maintainers)
Top Results From Across the Web
Courses for Adapting to Change - Skillsoft
In this course, you will learn the essential skills needed to handle organizational change and the steps to develop a change action plan....
Read more >Changing Behavior After Training - Crucial Learning
Review the skills you think will help you the most, discuss them candidly with your boss, and then tie them into your formal...
Read more >Training For Change | Movement and Capacity Building at the ...
We're dedicated to transformation – beyond training new skills and tools, we support folks to wage conflict, challenge self-limiting beliefs, practice ...
Read more >Change training: 4 key tools to help you keep on top of ...
Change training: 4 key tools to help you keep on top of Change Management risks · Change Impact Assessment · Capture and track...
Read more >Adjust to Change - Dale Carnegie Training
This course will enable you to take a creative and energetic approach to adapting to change. What You'll Learn. In this two-hour workshop,...
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Hi @PhilipMay A paper is upcoming for the paraphrase models.
These models were trained on various datasets with Millions of examples for paraphrases, mainly derived from Wikipedia edit logs, paraphrases mined from Wikipedia and SimpleWiki, paraphrases from news reports, AllNLI-entailment pairs with in-batch-negative loss etc.
In internal tests, they perform much better than the NLI+STSb models as they have see more and broader type of training data. NLI+STSb has the issue that they are rather narrow in their domain and do not contain any domain specific words / sentences (like from chemistry, computer science, math etc.). The paraphrase models has seen plenty of sentences from various domains.
More details with the setup, all the datasets, and a wider evaluation will follow soon.
@PhilipMay Sadly not yet, it is rather far down in the paper pipeline 😕