MiniLM: releasing all models?
See original GitHub issueHi there,
First of all: great work on distilling a strong teacher into a well performing student and eliminating the issue of parameter size discrepancy in teacher-student models! I am always happy to see smaller, usable models.
I was wondering if you plan to release the Small MiniLM model (L6xH384). It says We release the uncased 12-layer and 6-layer MiniLM models with 384 hidden size [...]
, but I can only find the link to the 12-layer model.
Thanks so much
Issue Analytics
- State:
- Created 3 years ago
- Comments:5 (3 by maintainers)
Top Results From Across the Web
README.md · sentence-transformers/all-MiniLM-L6-v2 at main
We're on a journey to advance and democratize artificial intelligence through open source and open science.
Read more >MiniLM: Deep Self-Attention Distillation for Task ... - arXiv
Comparison between the publicly released 6-layer models with 768 hidden size distilled from BERTBASE. We compare task-.
Read more >MINILM: Deep Self-Attention Distillation for ... - NIPS papers
Table 2: Comparison between the publicly released 6-layer models with 768 hidden size distilled from BERTBASE. We compare task-agnostic distilled models without ...
Read more >Compatible third party NLP models - Elastic
The Elastic Stack machine learning features support transformer models that ... All MiniLM L12 v2 Suitable similarity functions: dot_product , cosine , ...
Read more >https://raw.githubusercontent.com/microsoft/unilm/...
[Model Release] September, 2022: [**BEiT ... Both English and multilingual MiniLM models are released. "[MiniLMv2: Multi-Head Self-Attention ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hi @WenhuiWang0824
great, thank you for the great work and releasing the model. The MiniLMv1 models work great for bi-encoders and cross-encoders, so I’m eager to test the v2 models.
It would be great if the models could also be added to the huggingface model hub: https://huggingface.co/microsoft
This would make it easy to load and use the models.
Let me know if you need help putting the models on the hub.
Hi @volker42maru and @maksymbevza,
We have released the monolingual and multilingual minilmv2 models distilled from different teachers. Please find the model links in the MiniLM folder.
Thanks