question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Community contribution - `BetterTransformer` integration for more models!

See original GitHub issue

BetterTransformer integration for more models!

BetterTransformer API provides faster inference on CPU & GPU through a simple interface!

Models can benefit from very interesting speedups using a one liner and by making sure to install the latest version of PyTorch. A complete guideline on how to convert a new model has been created on the BetterTransformer documentation!

Here is a list of models that could be potentially supported, pick one of the architecture below and let’s discuss about the conversion!

Text models 🖊️ :

Vision models 📷 :

Audio models 🔉 :

Let us also know if you think that some architectures can be supported that we missed. Note that for encoder-decoder based models below, we expect to convert the encoder only.

Support for decoder-based models coming soon!

cc @michaelbenayoun @fxmarty

https://github.com/huggingface/optimum/issues/488

Issue Analytics

  • State:open
  • Created 10 months ago
  • Reactions:7
  • Comments:51 (35 by maintainers)

github_iconTop GitHub Comments

3reactions
younesbelkadacommented, Dec 5, 2022

Hi @ravenouse ! From what I got, this function is a C++ binding of the transformer encoder operation that is first defined here and fully defined here as you can see, the whole transformer encoder operations (self attention + ffn) is defined in a single operation

3reactions
michaelbenayouncommented, Nov 22, 2022

It is not in the list because DebertaV2 does not have a regular attention mechanism, so it is not possible to use it with BetterTransformer.

Read more comments on GitHub >

github_iconTop Results From Across the Web

Adding BetterTransformer support for new architectures
Adding BetterTransformer support for new architectures. You want to add a new model for BetterTransformer API? Check this guideline!
Read more >
[P] BetterTransformer: PyTorch-native free-lunch speedups for ...
Hi everyone, In the latest PyTorch stable release 1.13, the BetterTransformer feature was marked as stable! It is a free-lunch optimization to ...
Read more >
A BetterTransformer for Fast Transformer Inference - PyTorch
These fast paths are integrated in the standard PyTorch Transformer APIs, and will accelerate TransformerEncoder, TransformerEncoderLayer and ...
Read more >
تويتر \ younes (younesbelkada@) - Twitter
Big kudos to anyone who has contributed to this so far How to add more models ⬇️. github.com. Community contribution - `BetterTransformer` integration...
Read more >
BetterTransformer, Out of the Box Performance for Hugging ...
Hugging Face meets PyTorch to integrate 'BetterTransformer' in its ecosystem ... How to contribute and add support for more models.
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found