question-mark
Stuck on an issue?

Lightrun Answers was designed to reduce the constant googling that comes with debugging 3rd party libraries. It collects links to all the places you might be looking at while hunting down a tough bug.

And, if you’re still stuck at the end, we’re happy to hop on a call to see how we can help out.

Support for GPT-J

See original GitHub issue

Thanks for the great repo! I have tried it out, it’s really amazing to lead such a large model in multiple GPUs.

Describe a requested feature

Currently, GPT-J is supported only in HF 4.7.0 and by installing

pip install git+https://github.com/finetuneanon/transformers@gpt-j

In your requirement, there is HF 4.8.0, and needs to load several new models. Soon gpt-j will be fully integrated in HF: https://github.com/huggingface/transformers/pull/12243

I am wondering if is there an easy way to have back compatibility, or include GPT-J soon.

Thanks again for your great repo 👍🏻

– Andrea

Issue Analytics

  • State:closed
  • Created 2 years ago
  • Comments:11 (6 by maintainers)

github_iconTop GitHub Comments

2reactions
hyunwoongkocommented, Dec 5, 2021

We added GPTJ.

2reactions
andreamad8commented, Jul 19, 2021

Yeah, that’s was I thought.

I suggest adding this info in the README, for the people (like me 😃) that are not familiar with Megatron LM.

Read more comments on GitHub >

github_iconTop Results From Across the Web

GPT-J - Hugging Face
The GPTJ Model transformer with a language modeling head on top. This model inherits from FlaxPreTrainedModel. Check the superclass documentation ...
Read more >
How you can use GPT-J - Towards Data Science
Access GPT-J, a 6 billion parameter Natural language processing model. 3 ways how to inference it: browser, colab and local with huggingface ...
Read more >
How You Can Use GPT-J - forefront.ai
With its main commercial use case to help users by providing answers to their questions, chatbots are commonly used in a variety of...
Read more >
gptj - PyPI
gptj 3.0.9. pip install gptj. Copy PIP instructions. Latest version ... Status: Service Under Maintenance. Developed and maintained by the Python community, ...
Read more >
Deploying GPT-J and T5 with NVIDIA Triton Inference Server
FasterTransformer/examples/pytorch/gptj/utils/gptj_ckpt_convert.py to ... some custom layers into default NNs supported by FT (T5, GPT), ...
Read more >

github_iconTop Related Medium Post

No results found

github_iconTop Related StackOverflow Question

No results found

github_iconTroubleshoot Live Code

Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free

github_iconTop Related Reddit Thread

No results found

github_iconTop Related Hackernoon Post

No results found

github_iconTop Related Tweet

No results found

github_iconTop Related Dev.to Post

No results found

github_iconTop Related Hashnode Post

No results found