Image GPT
See original GitHub issue🌟 New model addition
Model description
OpenAI just announced Image GPT: https://openai.com/blog/image-gpt/
Although image rendering would be out of scope for Transformers, the RGB generation would still be in scope and it would be best to port the weights to a GPT2LMModel
.
However, it’s not immediately clear here how the tokenization is implemented in the downloaded model. (no separate vocab.json
)
Open source status
- the model implementation is available: https://github.com/openai/image-gpt
- the model weights are available: see README above
- who are the authors: @openai
Issue Analytics
- State:
- Created 3 years ago
- Reactions:34
- Comments:5 (1 by maintainers)
Top Results From Across the Web
Image GPT - OpenAI
Generative sequence modeling is a universal unsupervised learning algorithm: since all data types can be represented as sequences of bytes, a ...
Read more >openai/image-gpt - GitHub
This repository is meant to be a starting point for researchers and engineers to experiment with image GPT (iGPT). Our code forks GPT-2...
Read more >ImageGPT - Hugging Face
ImageGPT (iGPT) is a GPT-2-like model trained to predict the next pixel value, allowing for both unconditional and conditional image generation.
Read more >Image GPT | Discover AI use cases - GPT-3 Demo
In 2020, OpenAI released Image GPT (iGPT), a Transformer-based model that operates on sequences of pixels instead of sequences of text.
Read more >OpenAI's Image GPT Completes Your Images With Style!
Check out Weights & Biases and sign up for a free demo here: https://www.wandb.com/papers ❤️ Their mentioned post is available here: ...
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Hey @minimaxir! Here’s a colab which loads the weights into a subclass of
GPT2LMHeadModel
and demonstrates unconditional image generation and conditional image completion.Some differences I’ve found between Image-GPT and GPT2 which are reflected in the subclass.
[n_embd, 3*n_embd]
dimensional linear layer,c_attn
, which produces queries, keys, and values is instead split into 3 separate linear layers each with dimension[n_head, n_embd/n_head, n_embd]
in Image-GPT (this only affects how to load the weights and not the actual model).conv1d
module doesn’t have a bias termSo what’s our next step to add this to the repo?
This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.