OpenAIVae implementation
See original GitHub issuedo I see it correctly that the code fragments provided by OpenAI and the way you binded it in the vae.py file means that there is no actual codebook in the form of an explicit nn.Paramter or nn.Embeddings but the very first layer of the decoder serves as the vocabulary?
(decoder): Decoder(
(blocks): Sequential(
(input): Conv2d(n_in=8192, n_out=128, kw=1, use_float16=False, device=device(type='cpu'), requires_grad=False)
this would explain why I couldnt finy any oO
Issue Analytics
- State:
- Created 3 years ago
- Comments:27 (4 by maintainers)
Top Results From Across the Web
OpenAI's ChatGPT shows why implementation is key with ...
OpenAI has already made waves with its DALL-E image generation technology, and its GPT series has drawn attention with each successive release ( ......
Read more >lucidrains/DALLE-pytorch - [REPO]@Telematika
Implementation / replication of DALL-E (paper), OpenAI's Text to Image Transformer, in Pytorch. It will also contain CLIP for ranking the ...
Read more >github.com-lucidrains-DALLE-pytorch_-_2021-04-19_12-38-39
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, ... vae = OpenAIDiscreteVAE() # loads pretrained OpenAI VAE.
Read more >DALL-E in Pytorch download | SourceForge.net
Implementation / replication of DALL-E, OpenAI's Text to Image. ... In contrast to OpenAI's VAE, it also has an extra layer of downsampling, ......
Read more >365 Papers - Jordan Matelsky
BB-Graph is a new subisomorphism search algorithm that improves upon GraphQL and Cypher implementations. [ Read More ].
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
@CDitzel In pytorch implementation, they seem to be directly adding the logits to the sample from gumbel distribution.
I believe they divide the logits by temperature before sampling from categorical distribution in Figure 1 of the paper.
all right, so this is what I have come up with so far. It closely resembles Lucids implementations but parameterizes the gumbel softmax with the distance of the encoder output (logits) to the codebook vectors (described in this paper) and akin to VQ-VAEs, but in contrast to Lucids implementation which uses the logits directly as input to the Gumbel. Phils (and Karpathys implementation) never worked for me when I rightfully included the KL loss i.e. a kl loss > 0. With this implementation the KL loss can be included as it should with a uniform prior. However, the results on a larger data set are still underwhelming and not really satisfying in terms of reconstruction quality. Maybe someone can take a look at it and assess the correctness of this implementation?