Problems of IWAE ELBO Loss
See original GitHub issueHi Anand and all,
As weighting of samples, weight
should be detached from the current computational graph for the expected optimization objective, right? See
https://github.com/AntixK/PyTorch-VAE/blob/8700d245a9735640dda458db4cf40708caf2e77f/models/iwae.py#L155
Issue Analytics
- State:
- Created 2 years ago
- Reactions:1
- Comments:5
Top Results From Across the Web
How I learned to stop worrying and write ELBO (and its ...
Here, I will talk about the formultaion of IWAE and its 3 benefits: tighter lower bound estimate, importance-weighted gradients and complex ...
Read more >Fixing a Broken ELBO - arXiv
However, the fundamental problem is that these loss functions only depend on p(x|θ), and not on p(x, z|θ). Thus they do not measure...
Read more >Importance Weighted Autoencoders and Jackknife Methods
The right-hand side is called the evidence lower bound (ELBO), denoted by L. The inference problem then becomes an optimization problem that ...
Read more >CS 6785 Homework 2 | Canvas@Cornell
Problem 1: Implementing the Variational Autoencoder (VAE) (25 points) ... negative ELBO, (2) KL term, and (3) reconstruction loss as ...
Read more >Importance Weighted Autoencoders - DeepAI
We present the importance weighted autoencoder (IWAE), ... Tutorial: Deriving the Standard Variational Autoencoder (VAE) Loss Function.
Read more >Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start FreeTop Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
Top GitHub Comments
Besides, as the original paper said, “Vanilla VAE separated out the KL divergence in the bound in order to achieve a simpler and lower-variance update. Unfortunately, no analogous trick applies for k > 1” (Y. Burda et al., 2016). How are we still able to compute KL Divergence? https://github.com/AntixK/PyTorch-VAE/blob/8700d245a9735640dda458db4cf40708caf2e77f/models/iwae.py#L152
Kindly refers to PR: https://github.com/AntixK/PyTorch-VAE/pull/53