19 February 2017
notes on (Burda et al., 2016).
understanding: 8/10
code: https://github.com/yburda/iwae
importance weighted autoencoders (iwae) improve variational autoencoders. the main difference is in the loss function. iwae uses: \begin{align} \mathcal L_K = \E_{q_{\phi}(x \given y)}\left[\log \frac{1}{K} \sum_{k = 1}^K w_k\right] \end{align} where \(w_k = p_{\theta}(x_k, y) / q_{\phi}(x_k \given y)\) and \(x_k \sim q_{\phi}(x \given y)\). vae uses the same, but always with \(K = 1\). both objectives are lower bounds of \(p_{\theta}(y)\).
iwae is better because:
@inproceedings{burda2016importance, title = {Importance Weighted Autoencoders}, author = {Burda, Yuri and Grosse, Roger and Salakhutdinov, Ruslan}, year = {2016}, booktitle = {International Conference on Learning Representations (ICLR)} }
[back]