-
Notifications
You must be signed in to change notification settings - Fork 41
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Implementation of the MVAE model #80
Comments
@miguelsvasco
Yes. this loss function comes from not
This might be due to not training "unimodal" inferences of the PoE encoder, q(z|x) and q(z|y). Without it, inferred z from unimodal input (especially label or attribute) might be collapsed (a similar issue is also referred to our preprint paper as the "missing modality difficulty").
Yes, but you should use the Given your comments, I replaced the name of the previous notebook from Thank you! |
@masa-su |
@sgalkina For an example of the usage, please see the implementation of the M2 model, which is the well-known semi-supervised VAE model. If you have any trouble understanding how to use it, please feel free to ask! |
First of all thank you for the code! Just two slight remark:
However, the original formulation of the MVAE model (in the paper Multimodal Generative Models for Scalable Weakly-Supervised Learning), does not consider such terms, only a KL divergence term between the distribution of the POE encoder and the prior:
When I remove the kl_x and kl_y terms from the regularizer and train, the model seems unable to perform cross-modality inference:
The text was updated successfully, but these errors were encountered: