Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Should KL loss and reconstruction loss be in the same magnitude? #2

Open
zyj008 opened this issue Sep 5, 2018 · 0 comments
Open

Should KL loss and reconstruction loss be in the same magnitude? #2

zyj008 opened this issue Sep 5, 2018 · 0 comments

Comments

@zyj008
Copy link

zyj008 commented Sep 5, 2018

Hello! I have met KL collapsing problem when training VAE model. The KL loss is like this.

image

After reading your paper “Ladder Variational Autoencoders”, I decided to use warm up method for KL loss. But I still have some questions. In my task, the scale of reconstruction loss and kl loss is different greatly. The reconstruction loss is about 10^-210^-1, but the KL loss is 10^-510^-6. The value of mu and log_var are also 10^-5~10^-6 (mu or log_var is calculated by encoder results passing through one FC layer). So I have a question, will the difference between the scale of two kinds of loss strongly influence my training task? Should I do something to increase the value of KL loss before I implement warm up method?

Thank you!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant