-
Notifications
You must be signed in to change notification settings - Fork 36
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
aleatoric uncertainty loss term #1
Comments
Hi. Concerning the Concerning the Eq.(8) in the paper arXiv version seems to suggest that a negative loss is allowed. However, this is my personal understanding. Please feel free to find any fault in it. Best, P. |
Hi @pmorerio , does the variable As indicated in the paper, the To my understanding, the Is there any reason why you didn't Thank you! |
Hi @sdsy888, thanks for your question. The variable Hope this clarifies. |
Oh, I thought it in the wrong direction. I thought eq.8 will first predict variance and then get the real value using |
It seems that in your code of aleatoric uncertainty, mean and log_var are two separate parameters trained in the model. I thought that they should come from the output of the model (i.e. mean means the average of output y, log_var means the log variance of the output y). Why model them as two variables rather than results from y? Thanks! |
Thank you for your answer! I'm a little bit confused that how to 'use a single network to transform the input with its head split to predict both mean as well as variance'? Just replace them with the single output y in the model? Are there any supporting theories? Thanks! |
Sorry I did not really get your questions. |
Sorry for the disambiguated description. I mean that in your code, it seems that you first pass the image through each layer of the network, then decode the output by passing it through the same layers of the network to get mean and variance. Why is that? |
Hi, I hope this answers your question. |
Hi, is there a reason why you did not put activation function for the mu, logvar at the end of the decoder.
I am getting negative values for the loss. Is it okay to have a negative loss for the model??
The text was updated successfully, but these errors were encountered: