-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Regression performance #4
Comments
Adding more variables already improved these results significantly. See new plots below (again, for |
This is very cool Orel!! It's funny you were able to improve performance so quickly! Regarding the bias: weird! Although to tell you the truth, I'm not so worried about it, as what we want to do is rank the events... So biases will not really be a problem. Although lets keep an eye on it, as it could be the cause of some mistake somewhere... As soon as I'm back from vacation I will start playing around with the new code, and let you know if I'm able to come up with anything relevant. |
I think we can clearly see this bias in the confusion matrix in #2 (pasted below for convenience). The events in the lower left cells are a direct result of this bias. It would be great to figure out why it's there and fix it. |
Are you sure this is a result of the bias? From a very naive point of view: If you increase that bias by a factor 10, but keep the shape of the "blob" in the confusion matrix identical, then we would get exactly the same predictions (when you rank the events, the only thing that matters is the shape of that blob, not where it is with respect to the dashed line you show). Or perhaps I'm not properly understanding the effect of the bias? (you have thought more about this than me!) |
Oh, now I realise the mistake. I defined the event_type bins based on the true angular error, but I should actually define it based on the reconstruction angular error. Defining it based on the true value results in the bias translating to that lower edge cell. If we define it from the reco value (which is what we should do!) then I think you are right and that the bias will not have an effect. |
Calculated the regression performance with the Prod5 sample, using the baseline south array, on source gamma, pointing north, with minimum multiplicity of 2 telescopes. Full file name is Regression results are below for the MLP_small model (didn't try any others for now). The scores plot compares the results for all variables and for cases where we exclude the two new "promising" variables. I had hoped that those variables will provide an improvement in performance, but they don't. However, we anyway get a significant boost in performance with this new sample. Possible reasons for this are:
In particular I am happy to see we don't have a bias anymore. Not sure if an increase in statistics can be the reason for that, but maybe. |
Hi Orel, This really looks cool! Statistics definitely seem to play a role here indeed... I do have access to the path you linked, so I can try to copy these files and play around with them. One question: When you say statistics are larger, is it because more MCs were produced? Or because when including multiplicity 2 the statistics went up? Should we try to move to generating IRFs? I don't mind playing around with that... I guess Max will be happy to help! Best, |
Not entirely sure. Without digging into the previous file and testing the cuts, I can't answer whether that file already includes only >4 telescope multiplicity or not. It might be the case because the Prod3b file is significantly smaller (3.3GB vs. 14GB).
Yes, producing IRFs would be good. It would help answer also the question regarding the number of events. You can start with that if you wish and I will do the 2-3 things missing with the regression/classification. |
Added a few more variables to improve performance. Namely, we have a few variables which are given per telescope, like size, distance, crossing, asymmetry, etc. So far we only used the average of those values. However the distribution of these variables is also important of course. So, in order to add some additional information, I added also the median and the std of each distribution of sizes, distance, crossing, etc. This resulted in a significant improvement in performance. Below I put plots for the comparison of scores of each model,
The model with all characteristics is clearly the best, so we will stay with it. The confusion matrix is also significantly better and I think now we can clearly have 3 event types. |
Just for future reference: the plots shown from Feb 11, 2021 were artificially good, because we were adding too much information to the training (if I remember correctly, we were using point-like gammas as training? Can't remember...). We will update these plots and perhaps close the issue. |
That is true, the plots here were made using point-like gammas. I am not sure I would consider this adding too much information, but indeed we most likely cannot reach this level of performance with diffuse-like gammas. |
Performance isn't great at the moment. At low energies the prediction is quite poor. At higher energies we see some improvement, but it's still might not be good enough (see plots below).
Questions/ideas:
Plots below are for
gamma_onSource.S.3HB9-FD_ID0.eff-0.root
.The text was updated successfully, but these errors were encountered: