-
Notifications
You must be signed in to change notification settings - Fork 68
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Ia der term #1137
base: master
Are you sure you want to change the base?
Ia der term #1137
Conversation
Pull Request Test Coverage Report for Build 7133713817
💛 - Coveralls |
@CarterDW it seems like flake8 is failing, and the coverage (i.e. lines of code tested by new unit tests) has fallen a bit. I may be able to look at the actual implementation towards the end of the week. |
Attempted reformatting to fit flake8 format
Attempted to better fit flake8 formatting requirements
Im not sure about the coverage but I went through and reformatted the code so hopefully it should pass flake8 now |
Changed the statements from lines 192-203 so that it fits flake8
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks a lot @CarterDW and @jablazek (and apologies for the delay!). Just a couple of comments.
Also, we probably want to add a benchmark test, the same way we did for the previous FastPT terms.
BTW, I remember before FastPT couldn't compute all the galaxy x IA perturbative terms. Is that still the case? I remember we have a TODO for this :-)
# ak power spectrum | ||
pksa = {} | ||
if 'nonlinear' in [self.ak2_pk_kind]: | ||
pksa['nonlinear'] = np.array([cosmo.nonlin_matter_power( | ||
self.k_s, a)for a in self.a_s]) | ||
if 'linear' in [self.ak2_pk_kind]: | ||
pksa['linear'] = np.array([cosmo.linear_matter_power(self.k_s, a) | ||
for a in self.a_s]) | ||
if 'pt' in [self.ak2_pk_kind]: | ||
if 'linear' in pksa: | ||
pka = pksa['linear'] | ||
else: | ||
pka = np.array([cosmo.linear_matter_power(self.k_s, a) | ||
for a in self.a_s]) | ||
pka += self._g4T*self.one_loop_dd[0] | ||
pksa['pt'] = pka | ||
self.pk_ak = pksa[self.ak2_pk_kind] | ||
|
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Maybe I missed something, but wouldn't you achieve the same just adding self.ak2_pk_kind
to the previous if-else statement, and then doing
self.pk_ak = pks[self.ak2_pk_kind]
?
usefptk (::obj:`bool`):) if `True``, will use the FASTPT IA k2 | ||
term instead of the CCL IA k2 term |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Why not require that people have the latest version of FastPT? My first reaction is that this adds unnecessary complexity that will need to be deprecated afterwards anyway (with the headache it entails, since it changes the API).
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think there are two issues. One, for the basic k^2 term, we don't need FAST-PT, and we want the flexibility to choose the P_delta from the options CCL has access to. There may be future versions of this term that need some calculation to be done in FAST-PT.
Two, more generally, do we want to demand the latest (or fairly recent) FAST-PT for all CCL ept users? Probably good in the long term. Right now, FAST-PT is also changing rapidly, and is actually a bit behind in some functionality, so we don't expect all CCL users to actually have access to the bleeding edge.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
OK, I see, thanks @jablazek .
Then, how about:
- Remove the
usefptk
option from this function. - Allow for
ak2_pk_kind
to take an additional option (e.g.'pt_IA_k2'
), in which case the calculator uses the fast-pt prediction.
Then, if users select this option, you check if the fast-pt version allows for it and throw an error if it doesn't. My main aim here is not polluting the API with arguments that we may need to deprecate soon-ish, since they become a headache.
Thanks @damonge! (and Happy New Year!) Indeed, the nl bias x IA terms have been a goal for a long time, and we now have them! Thanks to some work from @CarterDW , as well as from the UTD group and a former student working with me, we have a couple of independent FAST-PT implementations of (most of) these terms. There is an active dev branch of FAST-PT where we are doing the cross checks. Regarding benchmarks, we can generate the Pks from FAST-PT at a given cosmology and redshift (or two) and then store as a static dat file for comparison. It's not independent, since they both use FAST-PT, but it does test the plumbing and API. |
Great news! Happy to have those terms in CCL when you think they're ready! Regarding benchmarks: yep, that sounds good. I agree it's not independent, but it's a good test nevertheless, in case we modify anything further down the line (e.g. interpolation/extrapolation schemes or whatever) that changes the accuracy of the calculation. This is what we did for all the other fast-pt terms (in fact, you can probably build on this benchmark test script and on the script that was used to generate the associated data from fastpt). |
(and Happy New Year!) |
Apologies for not seeing this email before, for the moment I think we’re
shelving this PR, as we work to also add in galaxy bias x IA terms and try
and fix some strange results, thanks!
…On Tue, Jun 18, 2024 at 3:49 AM David Alonso ***@***.***> wrote:
@CarterDW <https://github.com/CarterDW> @jablazek
<https://github.com/jablazek> can I check what the status is? Thanks!
—
Reply to this email directly, view it on GitHub
<#1137 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/BBRHUDGLPX4TCY4FUUGXXHDZH7YA3AVCNFSM6AAAAABJPR7G2OVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDCNZVGU2TINBTGU>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
OK, thanks! |
Hi @damonge . Sorry for the slow updates. A bit more info: we did our code comparison between the two implementations of the NL bias x IA terms. Most of them agreed, with two exceptions. We have been tracking down the differences and should have that resolved soon. However, @CarterDW , is there a reason to not finish the PR for the derivative term and then do a separate one for the NL bias terms and the t_ij terms? |
The main concern I would have is the behavior of the derivative term, since
that is one of the terms I’ve been looking at as the cause of some
divergences in the tracers when it is set to a non zero value.
I will try and do some more debugging this week and if I can sort that
then I agree, doing just the derivative term should provide no issues.
…On Tue, Jun 18, 2024 at 6:48 AM Jonathan Blazek ***@***.***> wrote:
Hi @damonge <https://github.com/damonge> . Sorry for the slow updates. A
bit more info: we did our code comparison between the two implementations
of the NL bias x IA terms. Most of them agreed, with two exceptions. We
have been tracking down the differences and should have that resolved soon.
However, @CarterDW <https://github.com/CarterDW> , is there a reason to
not finish the PR for the derivative term and then do a separate one for
the NL bias terms and the t_ij terms?
—
Reply to this email directly, view it on GitHub
<#1137 (comment)>, or
unsubscribe
<https://github.com/notifications/unsubscribe-auth/BBRHUDFVWY7MXLRFPUM22S3ZIANA7AVCNFSM6AAAAABJPR7G2OVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDCNZVHEYDQNBSHE>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
@CarterDW , we sorted out the unexpected high-k behavior , right? |
OK, great! |
I implemented in pyccl/ept.py a new IA k2 derivative term, adding it to the existing auto and cross correlations with a ck constant defined in pyccl/tracers.py based on the pre-existing c1 constant. The user can either choose to use an in-built CCL k2 term, or call FAST-PT and use the newly implemented FAST-PT derivative term. This is set to use CCL's term by default so older versions of FAST-PT can still be used with ept.py.
I also added in pyccl/ept.py functionality to check if the installed FAST-PT can be imported, and then checks if the installed FAST-PT has the newest derivative term if the user wishes to use the FAST-PT derivative term