maho3 / ltu-gobig

1 stars 0 forks source link

Quijote 1 Gpc/h inference #22

Open maho3 opened 3 months ago

maho3 commented 3 months ago

This issue is solely to keep notes for analyzing 1 Gpc/h inference using Quijote 1 Gpc/h ground-truths.

Notebook: https://github.com/maho3/ltu-gobig/blob/run_experiments/experiments/quijotelike/inference_quijote.ipynb

maho3 commented 3 months ago

Configuration

P(k) monopole coverage

image

PlotSinglePosterior

image

Posterior Coverage

image image image

Notes

bwandelt commented 3 months ago

is this because CHARM was trained on fastpm sims rather than borgpm?

maho3 commented 3 months ago

Yes, that is the case, and would explain the power bias of the halo field. By giving it a z=99 field with too much power (as in the BORG case), the halo emulator would infer less evolution, and thus less clustering.

bwandelt commented 3 months ago

that would explain why sigma_8 is underestimated. But omega_m seems to be more robust...

maho3 commented 3 months ago

I discovered today that this^ was probably not primarily driven by a CHARM error, but instead because CHARM halos were trained to recover the statistics of Rockstar halos in Quijote. However, we were using the Quijote FoF catalog as a reference, as in https://github.com/maho3/ltu-gobig/blob/main/scripts/preprocess_existing_sims/quijote.py

I am currently converting the Quijote Rockstar catalog to our framework, for a better comparison

maho3 commented 3 months ago

This was the issue! Testing now on the Quijote Rockstar catalogs, we have:

Configuration

Coverage

image

Loss

image

$k_{\rm min} = 0.1$ inference

image image image

$k_{\rm min} = 0.2$ inference

image image image

$k_{\rm min} = 0.4$ inference

image image image

$k_{\rm min} = 0.6$ inference

image image image

Notes

maho3 commented 3 months ago

Full hod inference at kmin=0.2! This is done training on 10 hod params per fastpm simulation, for 2000 simulations.

All the test points are quijote halos populated with the fiducial hod configuration, hence the small range of scales in the full coverage plots.

There's still something odd going on with the sigma8 predictions, I assume because the model is picking up at the CHARM discrepancy at high-k, like previously. I assume when we remedy charm, the sigma8 and HOD constraints will improve.

download.png

download (1).png

tlmakinen commented 1 month ago

From the LtU meeting comparing Pinnochio and CHARM Pk(halos) inference on Quiojote DM test set. Comparing different k_max cuts

image image

tlmakinen commented 1 month ago

some poor coverage tests for CHARM (albeit for incorrect min_logM setting):

for k_max=0.05 (not great):

Screenshot 2024-10-04 at 16 42 20 Screenshot 2024-10-04 at 16 43 21

for k_max=0.2 (worse):

Screenshot 2024-10-04 at 16 42 38 Screenshot 2024-10-04 at 16 43 51
tlmakinen commented 1 month ago

Did a cross-check of the analysis and re-did the coverage tests for the CHARM simulations at different k_max cuts. The test suite here is comprised of 500 Quijote simulations.

for k_max=0.05: image image image image


for k_max=0.2 image image image image

tlmakinen commented 1 month ago

for reference, we can check inference coverage of pinnocchio tested on quijote to see how cosmology is captured on small scales

for k_max=0.05: image image

remark: pinnocchio does ok on large scales

for k_max=0.2

NOTE that the coverage here failed (unable to sample properly over the full test suite, which implies that pinnochio lines up poorly with quijote). A sketch of the poor coverage for 88 test simulations:

Screenshot 2024-10-07 at 16 25 16

image

remark: pinnocchio does poorly on smaller scales