EiffL / sfh-inference-hackathon

Repository for SFH inference hackathon at AstroInfo 2021
MIT License
7 stars 2 forks source link

Dataset cleanup #26

Open EiffL opened 2 years ago

EiffL commented 2 years ago

Main dataset: TNG100

The proposal is to check that all the data we care about is in TNG100, and move the other datasets into a legacy folder.

ppfn commented 2 years ago

In order to use the shh/datasets/tng100 dataset:

from sfh.datasets.tng100 import tng100
nicolas-cerardi commented 2 years ago

tng100 implements interpolation for sfh dset - no need to implement the simpler sfh dset. However, tng100 does not have the "Mask" tensor (which is in sfh_interp), which flags 1 if the value is real and 0 if it's interpolated. Not sure that "Mask" is really necessary though.

EiffL commented 2 years ago

Ok then we can proceed most likely to just removing the datasets that are not necessary, and only keep sfh/datasets/tng100 . @yannick1974 if you want to open a PR for this cleaning, feel free to do so :-)

yannick1974 commented 2 years ago

I propose this to clean things up:

Also, I'd like to use environment variable for the location of data files, using the IDRIS variables when appropriate. This will make it easier to run the notebooks outside of Jean-Zay.

What do you all think about that?

EiffL commented 2 years ago

Yep that sounds good!

The only thing I usually do is that the dataset directory is a submodule of the code directory, which is itself a pip installable module.

yannick1974 commented 2 years ago

I can do like that too.