To run on iris and saga (General Atomics clusters):
module load toksearch
and, the first time you run, after you module load toksearch also pip install h5py==3.6.0
. Shouldn't be necessary, but if you have further difficulties consider things like module load defaults
, module load gcc7
, module load hdf5/gnu
For the basic case:
In dump_shots.py edit the min_shots and max_shots to be whatever range you want, for testing use e.g. 163300 to 163310. Then run it to dump the shots you want to collect (but only those with plasma; and you can also edit to only take from certain run days in the file) via
python dump_shots.py
Then collect the signals from those shots. By default, path_to_config above will be configs/example.yaml which has all the signals, draws from data/shots.npy (which dump_shots.py dumps to), and dumps to output_file (MAKE SURE TO EDIT THIS). You can take out signals, e.g. for testing
When the config file is ready, run
python new_database_maker.py path_to_config
For large runs, use launch_parallel_jobs.py
which manually dumps shots and splits into cases to run in parallel (toksearch theoretically can do this under the hood but in my experience it doesn't speed stuff up and is not robust). You can modify combine_shots.py
to combine the h5 files it dumps into one.
As side notes, omfit_run_dump.py can be used for grabbing text data, which is an OMFIT script (see top of file for how to use)