popgenmethods / pyrho

Fast inference of fine-scale recombination rates based on fused-LASSO
MIT License
44 stars 6 forks source link

Breaking up make_table into multiple files #26

Open uqmzhou8 opened 1 year ago

uqmzhou8 commented 1 year ago

Hi,

Apologies as I am relatively new to this population genetics work. I am trying to compute a lookup table for n=450 and N=600 but it is taking too long to run on a high performance computing cluster, exceeding the time limit even when using multiple threads. Therefore, I am trying to figure out how to construct multiple tables and join them into a complete lookup table or if there is a method to pause, save and continue at some predefined checkpoints?

Based on pyrho make_table --help, there are two arguments that seem to help me in my case as shown below: -S STORE_STATIONARY, --store_stationary STORE_STATIONARY Name of file to save stationary distributions -- useful for computing many lookup tables sequentially. -L LOAD_STATIONARY, --load_stationary LOAD_STATIONARY Name of file to load stationary distributions -- useful for computing many lookup tables tables sequentially.

Are these the correct arguments to use for storing information at different checkpoints and would you have any examples on how I can use this to build my way to n=450 or even more? Please suggest other methods that I can use to build a lookup table of n=450 or more but spread into smaller files/jobs.

Thank you in advance!

Regards, Charles

jeffspence commented 1 year ago

Hi Charles,

Sorry to hear you're running into issues. I don't have a good solution to your problem, but here are some thoughts in the hopes that they're helpful:

I hope this helps!

Jeff