NREL / reV

Renewable Energy Potential (reV) Model
https://nrel.github.io/reV/
BSD 3-Clause "New" or "Revised" License
107 stars 24 forks source link

Cloudify reV #335

Closed MRossol closed 2 years ago

MRossol commented 3 years ago

Why this feature is necessary: Enable the reV team and general public to run reV at scale in the cloud using OEDI datasets as inputs.

A possible solution is: Current cloud readiness of reV's modules:

Potential performance issue:

Cloud hardware / submission ideas:

Charge code reV

Urgency / Timeframe FY22 design Doc

grantbuster commented 3 years ago

Notes and stuff: Configure aws hpc: https://www.hpcworkshops.com/03-hpc-aws-parallelcluster-workshop/04-configure-pc.html Ssh into login node and squeue access: https://www.hpcworkshops.com/03-hpc-aws-parallelcluster-workshop/07-logon-pc.html Sbatch commands just like eagle: https://www.hpcworkshops.com/03-hpc-aws-parallelcluster-workshop/08-run-1stjob.html high performance file storage and transfer to s3 (not even really sure if necessary if reading from s3/hsds?): https://www.hpcworkshops.com/04-amazon-fsx-for-lustre.html

In this lab, the cluster has 0 compute nodes when starting and maximum size set to 8 instances. AWS ParallelCluster will grow and shrink between the min and max limits based on the cluster utilization and job queue backlog.

A GP2 Amazon EBS volume will be attached to the head-node then shared through NFS to be mounted by the compute nodes on /shared. It is generally a good location to store applications or scripts. Keep in mind that the /home directory is shared on NFS as well.

SLURM will be used as a job scheduler

grantbuster commented 2 years ago

PR #339

grantbuster commented 2 years ago

Implemented here: https://github.com/NREL/reV/tree/main/examples/aws_pcluster

Still need work to upload data for exclusions / transmission costs but no technical barriers for that.