HARPgroup / model_meteorology

0 stars 0 forks source link

PRISM Data #33

Open COBrogan opened 4 months ago

COBrogan commented 4 months ago

Maintained by Oregon State, the PRSIM dataset offers daily, monthly, and normals datasets for precipitation (ppt), mean temp, max temp, min temp, mean dew point, and min/max vapor pressure. From their homepage, "The PRISM Climate Group gathers climate observations from a wide range of monitoring networks, applies sophisticated quality control measures, and develops spatial climate datasets to reveal short- and long-term climate patterns. The resulting datasets incorporate a variety of modeling techniques and are available at multiple spatial/temporal resolutions, covering the period from 1895 to the present." See here for additional documentation and comparison to daymet

PRISM data is available at a 4km resolution. 800 m resolution raster files exist, but we need to purchase these rasters, which will have limits on distribution. For current pricing, see this page on the PRISM website.

PRISM data can be downloaded from the website, through an anonymous FTP site, or through their webservices. The web services is the most secure means to access this data programmatically, but their server traffic limits downloads of the same dataset to just twice so caution is advised when testing. Instructions for using this service can be found here. As noted on the Explorer page, PRISM considers the date downloaded to be the day that ends on the selected date:

"Note: Rather than calculating days based on the 24-hour period ending at midnight local time, PRISM defines a "day" as the 24 hours ending at 12:00 Greenwich Mean Time (GMT, or 7:00am Eastern Standard Time). This means that PRISM data for May 26, for example, actually refers to the 24 hours ending at 7:00am EST on May 26"

Per the PRISM website, data from the past 6 months is provisional. This means they are subject to change as reporting stations finalize their data. Per PRISM,

"The information [for provisional data] is based on whatever data is available from as many of the station networks and data sources as possible, so it changes as new data becomes available or as data quality is improved. The datasets are modeled using climatologically-aided interpolation (CAI), which uses the long-term average pattern (i.e., the 30-year normals) as first-guess of the spatial pattern of climatic conditions for a given month or day."

With PRISM, we need to:

Code/Examples

Import PRISM using meta-model

Inventory PRISM in postgres


#### Verify PRISM in postgres

select met.featureid, to_timestamp(met.tstime) as obs_date, extract(month from to_timestamp(met.tstime)) as mo, round((ST_summarystats(st_clip(met.rast, fgeo.dh_geofield_geom), 1, TRUE)).mean::numeric,5) as precip_kgm3, round(0.0393701 (ST_summarystats(st_clip(met.rast, fgeo.dh_geofield_geom), 1, TRUE)).mean::numeric,5) as precip_in, round(0.0393701 (ST_summarystats(st_clip(met.rast, fgeo.dh_geofield_geom), 1, TRUE)).max::numeric,5) as precip_max_in from dh_feature as f left outer join field_data_dh_geofield as fgeo on ( fgeo.entity_id = f.hydroid and fgeo.entity_type = 'dh_feature' ) left outer join dh_variabledefinition as v on ( v.varkey = 'prism_mod_daily' ) left outer join dh_feature as mcov on ( mcov.hydrocode = 'cbp6_met_coverage' ) left outer join dh_timeseries_weather as met on ( mcov.hydroid = met.featureid and met.varid = v.hydroid and extract(year from to_timestamp(met.tstime)) = 2021 ) where f.hydrocode = 'usgs_ws_01634000' order by met.tstime;

- After running tests on d.alpha which imported the 1st-5th (6th failed due to me downloading it a bunch of times and getting negged by PRISM).
- Returns:
   - note: zero days in early part may be due to bad downloads/imports

featureid | obs_date | mo | precip_kgm3 | precip_in -----------+------------------------+----+-----------------------+------------------------ 616654 | 2021-01-01 19:00:00-05 | 1 | 21.897387758890787 | 0.8621023458063062 616654 | 2021-01-02 19:00:00-05 | 1 | 1.5953182965517043 | 0.06280784086707025 616654 | 2021-01-03 19:00:00-05 | 1 | 1.2646758019924165 | 0.049790412792021635 616654 | 2021-01-04 19:00:00-05 | 1 | 0.014757499457725013 | 0.0005810042294005795 616654 | 2021-01-06 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-07 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-08 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-09 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-10 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-11 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-12 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-13 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-14 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-15 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-16 19:00:00-05 | 1 | 0 | 0 616654 | 2021-01-17 19:00:00-05 | 1 | 0.2218199946578049 | 0.008733075371677244 616654 | 2021-01-18 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-19 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-20 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-21 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-22 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-23 19:00:00-05 | 1 | 0.006076666540563261 | 0.00023923896936862965 616654 | 2021-01-24 19:00:00-05 | 1 | 11.850017201900481 | 0.4665363622405421 616654 | 2021-01-25 19:00:00-05 | 1 | 11.850017201900481 | 0.4665363622405421

COBrogan commented 4 months ago

Draft shell script to download prism data, project it to 4326, and clip it to a user-input mask vector. Inputs should be year and mask vector filepath. See prism branch.

rburghol commented 4 months ago

@COBrogan thanks for pushing your script up there. I’m going to do some parallel work on NLDAS2 today. I will feedback if I have anything that will cross over into PRISM.

COBrogan commented 4 months ago

Thanks @rburghol . It was easy enough to get the raster into the database. Next steps are:

rburghol commented 4 months ago

@COBrogan awesome, check out this issue I just added with respect to the ins and outs of consistent date and time handling #35

rburghol commented 4 months ago

@COBrogan - wuestion, is $maskExtent still a file, or is that a WKT string?

COBrogan commented 4 months ago

I forsaw us passing in a file (maybe csv) containing the WKT string. I think -cutline is pretty flexible so we can pass a number of different file types here

rburghol commented 4 months ago

Note - the final raster size after clipping with your file is 426k, which is down from 9 megs from the original tif (after reprojecting). This will make a big difference in data storage.

COBrogan commented 4 months ago

@rburghol that's good news! The clipping is fairly fast too. I've update the script to start getting this data into the database. If we have time tomorrow or Monday, I'd like to sync up on this download/insert script. I have the right idea, but my current code doesn't feel like the approach I should be taking. I've noticed that dh_update_timeseries_weather can probably do some of this work for me, but I thought we were trying to get this done only in shell. Or was it only the downloads we were concerned with?

Either way, I'm getting closer. I also want to learn how to call dh_handletimestamp. I'm guessing this is as easy as calling the module, but I may need directions on where to find that.

rburghol commented 4 months ago

I think that we need to take a crack at developing timestamp handling via shell as it will reduce complexity and enhance modularity for batch automation. We can verify it against the PHP and in the end if the shell method can’t hang, we can isolate the function from PHP and call it from the shell. We can compare notes tomorrow if you have like 30 minutes.

COBrogan commented 3 months ago

I'm updating the scripts such that the download + processing are handled by a function and the import to the database is handled by a separate function. My intention is that the database import function, addRasterToDBase2.sh, is generic enough to be used by all of our data import processes.

durelles commented 3 months ago

[like] Scott, Durelle reacted to your message:


From: C. Brogan @.> Sent: Tuesday, May 21, 2024 8:43:50 PM To: HARPgroup/model_meteorology @.> Cc: Subscribed @.***> Subject: Re: [HARPgroup/model_meteorology] PRISM Data (Issue #33)

I'm updating the scripts such that the download + processing are handled by a function and the import to the database is handled by a separate function. My intention is that the database import function, addRasterToDBase2.sh, is generic enough to be used by all of our data import processes.

— Reply to this email directly, view it on GitHubhttps://github.com/HARPgroup/model_meteorology/issues/33#issuecomment-2123410326, or unsubscribehttps://github.com/notifications/unsubscribe-auth/AC4K424HDVBNSDOFO4T7S4DZDOWYNAVCNFSM6AAAAABHYRNM6CVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMZDCMRTGQYTAMZSGY. You are receiving this because you are subscribed to this thread.Message ID: @.***>

rburghol commented 3 months ago

@COBrogan This seems like a really great idea. Thanks for modularising and I can't wait to take a look at it!