Closed 3zero2 closed 3 years ago
You did not specify an area of interest in the above command, so by default the script will download and process the NO2 data over the whole globe. The default resolution being 7x4km approximately, this is indeed a huge amount of data to resample. As far as I know, HARP tries to load the whole grid in memory and cannot work in batch. If you want to resample the data globally, i suggest either:
--resolution
optionYou are completely right. I provided an --aoi and now it completes.
For those reading this I had to put import geopandas
on the first line of the s5p_requests file. For some reason if geopandas was not imported first it was. giving off a segmentation fault.
Is it normal that the HARP tools for the L3 resampling use huge amounts of memory? It tried to download and process L2__NO2___ data with this command:
python s5p-request.py L2__NO2___ --date NOW-2HOUR NOW --num-workers 1
The download was quite small but when it got to the L3 resampling python started to use huge amounts of RAM until it crashed the computer. Is this expected when using HARP tools?