Closed thomasliebig closed 2 years ago
First: Thanks for the contribution!
I'm a bit worried that adding 45000 binaries will make the repository more or less unusable.
How about making it one file per hour or one per day?
Another thing: I'd actually prefer CVS over XML, if the data structure would support that. But that's just taste, because of the easier processing.
Thx, compressing day-wise is a great suggestion, I'll see how to achieve this ..
Data retrieved from (Source)
Description is included in Readme.md Utilized scraping shellscript is in fetch.sh (RAW data)
Sampling rate: 1 minute, results in approx. 45000 XML files that capture bike positions as WGS84 coordinates in the city of cologne; each file is zip compressed to reduce filesize
in total ~500MB