It's difficult for governments to figure out what open data sets they're already publishing. Realistically, an employee just has to search for site:example.gov inurl:json and start paging through the results. It'd be really helpful to have a tool that can automatically find every machine-readable file on a given domain, gather metadata for each of those, optionally filter out noise (e.g., XML config files), and generate a data.json file (or something like that) inventorying everything.
It's difficult for governments to figure out what open data sets they're already publishing. Realistically, an employee just has to search for
site:example.gov inurl:json
and start paging through the results. It'd be really helpful to have a tool that can automatically find every machine-readable file on a given domain, gather metadata for each of those, optionally filter out noise (e.g., XML config files), and generate adata.json
file (or something like that) inventorying everything.