Some large organizations do not have a centralized version control system (VCS). Most importantly, each component does not have a single individual or service to run the Scraper across the entire agency. Primary examples would be DHS and DOD, large organizations, the software is managed at the component level.
For instance, NSA and Code.mil have separate teams validating and authorizing the publication of the open source software (OSS). Similarly, DHS has 23 components, each component would most likely generate their own code.json file.
For that reason, we would like to see a feature on the Scraper that aggregates multiple code.json source files into a master code.json file.
Some large organizations do not have a centralized version control system (VCS). Most importantly, each component does not have a single individual or service to run the Scraper across the entire agency. Primary examples would be DHS and DOD, large organizations, the software is managed at the component level.
For instance, NSA and Code.mil have separate teams validating and authorizing the publication of the open source software (OSS). Similarly, DHS has 23 components, each component would most likely generate their own code.json file.
For that reason, we would like to see a feature on the Scraper that aggregates multiple code.json source files into a master code.json file.
Multiple sources files: www.domain1.gov/code.json www.domain2.gov/code.json www.domain3.gov/code.json = code.json file.
Resulting into an official master code.json file for an agency to be published on the national website.
Additionally, that would support the code.gov platform to always read the software inventory from one code.JSON source file for each agency.