- Build and maintain the most complete, accurate index of federal, public websites.
- Scan those websites daily in order to generate a variety of useful data for known stakeholders.
- Sites and repositories are current and all links work
- Scans are running daily
- Data quality had been analyzed and future steps for improvement are planned out
- Primary stakeholders are caught up with and know next steps
- Website traffic
- Data downloads
- API calls
- Number of stakeholders being served
- Number of estimated labor hours saved
Draft tracker spreadsheet here
- We work in the open.
- We only collect information that is available to anyone over the public internet.
- The program's products are machine-readable data files that are made publicly available, either as static flat files or queryable APIs.
- We do not make presentation layers.
- We only design and build scans in response to user needs articulated by known stakeholders.
- All scans run against the complete Federal Website Index.
- If a scan is no longer needed or used by a known stakeholder, we deprecate it.
- We follow the stakeholder experience.
- We prioritize reliability and accuracy of scans that we have launched.
- Our focus is on current data. Though scan data is snapshotted to an archive repo once a month, our system is ruthlessly focused on best providing current data and not on being a warehouse for historical data.
- All Target URLs included in Site Scanning data
- Number/Percent of scans completed
- Fields populated
- Total time of scanning