Skip to content

Latest commit

 

History

History
41 lines (31 loc) · 2.09 KB

goals-metrics-principles.md

File metadata and controls

41 lines (31 loc) · 2.09 KB

Goals

  • Build and maintain the most complete, accurate index of federal, public websites.
  • Scan those websites daily in order to generate a variety of useful data for known stakeholders.

Internal program measures of success

  • Sites and repositories are current and all links work
  • Scans are running daily
  • Data quality had been analyzed and future steps for improvement are planned out
  • Primary stakeholders are caught up with and know next steps

External program metrics

  • Website traffic
  • Data downloads
  • API calls
  • Number of stakeholders being served
  • Number of estimated labor hours saved

Draft tracker spreadsheet here

Principles

  • We work in the open.
  • We only collect information that is available to anyone over the public internet.
  • The program's products are machine-readable data files that are made publicly available, either as static flat files or queryable APIs.
  • We do not make presentation layers.
  • We only design and build scans in response to user needs articulated by known stakeholders.
  • All scans run against the complete Federal Website Index.
  • If a scan is no longer needed or used by a known stakeholder, we deprecate it.
  • We follow the stakeholder experience.
  • We prioritize reliability and accuracy of scans that we have launched.
  • Our focus is on current data. Though scan data is snapshotted to an archive repo once a month, our system is ruthlessly focused on best providing current data and not on being a warehouse for historical data.

Internal Reliability Metrics

  • All Target URLs included in Site Scanning data
  • Number/Percent of scans completed
  • Fields populated
  • Total time of scanning