6 Triage

Not all websites and data files are of equal risk of being unrecoverable. Thanks to the Internet Archive and the work of groups like ArchiveTeam, many copies of at-risk sites and data have already been made. In time-sensitive situations, care should be taken to prioritize data that is less likely to be backed up and retrievable at a later date.

6.1 Low concern: statically-linked files

Of least concern are files that are statically-linked, i.e. files that are shown at the bottom left of your browser window when you hover your mouse over them. For example, see the following screenshot of a page that lists data downloads from the CDC Youth Risk Behavior Surveillance System:

When we mouse over the “Access [ZIP]” link, we can see a small popup at the bottom left of the window displaying a URL, which indicates that this file is statically-linked:

Statically-linked files such these are the most likely to be picked up by web crawlers, such as those run by the Internet Archive. We can confirm whether or not these data have already been archived by consulting the Wayback Machine before proceeding (see the Wayback Machine URLs Tab or the querying the CDX API sections):

If time is extremely limited, then assume that pages and files that are statically-linked have already been captured and move on.

6.3 High concern: data behind portals

Data that exist behind portals requiring user interaction via the mouse or keyboard are of high concern, since basic web crawlers may not be able to interface with the portal depending on how it is set up. Unlike dynamically-loading pages, it’s more likely that these portals behave in drastically different ways. Of special concern are portals that use JavaScript to operate, though these are hard to identify without looking into the source HTML code behind a page.

Portals can be identified by their distinctive buttons and menus that do not display URLs at the bottom left of your browser window when your mouse hovers over them. Most of the time, these are HTTP <input> tags. See the MDN page on HTML <input> types for more examples of how these look in your browser.

See below for some examples of portals with their distinguishing features highlighted:

Screenshot of the CDC WONDER data portal.
Screenshot of the CDC WONDER data portal.
Screenshot of the U.S. Census Bureau TIGER/Line data portal. TIGER/Line is a special case because the data that it offers is available either through the form or through an FTP link co-located next to the form, which can easily be crawled by bots.
Screenshot of the U.S. Census Bureau TIGER/Line data portal. TIGER/Line is a special case because the data that it offers is available either through the form or through an FTP link co-located next to the form, which can easily be crawled by bots.

6.4 Highest concern: data served only by APIs

In some cases, data behind portal can be crawled if there is not heavy use of JavaScript or if the data is accessible through static links (for example the TIGER/Line data). However, data behind APIs most likely cannot be crawled, as there is no single standard for designing an API. Many different API protocols exist for data of different purposes, for example OpenAPI/Swagger, CKAN, GraphQL, etc.

As such, knowledge of

  1. What API protocol is being used (if the API is even using a standardized protocol to begin with),
  2. Which endpoints are responsible for serving the data of interest, and
  3. How to write a program to interface with those endpoints API and scrape data

are all required in order to scrape data from an API, and all of these generally require human intervention.

As with the previous sections, some data that are accessible from APIs can also be retrieved through other means. When determining what data is most vulnerable, be sure to check if this is true for any given API before assigning it a high priority.

6.5 Summary

  • When deciding what pages and data to prioritize for preservation, think: is a robot able to retrieve this data on its own, without human intervention? If so, then it’s likely that one already has and that you will be able to retrieve that data later.
  • To reiterate: if time is extremely limited, then assume that pages and files that are statically-linked have already been captured and move on.