Link rot

From Infogalactic: the planetary knowledge core
Jump to: navigation, search

<templatestyles src="Module:Hatnote/styles.css"></templatestyles>

Link rot (or linkrot), also known as link death, link breaking or reference rot, refers to the process by which hyperlinks on individual websites or the Internet in general point to web pages, servers or other resources that have become permanently unavailable. The phrase also describes the effects of failing to update out-of-date web pages that clutter search engine results. A link that does not work any more is called a broken link, dead link, or dangling link. Formally, this is a form of dangling reference: The target of the reference no longer exists.

Causes

One of the most common reasons for a broken link is that the web page to which it points no longer exists. This frequently results in a 404 error, which indicates that the web server responded but the specific page could not be found. Another type of dead link occurs when the server that hosts the target page stops working or relocates to a new domain name. The browser may return a DNS error or display a site unrelated to the content originally sought. The latter can occur when a domain name lapses and is reregistered by another party. Other reasons for broken links include:

  • Websites can be restructured or redesigned, or the underlying technology can be changed, altering or invalidating large numbers of inbound or internal links.
  • Many news sites keep articles freely accessible for only a short time period, and then move them behind a paywall. This causes a significant loss of supporting links in sites discussing news events and using media sites as references.
  • Links may expire.
  • Content may be intentionally removed by the owner.
  • Search results from social media such as Facebook and Tumblr are prone to link rot because of frequent changes in user privacy, the deletion of accounts, search result pointing to a dynamic page that has new results that differ from the cached result, or the deletion of links or photos.
  • Links can contain ephemeral, user-specific information such as session or login data. Because these are not universally valid, the result can be a broken link.
  • A link might be broken because of some form of blocking such as content filters or firewalls.
  • Dead links can also occur on the authoring side, when website content is assembled from Internet sources and deployed without properly verifying the link targets.

Prevalence

The 404 "Not Found" response is familiar to even the occasional web user. A number of studies have examined the prevalence of link rot on the web, in academic literature, and in digital libraries.[1] In a 2003 experiment, Fetterly et al. discovered that about one link out of every 200 disappeared each week from the Internet. McCown et al. (2005) discovered that half of the URLs cited in D-Lib Magazine articles were no longer accessible 10 years after publication, and other studies have shown link rot in academic literature to be even worse (Spinellis, 2003, Lawrence et al., 2001). Nelson and Allen (2002) examined link rot in digital libraries and found that about 3% of the objects were no longer accessible after one year. In 2014, bookmarking site Pinboard's owner Maciej Cegłowski reported a “pretty steady rate” of 5% link rot per year.[2]

A 2014 Harvard Law School study by Jonathan Zittrain, Kendra Albert and Lawrence Lessig, determined that approximately 50% of the URLs in U.S. Supreme Court opinions no longer link to the original information.[3] They also found that in a selection of legal journals published between 1999 and 2011, more than 70% of the links no longer functioned as intended. A 2013 study in BMC Bioinformatics analyzed nearly 15,000 links in abstracts from Thomson Reuters’ Web of Science citation index and found that the median lifespan of web pages was 9.3 years, and just 62% were archived.[4] In August 2015 Weblock analyzed more than 180'000 links from references in the full-text corpora of three major open access publishers and found that overall 24.5% of links cited were no longer available.[5]

Discovering

Discovering broken links might be done manually or automatically. Automated methods, including plug-ins for WordPress, Drupal and other content management system can be used to detect the presence of broken URLs. An alternative is using a specific broken link checker like Xenu's Link Sleuth. However, if a URL returns an HTTP 200 (OK) response, it may be accessible, but the contents of the page could have changed and may no longer be relevant. So manual checking links seems to be a must. Some web servers also return a soft 404, reporting to computers that the link works even though it doesn't. Bar-Yossef et al. (2004) [6] developed a heuristic for automatically discovering soft 404s.

Combating

There are numerous solutions for tackling broken links: Some work to prevent them in the first place, while others trying to resolve them when they have occurred. There are also numerous tools that have been developed to help combat link rot.

Authoring

  • Carefully select and implement hyperlinks, and verify them regularly after publication. Best practices include linking to primary rather than secondary sources and prioritizing stable sites. McCown et al., 2005, suggest avoiding URL citations that point to resources on researchers' personal pages.
  • Always look for the most compact and direct URL available, and ensure that it’s clean, with no unnecessary information after the core of the URL.[7] This process is often referred to as URL normalization or URL canonicalization.
  • When available, use digital object identifier (DOIs) and Persistent Uniform Resource Locators (PURLs) whenever possible.
  • Avoid linking to PDF documents if possible. Because PDFs are documents rather than web pages, their content can change without notice, and their names are more likely to contain characters such as spaces that must be translated into safe codes for URLs. Large PDFs may also download slowly and cause a timeout error.[7]
  • Avoiding linking to pages deep in a website, a practice known as deep linking.
  • Using web archiving services (for example, WebCite) to permanently archive and retrieve cited Internet references (Eysenbach and Trudel, 2005).

Server side

  • Never change URLs and never remove pages. If there is a reason to no longer have a page, such as a news site redacting a story, replace it with a message explaining its removal.
  • When URLs change, use redirection mechanisms such as "301: Moved Permanently" to automatically refer browsers and crawlers to the new location.
  • Content management systems may offer built-in solutions to the management of links, such as updating them when content is changed or moved on a site.
  • WordPress guards against link rot by replacing non-canonical URLs with their canonical versions.[8]
  • IBM's Peridot attempts to automatically fix broken links.
  • Permalinking stops broken links by guaranteeing that the content will not move for the foreseeable future. Another form of permalinking is linking to a permalink that then redirects to the actual content, ensuring that even though the real content may be moved etc., links pointing to the resources stay intact.
  • Design URLs—for example, semantic URLs—such that they won't need to change when a different person takes over maintenance of a document or when different software is used on the server.[9]

User side

  • The Linkgraph widget gets the URL of the correct page based upon the old broken URL by using historical location information.
  • The Google 404 Widget attempts to "guess" the correct URL, and also provides the user with a search box to find the correct page.
  • When a user receives a 404 response, the Google Toolbar attempts to assist the user in finding the missing page.[10]

Web archiving

<templatestyles src="Module:Hatnote/styles.css"></templatestyles>

To combat link rot, web archivists are actively engaged in collecting the Web or particular portions of the Web and ensuring the collection is preserved in an archive, such as an archive site, for future researchers, historians, and the public. The goal of the Internet Archive is to maintain an archive of the entire Web, taking periodic snapshots of pages that can then be accessed for free via the Wayback Machine. In January 2013 the company announced that it had reached the milestone of 240 billion archived URLs.[11] National libraries, national archives and other organizations are also involved in archiving culturally important Web content.

Individuals may use a number of tools that allow them to archive web resources that may go missing in the future:

  • The WayBack Machine, at the Internet Archive,[12] is a free website that archives old web pages. It does not archive websites whose owners have stated they do not want their website archived.
  • WebCite, a tool specifically for scholarly authors, journal editors and publishers to permanently archive "on-demand" and retrieve cited Internet references (Eysenbach and Trudel, 2005).
  • Perma, which is supported by the Harvard Law School together with a broad coalition of university libraries, takes a snapshot of a URL's content and returns a permanent link.[3]
  • The Hiberlink project, a collaboration between the University of Edinburgh, the Los Alamos National Laboratory and others, is working to measure “reference rot” in online academic articles, and also to what extent Web content has been archived.[13] A related project, Memento, has established a technical standard for accessing online content as it existed in the past.[14]
  • Some social bookmarking websites allow users to make online clones of any web page on the internet, creating a copy at an independent url which remains online even if the original page goes down.
  • Amber, created by the Harvard Berkman Center, is a new tool built to fight link rot through archiving links on Wordpress and Drupal sites to prevent web censorship and bolster content preservation.[15]

However, such preserving systems may encounter on and off service interruption so that the preserved URLs are not available now and then.[16]

See also

<templatestyles src="Div col/styles.css"/>

Further reading

Link rot on the Web

  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.

In academic literature

  • Habibzadeh, P.; Sciences, Schattauer GmbH - Publishers for Medicine and Natural (2013-01-01). "Decay of References to Web sites in Articles Published in General Medical Journals: Mainstream vs Small Journals". Applied Clinical Informatics 4 (4)[1]
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.
  • Lua error in package.lua at line 80: module 'strict' not found.

In digital libraries

  • Lua error in package.lua at line 80: module 'strict' not found.

References

  1. 1.0 1.1 Lua error in package.lua at line 80: module 'strict' not found.
  2. Lua error in package.lua at line 80: module 'strict' not found.
  3. 3.0 3.1 Lua error in package.lua at line 80: module 'strict' not found.
  4. Lua error in package.lua at line 80: module 'strict' not found.
  5. Lua error in package.lua at line 80: module 'strict' not found.
  6. Lua error in package.lua at line 80: module 'strict' not found.
  7. 7.0 7.1 Lua error in package.lua at line 80: module 'strict' not found.
  8. Lua error in package.lua at line 80: module 'strict' not found.
  9. Lua error in package.lua at line 80: module 'strict' not found.
  10. Lua error in package.lua at line 80: module 'strict' not found.
  11. Lua error in package.lua at line 80: module 'strict' not found.
  12. Lua error in package.lua at line 80: module 'strict' not found.
  13. Lua error in package.lua at line 80: module 'strict' not found.
  14. Lua error in package.lua at line 80: module 'strict' not found.
  15. Lua error in package.lua at line 80: module 'strict' not found.
  16. Lua error in package.lua at line 80: module 'strict' not found.

External links