Stories
Slash Boxes
Comments

SoylentNews is people

posted by janrinok on Tuesday January 30, @06:50PM   Printer-friendly
from the the-net-never-forgets-ha dept.

Web developer Trevor Morris has a short post on the attrition of web sites over the years.

I have run the Laravel Artisan command I built to get statistics on my outgoing links section. Exactly one year later it doesn't make good reading.

[...] The percentage of total broken links has increased from 32.8% last year to 35.7% this year. Links from over a decade ago have a fifty per cent chance of no longer working. Thankfully, only three out of over 550 have gone missing in the last few years of links, but only time will tell how long they'll stick around.

As pointed out in the early and mid 1990s, the inherent centralization of sites, later web sites, is the basis for this weakness. That is to say one single copy exists which resides under the control of the publisher / maintainer. When that one copy goes, it is gone.


Original Submission

 
This discussion was created by janrinok (52) for logged-in users only, but now has been archived. No new comments can be posted.
Display Options Threshold/Breakthrough Mark All as Read Mark All as Unread
The Fine Print: The following comments are owned by whoever posted them. We are not responsible for them in any way.
  • (Score: 2) by JoeMerchant on Tuesday January 30, @08:49PM (6 children)

    by JoeMerchant (3937) on Tuesday January 30, @08:49PM (#1342439)

    The wayback machine doesn't work nearly as well as it used to.

    --
    🌻🌻 [google.com]
    Starting Score:    1  point
    Karma-Bonus Modifier   +1  

    Total Score:   2  
  • (Score: 1, Informative) by Anonymous Coward on Tuesday January 30, @08:59PM (4 children)

    by Anonymous Coward on Tuesday January 30, @08:59PM (#1342442)

    > The wayback machine doesn't work nearly as well as it used to.

    Istr seeing something about this, might be due to all the variable content that is served out of a database (page created on the fly) and/or all the code that is behind "modern" (aka bloated) web pages?

    When I send the Wayback Machine a link to a fairly simple page it seems to work same as ever. Often those are the pages with the most archival value anyway...but ymmv.
     

    • (Score: 2) by JoeMerchant on Tuesday January 30, @10:30PM (3 children)

      by JoeMerchant (3937) on Tuesday January 30, @10:30PM (#1342448)

      Wayback works great on the website I made in 1997 (and continue to maintain in the same style through today).

      I threw up a Wordpress blog somewhere around 2005, I don't think Wayback ever picked it up.

      --
      🌻🌻 [google.com]
      • (Score: 4, Funny) by Anonymous Coward on Wednesday January 31, @03:52AM (2 children)

        by Anonymous Coward on Wednesday January 31, @03:52AM (#1342472)

        > I threw up a Wordpress blog ...

        I can't imagine how you ever swallowed it in the first place, but it's good it came back up rather than causing an intestinal perforation.

        • (Score: 3, Funny) by The Vocal Minority on Wednesday January 31, @05:22PM (1 child)

          by The Vocal Minority (2765) on Wednesday January 31, @05:22PM (#1342527) Journal

          Contrary to popular belief, a swallowed Wordpress blog, although providing no nutritional value, is unlikely to cause intestinal perforation. A particularly large blog may become lodged, however, and require extensive archiving to remove.

          • (Score: 1, Informative) by Anonymous Coward on Thursday February 01, @05:44AM

            by Anonymous Coward on Thursday February 01, @05:44AM (#1342591)

            Ah, so it becomes a bezoar [webpathology.com]. That sounds about right.

  • (Score: 1, Informative) by Anonymous Coward on Wednesday January 31, @07:29AM

    by Anonymous Coward on Wednesday January 31, @07:29AM (#1342480)

    Because it respects robots.txt, etc. So when a domain squatter takes over the previous archived site can vanish if the new robots.txt or similar tells IA to not archive:

    https://help.archive.org/help/using-the-wayback-machine/ [archive.org]

    Some sites are not available because of robots.txt or other exclusions. What does that mean?

    Such sites may have been excluded from the Wayback Machine due to a robots.txt file on the site or at a site owner’s direct request.