The Wild Wild Web

  • Surf Mode new!
  • About
  • Latest

    rainbow separator

    Curating dead websites

    I recently added two new categories to the directory, /~/learning and /~/fediverse, which brings the directory closer to 400 websites.

    With that many websites, which were added in a span of 3 months, I started experiencing the inevitable and universal problem of any web directory: some websites started to die.

    I decided to tackle this issue sooner than later with a simple system:

    - A website can be declared dead.

    - When a website is dead, it goes in a graveyard. For example, here is the ~/social graveyard. Of course, a dead website won't appear anymore in other sections, including the Surf Mode.

    - A website can be brought back to life if it becomes alive again. It can also be permanently archived when the curator is certain the website cannot be resuscitated.

    To help the curator handle dead websites, there is a new background job that pings websites every day. That job marks websites unreachable in case of failure and unmark them when they become reachable again.

    I decided against a 100% automated process because it's not trivial to automatically detect that a website is unreachable. For example, the website may require a challenge to access or block crawler requests.

    With the solution I implemented, curating dead websites takes minutes per week, which is good enough for now.
    - j0d1,

    Previously

    rainbow separator