What can we do to prevent linkrot?
Image credit: https://uxdesign.cc/how-to-design-a-404-error-page-that-keeps-users-on-your-site-f3443a980ece
Please leave the feedback on this challenge
Is the problem still unsolved?
Is it concisely described?
Bounty for the best solution
Provide a bounty for the best solution
Bounties attract serious brainpower to the challenge.
Blockchain of the internet + page ID of every historic version
- What if every page on the internet received its own ID as soon as it was detected by a scanner/bot. And every modification of every page would save it as a newer version.
- Blockchain archive of the internet (minus AI-detected malicious/nonsense/spam pages)
- Some huge institutions on par with the library of Alexandria would keep the blockchain copies safe and up to date. This would be funded by nation-states and philanthropy.
Please leave the feedback on this idea
The linkrot (LR) number of as a webpage "up-to-date" metrics
- It could be a nice metric that tells you about how up-to-date is the website. Websites with many rotten links would mean that the site is not regularly maintained and that you should be careful about the facts found there.
- It would force people to edit their rotten links immediately in order to keep a high LR ranking, since it would become a metric considered by searching algorithms.