A good read about history of networks

Pretty humorously framed: https://apenwarr.ca/log/20170810

I would like to build an internet archiver to archive these posts.

What it will essentially do is:

1) Get HTML content of the link that I gave it 2) Save it in a key value store with all links in the page updated 3) Recursively save all links to to the key value store

It should allow for:

1) Content to be loaded just as it looks (No dynamic content) 2) Links should work and point to a valid page

Thinking a bot more about its design:

1) Rather than referring to the actual link in the query parameters, it will be better to use a hash and have a reverse map of hash to url if required in future (or perhaps this can be stored in a comment in the HTML body that will be stored) 2) So the crawler will work like this: * Get HTML content * Insert the url crawled as a comment in the head of the HTML * Save this content to a key value db with the url hash as the key and modified HTML as the body. * Add a link to the above in some other place to start the navigation – something like sections – or index it using elasticsearch so that we can build a search engine on the articles.

This does sound like the wayback machine, but its objective is just to preserve content that I want to preserve – knowing the changes is not required at all, and this will mostly be a personal collection of things.

Some day I will find the time to implement this.