Id like to put my lab servers to work archiving US federal data thats likely to get pulled - climate and biomed data seems mostly likely. The most obvious strategy to me seems like setting up mirror torrents on academictorrents. Anyone compiling a list of at-risk data yet?
archive.org is hosted in the US and could end up being a valid target. It doesn't strike me as being a very good place to securely store anything nowadays. I'd consider anything hosted in the US to be out.
For myself: Wayback It saves link to multiple different web archives and gives me pdf and warc files.
For others: Archive team have a few active projects to save at risk data and there is IRC channel in which people can suggest adding other websites for saving. They also have wiki with explanations how people can help.
I archive youtube videos that I like with TubeArchivist, I have a playlist for random videos i'd like to keep, and also subscribe to some of my favourite creator so I can keeptheir videos, even when I'm offline
Seems nice, but you need an external Player to watch the content, which can be goof for some people, but I like the webUI of TubeArchivist (even though it can be enhanced for sure)
Monolith can be particularly handy for this. I used it in a recent project to archive the outgoing links from my own site. Coincidentally, if anyone is interested in that, it's called django-cool-urls.
I am using archivebox, it is pretty straight-forward to self-host and use.
However, it is very difficult to archive most news sites with it and many other sites as well. Most cookie etc pop ups on a site will render the archived page unusable and often archiving won’t work at all because some bot protection (Cloudflare etc.) will kick-in when archivebox tries to access a site.
If anyone else has more success using it, please let me know if I am doing something wrong…
Monolith has the same problem here. I think the best resolution might be some sort of browser-plugin based solution where you could say "archive this" and have it push the result somewhere.
I wonder if I could combine a dumb plugin with Monolith to do that... A weekend project perhaps.
EDIT: Added script. Note that the script doesn't include archiving to
archivebox, since its API isn't available in stable verison yet. You can add a
function depending on your setup. Personally, I am depending on Caddy and
docker, so I am using caddy module [1] to execute commands with this in my Caddyfile:
Thank you for the warning. You are correct. It's prune to command injection. I will validate the URL before executing it. This shoud suffice until archivebox's rest API is available in stable.
In that they're a single organization, yes, but I'm a single person with significantly fewer resources. Non-availability is a significantly higher risk for things I host personally.
There was the attack on the Internet archive recently, are there any good options out there to help mirror some of the data or otherwise provide redundancy?