Table of Contents

related [[infra]] [[linkrot]]

* why?

motivation: If I do it, I would be able to search on all pages I ever visited [[search]] [[memex]]

Archiving-URLs -

The most ambitious & total approach to local caching is to set up a proxy to do your browsing through, and record literally all your web traffic; for example, using Live Archiving Proxy (LAP) or WarcProxy which will save as WARC files every page you visit through it. (Zachary Vance explains how to set up a local HTTPS certificate to MITM your HTTPS browsing as well.)

One may be reluctant to go this far, and prefer something lighter-weight, such as periodically extracting a list of visited URLs from one’s web browser and then attempting to archive them.

[2018-11-05] just backup everything you can find in promnesia? [[promnesia]]

* archivebox [[archivebox]]

The tool I'm currently using, very decent

ok, first instapaper run

[√] 2020-08-11 01:33:33 Update of 252 pages complete (146.68 min)
    - 0 links skipped
    - 228 links updated
    - 24 links had errors
535M	./1597100812.87
609M	./1597100812.31
757M	./1597100812.221
1.1G	./1597100812.173
8.5G	.

Ok, and second run the next day said it's already added all of them to index. Nice!


hmm wonder how did it manage to do user mapping??? is 1000 just dome default docker thing?

suggest to use run --rm

crap, timestamps, not shas are used… again??

ok, need to multithread..

add command – set maximum limit for data transferred?

prune command – I think I had some scripts already…

index web interface – might be useful to have size? for detecting largest offenders

index web interface – would be nice to mark sites that errored? Not sure what's the actionable outcome of that though

this issue

[2020-08-11] ok, need to starti without the pdf, screenshot etc… takes too long

also make sure it's possibe to add pdfs as an afterthought?

[2020-08-05] Release Major new ArchiveBox version, with a brand new CLI, UI, and SQLite index · pirate/ArchiveBox

Major new ArchiveBox version, with a brand new CLI, UI, and SQLite index

trying out the new one [[archivebox]]

[2020-10-25] would be nice to have parallel execution or something..

[2020-10-25] hmm, if archiving is interrupted, how to carry on? apparently 'archivebox update'?

[2020-10-25] media – could def download later/in parallel..

ok, I think I just want to take promnesia and run it against all non-browser sources [[promnesia]]

would be nice to mark different sources as well if possible?

bookmark Archiver

maybe just feed promnesia database to it??

right, so just archive redoes the index? Should run in against wereyouhere I suppose…

commit my changes to archiver, maybe even add the scripts?

figure out 404 etc

[2019-04-06] should run it after I normalise all the wereyouhere links?

I guess filter out all suspicious ones, containing special characters?

[2019-04-16] ok, he's working on django backend where we can use hashes

I guess some sites (with comments) – useful to update regularly, but most are okay with one snapshot?

status command is kinda similar to my old blame script? (might be on a branch)

only save mp3 for youtube videos? I guess it should be selective… or maybe dpeend on number of views

wonder if my exporters could be useful for archivebox [[orger]] [[promnesia]]

[2019-04-16] pirate/ArchiveBox: 🗃 The open source self-hosted web archive. Takes browser history/bookmarks/Pocket/Pinboard/etc., saves HTML, JS, PDFs, media, and more…

Storage Requirements
Because ArchiveBox is designed to ingest a firehose of browser history and bookmark feeds to a local disk, it can be much more disk-space intensive than a centralized service like the Internet Archive or However, as storage space gets cheaper and compression improves, you should be able to use it continuously over the years without having to delete anything. In my experience, ArchiveBox uses about 5gb per 1000 articles, but your milage may vary depending on which options you have enabled and what types of sites you're archiving. By default, it archives everything in as many formats as possible, meaning it takes more space than a using a single method, but more content is accurately replayable over extended periods of time. Storage requirements can be reduced by using a compressed/deduplicated filesystem like ZFS/BTRFS, or by setting FETCH_MEDIA=False to skip audio & video files.

[2019-04-16] pirate/ArchiveBox: 🗃 The open source self-hosted web archive. Takes browser history/bookmarks/Pocket/Pinboard/etc., saves HTML, JS, PDFs, media, and more…

Support for saving multiple snapshots of each site over time will be added soon (along with the ability to view diffs of the changes between runs).

[2019-04-16] [pirate/ArchiveBox] Bugfixes, new data integrity and invariant checks, remove title prefetching

re-save index after archiving completes to update titles and urls
emove title prefetching in favor of new FETCH_TITLE archive method

backup config?

prioritise never bookmarked over bookmarked with errors

commit it??

some links are pretty crazy… maybe prune huge pages manually and ignore

e.g. wget -N -E -np -x -H -k -K -S –restrict-file-names=unix -p –user-agent=Bookmark Archiver –no-check-certificate
– about 150M

[2018-10-03] kanishka-linux/reminiscence: Self-Hosted Bookmark And Archive Manager

[2018-10-05] wonder how is it different from my bookmark archiver? [[webarchive]]

Tweet from @gwern [[linkrot]]

pull tweet">

@gwern: @karlicoss @thomas536 Not documented in there yet is my latest archiving tool: Currently costs 20GB for 7,677 PDFs & self-contained single-file HTML mirrors.

[2019-12-20] Web Archiving Community · pirate/ArchiveBox Wiki [[linkrot]]

[2019-12-11] Verizon/Yahoo Blocking Attempts to Archive Yahoo Groups – Deletion: Dec. 14 | Hacker News

ven if we still had the Library of Alexandria, it may have shed zero light on the actual lives of citizens. Archiving content on the internet means capturing thousands of individual level perspectives and experiences. We don't know what will end up being important to historians 50 or 100 years from now. I would bet there are dozens if not hundreds of historians that would give anything for a record of their favorite time period that contains even a fraction of the amount of content today's archive efforts are storing.

[2020-05-28] site-deaths - IndieWeb [[linkrot]]

[2019-04-19] Bret Victor on Twitter: "60% of my fav links from 10 yrs ago are 404. I wonder if Library of Congress expects 60% of their collection to go up in smoke every decade." [[linkrot]]

pull tweet">

[2019-06-13] Time Travel: Find Mementos in Internet Archive, Archive-It, British Library,, GitHub and many more! [[search]] [[linkrot]]

[2019-12-22] This Page is Designed to Last [[linkrot]]

[2019-07-08] Fund: On-Demand Web Archiving of Annotated Pages – Hypothesis [[linkrot]]

[2020-03-06] Archiving URLs | Hacker News

[2021-02-25] Wikipedia:Database download - Wikipedia [[wikipedia]]

pages-articles-multistream.xml.bz2 – Current revisions only, no talk or user pages; this is probably what you want, and is approximately 18 GB compressed (expands to over 78 GB when decompressed).

ugh. image preservation is a mess… [[wikipedia]] [[webarchive]]

[2021-02-25] Wikipedia:Database download - Wikipedia

pages-articles.xml.bz2 and pages-articles-multistream.xml.bz2 both contain the same xml contents. So if you unpack either, you get the same data. But with multistream, it is possible to get an article from the archive without unpacking the whole thing.

[2021-02-25] Main Page - Kiwix [[prepping]] [[wikipedia]]

[2021-02-25] jeharu comments on The full English Wikipedia on Kiwix now weighs 79Gb instead of 94Gb thanks to improvements in image compression [[kiwix]] [[prepping]]

[–]jeharu54TB 46 points 2 months ago
no support yet for incremental updating, right? bummer.

    permalinkembedsavereportgive awardreply
[–]The_other_kiwix_guy[S] 66 points 2 months ago
We've started working on a prototype but that'll take time and a lot more money than we have. Would not expect anything before another 2-3 years.

hm okay sad.. guess I can do a backup per year or smth for now

[2021-02-25] Wikipedia:Database download - Wikipedia

The only downside to multistream is that it is marginally larger

would be nice to maybe tag urls… e.g. which source they are coming from [[archivebox]]

or just have a special source for manual notes/exobrainy stuff and another one for the rest?

def archive things I post (e.g. referenced in my own tweets/comments etc) [[self]] [[archivebox]]

could also check api? [[webarchive]]

e.g. it archives medium-like stuff?

hmm, a bit confused about – how reliable is it? is it backed up somewhere? perhaps still should save stuff from there locally… [[webarchive]]

pdfs on the other hand are a bit of higher priority? [[webarchive]]

[2021-03-26] AdGuardHome/whotracksme.json at master · AdguardTeam/AdGuardHome [[archivebox]]

could use this to prune?