Comment by π decant
Re: "pandoc and website preservation"
@Half_Elf_Monk For wikipedia, I use to tricks: 1. zim archive and python library, zimply, I don't know if zimply works with minimal text browser. 2. there is always gempedia, I just save the .gmi file for good articles. But as @stack said there are many irredeemable java heavy sites where I will just download with wget or just ctrl-c ctrl-v the text part.
2025-11-29 Β· 5 months ago
2 Later Comments β
π² Half_Elf_Monk Β· Dec 12 at 16:16:
@stack - That would work, but I simply don't trust the LLMs to get details right. @decant - interesting. Realistically, shouldn't there be a way to simply download a snapshot of wikipedea as a whole? As mostly-text, it shouldn't be that big. Why are there no local wikipedia browsers?
π stack Β· Dec 12 at 19:31:
@Half_Elf_Monk -- there is a stripped down wikipedia distribution, people often put it on local devices feaured on Gizmodo and such... Can't remember where to get it, and not at a decent computer now, but I am sure you can easily find it.
Original Post
pandoc and website preservation β Back when I used firefox/chromium. I use their print function to save full web page to a pdf file. For example, Paul V Bolotoff wrote articles on the history of DEC alpha CPUs, but his website is long gone, the only copy of the article I could find is on the archive section of someoneβs personal site. But I found out I could use pandoc accomplish this task: pandoc [http link] -o oldarticle.epub I find the epub family of formats better suit my needs. PDF is...