I’ve found that all the web archiving software I’ve encountered are either manual (you have to archive everything individually in a separate application) or crawler-based (which can end up putting a lot of extra load on smaller web server, and could even get your ip blocked).

Are there any solutions that simply automatically archive web pages as you load them in your browser? If not, why aren’t there?

I could also see something like that being useful as a self-hosted web indexer, where if you ever go “I think I’ve seen this name before”, you can click on it, and your computer will say something like “this name appeared in a news headline you scrolled past two weeks ago”

  • detonational_VuSE@lemmy.ml
    link
    fedilink
    English
    arrow-up
    1
    ·
    10 hours ago

    wget is the command line program to do what you’re saying. Or what I use, anyway. Not tied to a browser, though. Maybe you could export your history and pipe it into wget if you’re using Linux or have a Linux-like command line?

    I also use the FF SingleFile plugin. Again, not automatic, though.