I’ve been looking online for ways to download websites (game wikis mostly), in order to have them all in my collection and ensure that they dont get taken down or changed.
After trying linkwarden, which is fine for singular web pages, one has to manually link each individual page of the wiki in order to make a pdf.
With this in mind, the only other option that I’ve discovered is using wget recursively. Do any you of you have experience with this or reccomend alternative ideas? Any and all help is appreciated.
PS: I will most likely download official game guides which will cover most of the games, but looking for something to cover all my games library.
I’ve used HTTrack in the past
It worked well for the websites that I tried it on and it’s pretty easy to run. You could give it a go?
https://en.wikipedia.org/wiki/HTTrack
I’ve got a container set up of this. Drops the output on the NAS and can be accessed from any box on the local net. The only issue is it has a tendency to need the container recycled every so often, like it just gets board sitting there and quits. 🤔
Cool! I didn’t know that was an option