Re: website saver for linux?



Philip wrote:
I'm looking for a tool which spiders a site, and downloads every page in
the domain that it finds linked from a particular url and linked urls
in the domain, creating a local site that can be manipulated offline as
static html.

Is there such a tool for linux (better still debian)?

thanks, Philip


look for the httrack package.

WT


--
To UNSUBSCRIBE, email to debian-user-REQUEST@xxxxxxxxxxxxxxxx with a subject of "unsubscribe". Trouble? Contact listmaster@xxxxxxxxxxxxxxxx