[Noisebridge-discuss] archiving a password protected website
ryanobjc at gmail.com
Sun May 13 23:41:28 PDT 2012
wget --username=<username> --password=<password> -r -np <URL>
be careful, sometimes auto-generated sites can produce endless loops
of content that confuses wget.
If things get hairy, put:
to limit how 'deep' the get should follow links.
On Sun, May 13, 2012 at 11:33 PM, Lee Sonko <lee at lee.org> wrote:
> I'm trying to make an archive of a website subscription I belong to,
> my 23andme.com account. I can't find a tool that will download this website.
> I'd rather not copy-and-paste 300 pages. I tried WinHTTrack. Maybe WGet
> excels at this but it's a steep learning curve; it'd be nice if someone
> point me in the direction of a tool that could do it.
> I can see two obvious hurdles. Logging in might be designed to be an
> interactive process (I tried dragging cookies around in WinHTrack to no
> avail). And maybe so much depends on the server on a modern website that it
> might not be possible to have a web page without a server. What say the
> Noisebridge-discuss mailing list
> Noisebridge-discuss at lists.noisebridge.net
More information about the Noisebridge-discuss