Anybody knows any good freeware util that can easy grab a full web site to HDD ?
Including pictures and subfolders
Posted on 2003-04-20 07:27:29 by BogdanOntanu
Try wget, it has an option to recursively download html files and images/stylesheets/etc it uses.

Thomas
Posted on 2003-04-20 08:23:20 by Thomas
What about Internet Explorer option of making your favourites avaible offline? Would it work?
Posted on 2003-04-20 09:01:05 by roticv

Try wget, it has an option to recursively download html files and images/stylesheets/etc it uses.

Thomas


Yep, wget is the weapon of choice.

http://www.interlog.com/~tcharron/wgetwin.html



bazik@exodus bazik $ wget --help
GNU Wget 1.8.2, a non-interactive network retriever.
Usage: wget [OPTION]... [URL]...

Mandatory arguments to long options are mandatory for short options too.

Startup:
-V, --version display the version of Wget and exit.
-h, --help print this help.
-b, --background go to background after startup.
-e, --execute=COMMAND execute a `.wgetrc'-style command.

Logging and input file:
-o, --output-file=FILE log messages to FILE.
-a, --append-output=FILE append messages to FILE.
-d, --debug print debug output.
-q, --quiet quiet (no output).
-v, --verbose be verbose (this is the default).
-nv, --non-verbose turn off verboseness, without being quiet.
-i, --input-file=FILE download URLs found in FILE.
-F, --force-html treat input file as HTML.
-B, --base=URL prepends URL to relative links in -F -i file.
--sslcertfile=FILE optional client certificate.
--sslcertkey=KEYFILE optional keyfile for this certificate.
--egd-file=FILE file name of the EGD socket.

Download:
--bind-address=ADDRESS bind to ADDRESS (hostname or IP) on local host.
-t, --tries=NUMBER set number of retries to NUMBER (0 unlimits).
-O --output-document=FILE write documents to FILE.
-nc, --no-clobber don't clobber existing files or use .# suffixes.
-c, --continue resume getting a partially-downloaded file.
--progress=TYPE select progress gauge type.
-N, --timestamping don't re-retrieve files unless newer than local.
-S, --server-response print server response.
--spider don't download anything.
-T, --timeout=SECONDS set the read timeout to SECONDS.
-w, --wait=SECONDS wait SECONDS between retrievals.
--waitretry=SECONDS wait 1...SECONDS between retries of a retrieval.
--random-wait wait from 0...2*WAIT secs between retrievals.
-Y, --proxy=on/off turn proxy on or off.
-Q, --quota=NUMBER set retrieval quota to NUMBER.
--limit-rate=RATE limit download rate to RATE.

Directories:
-nd --no-directories don't create directories.
-x, --force-directories force creation of directories.
-nH, --no-host-directories don't create host directories.
-P, --directory-prefix=PREFIX save files to PREFIX/...
--cut-dirs=NUMBER ignore NUMBER remote directory components.

HTTP options:
--http-user=USER set http user to USER.
--http-passwd=PASS set http password to PASS.
-C, --cache=on/off (dis)allow server-cached data (normally allowed).
-E, --html-extension save all text/html documents with .html extension.
--ignore-length ignore `Content-Length' header field.
--header=STRING insert STRING among the headers.
--proxy-user=USER set USER as proxy username.
--proxy-passwd=PASS set PASS as proxy password.
--referer=URL include `Referer: URL' header in HTTP request.
-s, --save-headers save the HTTP headers to file.
-U, --user-agent=AGENT identify as AGENT instead of Wget/VERSION.
--no-http-keep-alive disable HTTP keep-alive (persistent connections).
--cookies=off don't use cookies.
--load-cookies=FILE load cookies from FILE before session.
--save-cookies=FILE save cookies to FILE after session.

FTP options:
-nr, --dont-remove-listing don't remove `.listing' files.
-g, --glob=on/off turn file name globbing on or off.
--passive-ftp use the "passive" transfer mode.
--retr-symlinks when recursing, get linked-to files (not dirs).

Recursive retrieval:
-r, --recursive recursive web-suck -- use with care!
-l, --level=NUMBER maximum recursion depth (inf or 0 for infinite).
--delete-after delete files locally after downloading them.
-k, --convert-links convert non-relative links to relative.
-K, --backup-converted before converting file X, back up as X.orig.
-m, --mirror shortcut option equivalent to -r -N -l inf -nr.
-p, --page-requisites get all images, etc. needed to display HTML page.

Recursive accept/reject:
-A, --accept=LIST comma-separated list of accepted extensions.
-R, --reject=LIST comma-separated list of rejected extensions.
-D, --domains=LIST comma-separated list of accepted domains.
--exclude-domains=LIST comma-separated list of rejected domains.
--follow-ftp follow FTP links from HTML documents.
--follow-tags=LIST comma-separated list of followed HTML tags.
-G, --ignore-tags=LIST comma-separated list of ignored HTML tags.
-H, --span-hosts go to foreign hosts when recursive.
-L, --relative follow relative links only.
-I, --include-directories=LIST list of allowed directories.
-X, --exclude-directories=LIST list of excluded directories.
-np, --no-parent don't ascend to the parent directory.

Mail bug reports and suggestions to <bug-wget@gnu.org>.


:grin:
Posted on 2003-04-20 13:41:54 by bazik
Wget is a powerfull toll, but I like a lot also WinHTTrack (http://www.httrack.com). It is open source too. Is is one of the few which let me grab msdn.microsoft.com pages without coping on my hard disk half of the whole internet network.
Posted on 2003-04-20 17:00:22 by LuHa
wget, I got it to work (I like it), I just had to RTFM (and configure the proxy settings) :)
Posted on 2003-04-20 18:25:39 by scientica
Thanks LuHa:alright:
Posted on 2003-04-21 02:57:28 by Vortex
Originally posted by LuHa
Wget is a powerfull toll, but I like a lot also WinHTTrack (http://www.httrack.com). It is open source too. Is is one of the few which let me grab msdn.microsoft.com pages without coping on my hard disk half of the whole internet network. [

:grin: So you have the whole msdn down on your computer? Cool :alright:
Posted on 2003-04-21 03:41:05 by roticv


:grin: So you have the whole msdn down on your computer? Cool :alright:


Lol :grin:. No, only the MSJ section.
Posted on 2003-04-21 05:11:35 by LuHa
I use http://www.webstripper.net/ ... I've used a few times and so far I hadn't had any trouble.

Later,
Posted on 2003-04-21 12:28:14 by gorshing