Go to the first, previous, next, last section, table of contents.
-
If you wish Wget to keep a mirror of a page (or FTP
subdirectories), use `--mirror' (`-m'), which is the shorthand
for `-r -N'. You can put Wget in the crontab file asking it to
recheck a site each Sunday:
crontab
0 0 * * 0 wget --mirror ftp://ftp.xemacs.org/pub/xemacs/ -o /home/me/weeklog
-
You may wish to do the same with someone's home page. But you do not
want to download all those images -- you're only interested in
HTML.
wget --mirror -A.html http://www.w3.org/
-
But what about mirroring the hosts networkologically close to you? It
seems so awfully slow because of all those DNS resolving. Just
use `-D' (See section 6.3 Domain Acceptance).
wget -rN -Dsrce.hr http://www.srce.hr/
Now Wget will correctly find out that `regoc.srce.hr' is the same
as `www.srce.hr', but will not even take into consideration the
link to `www.mit.edu'.
-
You have a presentation and would like the dumb absolute links to be
converted to relative? Use `-k'.
wget -k -r URL
-
Suppose you want Wget not to download anything, but just to fill your
PROXY cache. You would probably use options like this:
wget -r -nd -P/tmp/trash http://whatever/
`-r' to retrieve recursively, `-nd' not to create directories
and -P/tmp/trash for all the output to go where you can easily remove
it.
-
You would like the output documents to go to stdout instead of to files?
OK, but Wget will automatically shut up (turn on `--quiet') to
prevent mixing of Wget output and the retrieved documents.
wget -O - http://jagor.srce.hr/ http://www.srce.hr/
You can also combine the two options and make weird pipelines to
retrieve the documents from remote hotlists:
wget -O - http://cool.list.com/ | wget --force-html -i -
Go to the first, previous, next, last section, table of contents.