Next: Very Advanced Usage, Previous: Simple Usage, Up: Examples [Contents][Index]
You have a file that contains the URLs you want to download? Use the
‘-i
’ switch:
wget -i file
If you specify ‘-
’ as file name, the URLs will be read from
standard input.
gnulog
:
wget -r https://www.gnu.org/ -o gnulog
wget --convert-links -r https://www.gnu.org/ -o gnulog
Retrieve only one HTML page, but make sure that all the elements needed for the page to be displayed, such as inline images and external style sheets, are also downloaded. Also make sure the downloaded page references the downloaded links.
wget -p --convert-links http://www.example.com/dir/page.html
The HTML page will be saved to www.example.com/dir/page.html
, and
the images, stylesheets, etc., somewhere under www.example.com/
,
depending on where they were on the remote server.
www.example.com/
directory.
In fact, I don’t want to have all those random server directories
anyway—just save all those files under a download/
subdirectory of the current directory.
wget -p --convert-links -nH -nd -Pdownload \ http://www.example.com/dir/page.html
www.lycos.com
’, showing the original
server headers:
wget -S http://www.lycos.com/
wget --save-headers http://www.lycos.com/ more index.html
wuarchive.wustl.edu
’, saving them
to /tmp
.
wget -r -l2 -P/tmp ftp://wuarchive.wustl.edu/
You want to download all the GIFs from a directory on an HTTP
server. You tried ‘wget http://www.example.com/dir/*.gif
’, but that
didn’t work because HTTP retrieval does not support globbing. In
that case, use:
wget -r -l1 --no-parent -A.gif http://www.example.com/dir/
More verbose, but the effect is the same. ‘-r -l1
’ means to
retrieve recursively (see Recursive Download), with maximum depth
of 1. ‘--no-parent
’ means that references to the parent directory
are ignored (see Directory-Based Limits), and ‘-A.gif
’ means to
download only the GIF files. ‘-A "*.gif"
’ would have worked
too.
wget -nc -r https://www.gnu.org/
If you want to encode your own username and password to HTTP or FTP, use the appropriate URL syntax (see URL Format).
wget ftp://hniksic:[email protected]/.emacs
Note, however, that this usage is not advisable on multi-user systems
because it reveals your password to anyone who looks at the output of
ps
.
You would like the output documents to go to standard output instead of to files?
wget -O - http://jagor.srce.hr/ http://www.srce.hr/
You can also combine the two options and make pipelines to retrieve the documents from remote hotlists:
wget -O - http://cool.list.com/ | wget --force-html -i -
Next: Very Advanced Usage, Previous: Simple Usage, Up: Examples [Contents][Index]