I keep having to Google wget incantations. So, I’m going to just write some common ones down here. The spell at the moment is below and can be used with my previous post about processing EPA WQX/STORET domain values into useful tables:
wget -r -c -np -l2 -H -nd -A.zip,.xslx -erobots=off -o download.log --user-agent="Mozilla/5.0 (X11; U; Linux i686; en-US; rv:188.8.131.52) Gecko/2008092416 Firefox/3.0.3" http://www.epa.gov/storet/wqx/wqx_getdomainvalueswebservice.html
This is derived from the following notes:
- if you omit ‘-A.mid’ it would just download everything.
- -l1 means 1 level deep.
- -np means ignore parent links.
- -r means recursively download links.
- -nc means don’t download stuff that is already downloaded (if you want to resume later, or check for new files some other time)
- If the links are to different subdomains, you can specify host-spanning using the -H option,
- e.g. if bar.html contains links to files on host src.foobar.com, it won’t fetch them unless you specify -H.
- It’s also a good idea in that case to limit spanning to a domain using -D foobar.com.
- The -w option waits 30 seconds between retrievals. Not used here.
- You can also use –limit-rate=20k to limit the download speed to 20kb per second.
- -nd, –no-directories or don’t create directories as files are found on the server. Just stick everything into a single directory.
- -c Continue the Incomplete Download Using wget -c
- –user-agent Some websites can disallow you to download its page by identifying that the user agent is not a browser. So you can mask the user agent by using –user-agent options and show wget like a browser as shown below.
--user-agent="Mozilla/5.0 (X11; U; Linux i686; en-US; rv:184.108.40.206) Gecko/2008092416 Firefox/3.0.3" URL-TO-DOWNLOAD
- If the internet connection has problem, and if the download file is large there is a chance of failures in the download. By default wget retries 20 times to make the download successful. If needed, you can increase retry attempts using –tries option as shown below.
wget --tries=75 DOWNLOAD-URL
- Log messages to a log file instead of stderr Using wget -o. When you wanted the log to be redirected to a log file instead of the terminal.
wget -o download.log DOWNLOAD-URL
Many sites now employ a means of blocking robots like wget from accessing their files. Most of the time they use .htaccess to do this. So a permanent workaround has wget mimic a normal browser. Just add the -d option. Like: $ wget -O/dev/null -d http://www.askapache.com If you run the command at the top, you’ll get a directory of files as below
$ ls -lht total 1.6M -rw-r--r-- 1 john.zastrow Domain Users 28K Feb 20 09:25 download.log -rw-r--r-- 1 john.zastrow Domain Users 77K Feb 20 09:16 All.zip -rw-r--r-- 1 john.zastrow Domain Users 33K Feb 20 09:16 Organization.zip -rw-r--r-- 1 john.zastrow Domain Users 101K Feb 20 09:16 Characteristic.zip -rw-r--r-- 1 john.zastrow Domain Users 6.7K Feb 20 09:16 MeasureUnit.zip ----- snip ------
wget --referer="http://www.google.com" --user-agent="Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:220.127.116.11) Gecko/20070725 Firefox/18.104.22.168" --header="Accept: text/xml,application/xml,application/xhtml+xml,text/html;q=0.9,text/plain;q=0.8,image/png,*/*;q=0.5" --header="Accept-Language: en-us,en;q=0.5" --header="Accept-Encoding: gzip,deflate" --header="Accept-Charset: ISO-8859-1,utf-8;q=0.7,*;q=0.7" --header="Keep-Alive: 300" -dnv http://www.askapache.com/sitemap.xml