GNU Wget NEWS -- history of user-visible changes. Copyright (C) 1997, 1998, 2000, 2001, 2002, 2003 Free Software Foundation, Inc. See the end for copying conditions. Please send GNU Wget bug reports to . * Changes in Wget 1.9. ** It is now possible to specify that POST method be used for HTTP requests. For example, `wget --post-data="id=foo&data=bar" URL' will send a POST request with the specified contents. ** IPv6 is experimentally supported. ** The `--timeout' option now affects DNS lookup and establishing the connection as well. Previously it only affected reading and writing data. Those three timeouts can be set separately using `--dns-timeout', `--connection-timeout', and `--read-timeout', respectively. ** Download speed shown by the progress bar is based on the data recently read, rather than the average speed of the entire download. The ETA projection is still based on the overall average. ** It is now possible to connect to FTP servers through FWTK firewalls. Set ftp_proxy to an FTP URL, and Wget will automatically log on to the proxy as "username@host". ** The new option `--retry-connrefused' makes Wget retry downloads even in the face of refused connections, which are otherwise considered a fatal error. ** The new option `--dns-cache=off' may be used to prevent Wget from caching DNS lookups. ** Wget no longer escapes characters in local file names based on whether they're appropriate in URLs. Escaping can still occur for nonprintable characters or for '/', but no longer for frequent characters such as space. You can use the new option --restrict-file-names to relax or strengthen these rules, which can be useful if you dislike the default or if you're downloading to non-native partitions. ** Handling of HTML comments has been dumbed down to conform to what users expect and other browsers do: instead of being treated as SGML declaration, a comment is terminated at the first occurrence of "-->". Use `--strict-comments' to revert to the old behavior. ** Wget now correctly handles relative URIs that begin with "//", such as "//img.foo.com/foo.jpg". ** Boolean options in `.wgetrc' and on the command line now accept values "yes" and "no" along with the traditional "on" and "off". ** It is now possible to specify decimal values for --timeout, --wait, --waitretry, and --limit-rate. For instance, `--wait=0.5' now works as expected, and so does --limit-rate=2.5k. * Wget 1.8.2 is a bugfix release with no user-visible changes. * Wget 1.8.1 is a bugfix release with no user-visible changes. * Changes in Wget 1.8. ** A new progress indicator is now available and used by default. You can choose the progress bar type with `--progress=TYPE'. Two types are available, "bar" (the new default), and "dot" (the old dotted indicator). You can permanently revert to the old progress indicator by putting `progress = dot' in your `.wgetrc'. ** You can limit the download rate of the retrieval using the `--limit-rate' option. For example, `wget --limit-rate=15k URL' will tell Wget not to download the body of the URL faster than 15 kilobytes per second. ** Recursive retrieval and link conversion have been revamped: *** Wget now traverses links breadth-first. This makes the calculation of depth much more reliable than before. Also, recursive downloads are faster and consume *significantly* less memory than before. *** Links are converted only when the entire retrieval is complete. This is the only safe thing to do, as only then is it known what URLs have been downloaded. *** BASE tags are handled correctly when converting links. Since Wget already resolves when resolving handling URLs, link conversion now makes the BASE tags point to an empty string. *** HTML anchors are now handled correctly. Links to an anchor in the same document (), which used to confuse Wget, are now converted correctly. *** When in page-requisites (-p) mode, no-parent (-np) is ignored when retrieving for inline images, stylesheets, and other documents needed to display the page. *** Page-requisites (-p) mode now works with frames. In other words, `wget -p URL-THAT-USES-FRAMES' will now download the frame HTML files, and all the files that they need to be displayed properly. ** `--base' now works conjunction with `--input-file', providing a base for each URL and thereby allowing the URLs in the file to be relative. ** If a host has more than one IP address, Wget uses the other addresses when accessing the first one fails. ** Host directories now contain port information if the URL is at a non-standard port. ** Wget now supports the robots.txt directives specified in . ** URL parser has been fixed, especially the infamous overzealous quoting. Wget no longer dequotes reserved characters, e.g. `%3F' is no longer translated to `?', nor `%2B' to `+'. Unsafe characters which are not reserved are still escaped, of course. ** No more than 20 successive redirections are allowed. * Wget 1.7.1 is a bugfix release with no user-visible changes. * Changes in Wget 1.7. ** SSL (`https') pages now work if you compile Wget with SSL support; use the `--with-ssl' configure flag. You need to have OpenSSL installed. ** Cookies are now supported. Wget will accept cookies sent by the server and return them in later requests. Additionally, it can load and save cookies to disk, in the same format that Netscape uses. ** "Keep-alive" (persistent) HTTP connections are now supported. Using keep-alive allows Wget to share one TCP/IP connection for many retrievals, making multiple-file downloads faster and less stressing for the server and the network. ** Wget now recognizes FTP directory listings generated by NT and VMS servers. ** It is now possible to recurse through FTP sites where logging in puts you in some directory other than '/'. ** You may now use `~' to mean home directory in `.wgetrc'. For example, `load_cookies = ~/.netscape/cookies.txt' works as you would expect. ** The HTML parser has been rewritten. The new one works more reliably, allows finer-grained control over which tags and attributes are detected, and has better support for some features like correctly skipping comments and declarations, decoding entities, etc. It is also more general. ** tags are now respected. ** Wget's internal tables now use hash tables instead of linked lists where appropriate. This results in huge speedups when retrieving large sites (thousands of documents). ** Wget now has a man page, automatically generated from the Texinfo documentation. (The last version that shipped with a man page was 1.4.5). To get this, you need to have pod2man from the Perl distribution installed on your system. * Changes in Wget 1.6 ** Administrative changes. *** Maintainership. Due to Hrvoje being plagued with a "real job", Dan Harkless is the most active maintainer (not that he doesn't have a real job as well). Hrvoje still participates occasionally, and both are being helped by many other people. *** Web page. Thanks to Jan Prikryl, Wget has an "official" web page. Take a look at: http://sunsite.dk/wget/ *** Anonymous CVS. Thanks to ever-helpful Karsten Thygesen, Wget sources are now available at an anonymous CVS server. Take a look at the web page for downloading instructions. ** New -K / --backup-converted / backup_converted = on option causes files modified due to -k to be saved with a .orig prefix before being changed. When using -N as well, it is these .orig files that are compared against the server. ** New --follow-tags / follow_tags = ... option allows you to restrict Wget to following only certain HTML tags when doing a recursive retrieval. -G / --ignore-tags / ignore_tags = ... is just the opposite -- all tags but the ones you specify will be followed. ** New --waitretry / waitretry = SECONDS option allows waiting between retries of failed downloads. Wget will use "linear" backoff, waiting 1 second after the first failure, 2 after the second, up to SECONDS. waitretry is set to 10 by default in the system wgetrc. ** New -p / --page-requisites / page_requisites = on option causes Wget to download all ancillary files necessary to display a given HTML page properly (e.g. inlined images). ** New -E / --html-extension / html_extension = on option causes Wget to append ".html" to text/html filenames not ending in regexp "\.[Hh][Tt][Mm][Ll]?". ** New type of .wgetrc command -- "lockable Boolean". Can be set to on, off, always, or never. This allows the .wgetrc to override the commandline. So far, passive_ftp is the only .wgetrc command which takes a lockable Boolean. ** A number of new translation files have been added. ** New --bind-address / bind_address =
option for people on hosts bound to multiple IP addresses. ** wget now accepts (illegal per HTTP spec) relative URLs in HTTP redirects. * Wget 1.5.3 is a bugfix release with no user-visible changes. * Wget 1.5.2 is a bugfix release with no user-visible changes. * Wget 1.5.1 is a bugfix release with no user-visible changes. * Changes in Wget 1.5.0 ** Wget speaks many languages! On systems with gettext(), Wget will output messages in the language set by the current locale, if available. At this time we support Czech, German, Croatian, Italian, Norwegian and Portuguese. ** Opie (Skey) is now supported with FTP. ** HTTP Digest Access Authentication (RFC2069) is now supported. ** The new `-b' option makes Wget go to background automatically. ** The `-I' and `-X' options now accept wildcard arguments. ** The `-w' option now accepts suffixes `s' for seconds, `m' for minutes, `h' for hours, `d' for days and `w' for weeks. ** Upon getting SIGHUP, the whole previous log is now copied to `wget-log'. ** Wget now understands proxy settings with explicit usernames and passwords, e.g. `http://user:password@proxy.foo.com/'. ** You can use the new `--cut-dirs' option to make Wget create less directories. ** The `;type=a' appendix to FTP URLs is now recognized. For instance, the following command will retrieve the welcoming message in ASCII type transfer: wget "ftp://ftp.somewhere.com/welcome.msg;type=a" ** `--help' and `--version' options have been redone to to conform to standards set by other GNU utilities. ** Wget should now be compilable under MS Windows environment. MS Visual C++ and Watcom C have been used successfully. ** If the file length is known, percentages are displayed during download. ** The manual page, now hopelessly out of date, is no longer distributed with Wget. * Wget 1.4.5 is a bugfix release with no user-visible changes. * Wget 1.4.4 is a bugfix release with no user-visible changes. * Changes in Wget 1.4.3 ** Wget is now a GNU utility. ** Can do passive FTP. ** Reads .netrc. ** Info documentation expanded. ** Compiles on pre-ANSI compilers. ** Global wgetrc now goes to /usr/local/etc (i.e. $sysconfdir). ** Lots of bugfixes. * Changes in Wget 1.4.2 ** New mirror site at ftp://sunsite.auc.dk/pub/infosystems/wget/, thanks to Karsten Thygesen. ** Mailing list! Mail to wget-request@sunsite.auc.dk to subscribe. ** New option --delete-after for proxy prefetching. ** New option --retr-symlinks to retrieve symbolic links like plain files. ** rmold.pl -- script to remove files deleted on the remote server ** --convert-links should work now. ** Minor bugfixes. * Changes in Wget 1.4.1 ** Minor bugfixes. ** Added -I (the opposite of -X). ** Dot tracing is now customizable; try wget --dot-style=binary * Changes in Wget 1.4.0 ** Wget 1.4.0 [formerly known as Geturl] is an extensive rewrite of Geturl. Although many things look suspiciously similar, most of the stuff was rewritten, like recursive retrieval, HTTP, FTP and mostly everything else. Wget should be now easier to debug, maintain and, most importantly, use. ** Recursive HTTP should now work without glitches, even with Location changes, server-generated directory listings and other naughty stuff. ** HTTP regetting is supported on servers that support Range specification. WWW authorization is supported -- try wget http://user:password@hostname/ ** FTP support was rewritten and widely enhanced. Globbing should now work flawlessly. Symbolic links are created locally. All the information the Unix-style ls listing can give is now recognized. ** Recursive FTP is supported, e.g. wget -r ftp://gnjilux.cc.fer.hr/pub/unix/util/ ** You can specify "rejected" directories, to which you do not want to enter, e.g. with wget -X /pub ** Time-stamping is supported, with both HTTP and FTP. Try wget -N URL. ** A new texinfo reference manual is provided. It can be read with Emacs, standalone info, or converted to HTML, dvi or postscript. ** Fixed a long-standing bug, so that Wget now works over SLIP connections. ** You can have a system-wide wgetrc (/usr/local/lib/wgetrc by default). Settings in $HOME/.wgetrc override the global ones, of course :-) ** You can set up quota in .wgetrc to prevent sucking too much data. Try `quota = 5M' in .wgetrc (or quota = 100K if you want your sysadmin to like you). ** Download rate is printed after retrieval. ** Wget now sends the `Referer' header when retrieving recursively. ** With the new --no-parent option Wget can retrieve FTP recursively through a proxy server. ** HTML parser, as well as the whole of Wget was rewritten to be much faster and less memory-consuming (yes, both). ** Absolute links can be converted to relative links locally. Check wget -k. ** Wget catches hangup, filtering the output to a log file and resuming work. Try kill -HUP %?wget. ** User-defined headers can be sent. Try wget http://fly.cc.her.hr/ --header='Accept-Charset: iso-8859-2' ** Acceptance/Rejection lists may contain wildcards. ** Wget can display HTTP headers and/or FTP server response with the new `-S' option. It can save the original HTTP headers with `-s'. ** socks library is now supported (thanks to Antonio Rosella ). Configure with --with-socks. ** There is a nicer display of REST-ed output. ** Many new options (like -x to force directory hierarchy, or -m to turn on mirroring options). ** Wget is now distributed under GNU General Public License (GPL). ** Lots of small features I can't remember. :-) ** A host of bugfixes. * Changes in Geturl 1.3 ** Added FTP globbing support (ftp://fly.cc.fer.hr/*) ** Added support for no_proxy ** Added support for ftp://user:password@host/ ** Added support for %xx in URL syntax ** More natural command-line options ** Added -e switch to execute .geturlrc commands from the command-line ** Added support for robots.txt ** Fixed some minor bugs * Geturl 1.2 is a bugfix release with no user-visible changes. * Changes in Geturl 1.1 ** REST supported in FTP ** Proxy servers supported ** GNU getopt used, which enables command-line arguments to be ordered as you wish, e.g. geturl http://fly.cc.fer.hr/ -vo log is the same as geturl -vo log http://fly.cc.fer.hr/ ** Netscape-compatible URL syntax for HTTP supported: host[:port]/dir/file ** NcFTP-compatible colon URL syntax for FTP supported: host:/dir/file ** supported ** autoconf supported ---------------------------------------------------------------------- Copyright information: Copyright (C) 1997, 1998, 2000, 2001 Free Software Foundation, Inc. Permission is granted to anyone to make or distribute verbatim copies of this document as received, in any medium, provided that the copyright notice and this permission notice are preserved, thus giving the recipient permission to redistribute in turn. Permission is granted to distribute modified versions of this document, or of portions of it, under the above conditions, provided also that they carry prominent notices stating who last changed them.