X-Git-Url: http://sjero.net/git/?p=wget;a=blobdiff_plain;f=src%2Fretr.c;h=b667ca2ff3cf6ecb4edca9c1148e856f68b0ecbe;hp=a19262a03c1fce7ad30b35341781132f79036431;hb=d763f8bf6d6e13ce006ffab616cc8a77e747a633;hpb=7c044778bc83f3714e8d91d31c992a76d78e42ad diff --git a/src/retr.c b/src/retr.c index a19262a0..b667ca2f 100644 --- a/src/retr.c +++ b/src/retr.c @@ -1,11 +1,12 @@ /* File retrieval. - Copyright (C) 1995, 1996, 1997, 1998, 2000, 2001 Free Software Foundation, Inc. + Copyright (C) 1996, 1997, 1998, 1999, 2000, 2001, 2002, 2003, + 2004, 2005, 2006, 2007, 2008 Free Software Foundation, Inc. This file is part of GNU Wget. GNU Wget is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by -the Free Software Foundation; either version 2 of the License, or (at +the Free Software Foundation; either version 3 of the License, or (at your option) any later version. GNU Wget is distributed in the hope that it will be useful, @@ -14,57 +15,50 @@ MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License -along with Wget; if not, write to the Free Software -Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. +along with Wget. If not, see . -In addition, as a special exception, the Free Software Foundation -gives permission to link the code of its release of Wget with the -OpenSSL project's "OpenSSL" library (or with modified versions of it -that use the same license as the "OpenSSL" library), and distribute -the linked executables. You must obey the GNU General Public License -in all respects for all of the code used other than "OpenSSL". If you -modify this file, you may extend this exception to your version of the -file, but you are not obligated to do so. If you do not wish to do -so, delete this exception statement from your version. */ +Additional permission under GNU GPL version 3 section 7 -#include +If you modify this program, or any covered work, by linking or +combining it with the OpenSSL project's OpenSSL library (or a +modified version of that library), containing parts covered by the +terms of the OpenSSL or SSLeay licenses, the Free Software Foundation +grants you additional permission to convey the resulting work. +Corresponding Source for a non-source form of such a combination +shall include the source code for the parts of OpenSSL used as well +as that of the covered work. */ + +#include "wget.h" #include #include -#include #ifdef HAVE_UNISTD_H # include #endif /* HAVE_UNISTD_H */ #include -#ifdef HAVE_STRING_H -# include -#else -# include -#endif /* HAVE_STRING_H */ +#include #include -#include "wget.h" #include "utils.h" #include "retr.h" #include "progress.h" #include "url.h" #include "recur.h" #include "ftp.h" +#include "http.h" #include "host.h" #include "connect.h" #include "hash.h" #include "convert.h" - -#ifdef HAVE_SSL -# include "gen_sslfunc.h" /* for ssl_iread */ -#endif - -#ifndef errno -extern int errno; -#endif +#include "ptimer.h" +#include "html-url.h" +#include "iri.h" /* Total size of downloaded files. Used to enforce quota. */ -LARGE_INT total_downloaded_bytes; +SUM_SIZE_INT total_downloaded_bytes; + +/* Total download time in seconds. */ +double total_download_time; /* If non-NULL, the stream to which output should be written. This stream is initialized when `-O' is used. */ @@ -72,7 +66,7 @@ FILE *output_stream; /* Whether output_document is a regular file we can manipulate, i.e. not `-' or a device file. */ -int output_stream_regular; +bool output_stream_regular; static struct { wgint chunk_bytes; @@ -83,8 +77,7 @@ static struct { static void limit_bandwidth_reset (void) { - limit_data.chunk_bytes = 0; - limit_data.chunk_start = 0; + xzero (limit_data); } /* Limit the bandwidth by pausing the download for an amount of time. @@ -92,9 +85,9 @@ limit_bandwidth_reset (void) is the timer that started at the beginning of download. */ static void -limit_bandwidth (wgint bytes, struct wget_timer *timer) +limit_bandwidth (wgint bytes, struct ptimer *timer) { - double delta_t = wtimer_read (timer) - limit_data.chunk_start; + double delta_t = ptimer_read (timer) - limit_data.chunk_start; double expected; limit_data.chunk_bytes += bytes; @@ -102,37 +95,42 @@ limit_bandwidth (wgint bytes, struct wget_timer *timer) /* Calculate the amount of time we expect downloading the chunk should take. If in reality it took less time, sleep to compensate for the difference. */ - expected = 1000.0 * limit_data.chunk_bytes / opt.limit_rate; + expected = (double) limit_data.chunk_bytes / opt.limit_rate; if (expected > delta_t) { double slp = expected - delta_t + limit_data.sleep_adjust; double t0, t1; - if (slp < 200) - { - DEBUGP (("deferring a %.2f ms sleep (%s/%.2f).\n", - slp, number_to_static_string (limit_data.chunk_bytes), - delta_t)); - return; - } + if (slp < 0.2) + { + DEBUGP (("deferring a %.2f ms sleep (%s/%.2f).\n", + slp * 1000, number_to_static_string (limit_data.chunk_bytes), + delta_t)); + return; + } DEBUGP (("\nsleeping %.2f ms for %s bytes, adjust %.2f ms\n", - slp, number_to_static_string (limit_data.chunk_bytes), - limit_data.sleep_adjust)); + slp * 1000, number_to_static_string (limit_data.chunk_bytes), + limit_data.sleep_adjust)); - t0 = wtimer_read (timer); - xsleep (slp / 1000); - wtimer_update (timer); - t1 = wtimer_read (timer); + t0 = ptimer_read (timer); + xsleep (slp); + t1 = ptimer_measure (timer); /* Due to scheduling, we probably slept slightly longer (or - shorter) than desired. Calculate the difference between the - desired and the actual sleep, and adjust the next sleep by - that amount. */ + shorter) than desired. Calculate the difference between the + desired and the actual sleep, and adjust the next sleep by + that amount. */ limit_data.sleep_adjust = slp - (t1 - t0); + /* If sleep_adjust is very large, it's likely due to suspension + and not clock inaccuracy. Don't enforce those. */ + if (limit_data.sleep_adjust > 0.5) + limit_data.sleep_adjust = 0.5; + else if (limit_data.sleep_adjust < -0.5) + limit_data.sleep_adjust = -0.5; } limit_data.chunk_bytes = 0; - limit_data.chunk_start = wtimer_read (timer); + limit_data.chunk_start = ptimer_read (timer); } #ifndef MIN @@ -145,7 +143,7 @@ limit_bandwidth (wgint bytes, struct wget_timer *timer) static int write_data (FILE *out, const char *buf, int bufsize, wgint *skip, - wgint *written) + wgint *written) { if (!out) return 1; @@ -160,7 +158,7 @@ write_data (FILE *out, const char *buf, int bufsize, wgint *skip, bufsize -= *skip; *skip = 0; if (bufsize == 0) - return 1; + return 1; } fwrite (buf, 1, bufsize, out); @@ -170,7 +168,18 @@ write_data (FILE *out, const char *buf, int bufsize, wgint *skip, performance: fast downloads will arrive in large 16K chunks (which stdio would write out immediately anyway), and slow downloads wouldn't be limited by disk speed. */ + + /* 2005-04-20 SMS. + Perhaps it shouldn't hinder performance, but it sure does, at least + on VMS (more than 2X). Rather than speculate on what it should or + shouldn't do, it might make more sense to test it. Even better, it + might be nice to explain what possible benefit it could offer, as + it appears to be a clear invitation to poor performance with no + actual justification. (Also, why 16K? Anyone test other values?) + */ +#ifndef __VMS fflush (out); +#endif /* ndef __VMS */ return !ferror (out); } @@ -187,7 +196,7 @@ write_data (FILE *out, const char *buf, int bufsize, wgint *skip, is incremented by the amount of data read from the network. If QTYWRITTEN is non-NULL, the value it points to is incremented by the amount of data written to disk. The time it took to download - the data (in milliseconds) is stored to ELAPSED. + the data is stored to ELAPSED. The function exits and returns the amount of data read. In case of error while reading data, -1 is returned. In case of error while @@ -195,14 +204,14 @@ write_data (FILE *out, const char *buf, int bufsize, wgint *skip, int fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, - wgint *qtyread, wgint *qtywritten, double *elapsed, int flags) + wgint *qtyread, wgint *qtywritten, double *elapsed, int flags) { int ret = 0; static char dlbuf[16384]; int dlbufsize = sizeof (dlbuf); - struct wget_timer *timer = NULL; + struct ptimer *timer = NULL; double last_successful_read_tm = 0; /* The progress gauge, set according to the user preferences. */ @@ -212,9 +221,9 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, continually update the display. When true, smaller timeout values are used so that the gauge can update the display when data arrives slowly. */ - int progress_interactive = 0; + bool progress_interactive = false; - int exact = flags & rb_read_exactly; + bool exact = !!(flags & rb_read_exactly); wgint skip = 0; /* How much data we've read/written. */ @@ -227,9 +236,10 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, if (opt.verbose) { /* If we're skipping STARTPOS bytes, pass 0 as the INITIAL - argument to progress_create because the indicator doesn't - (yet) know about "skipping" data. */ - progress = progress_create (skip ? 0 : startpos, startpos + toread); + argument to progress_create because the indicator doesn't + (yet) know about "skipping" data. */ + wgint start = skip ? 0 : startpos; + progress = progress_create (start, start + toread); progress_interactive = progress_interactive_p (progress); } @@ -241,7 +251,7 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, the timer. */ if (progress || opt.limit_rate || elapsed) { - timer = wtimer_new (); + timer = ptimer_new (); last_successful_read_tm = 0; } @@ -261,61 +271,61 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, int rdsize = exact ? MIN (toread - sum_read, dlbufsize) : dlbufsize; double tmout = opt.read_timeout; if (progress_interactive) - { - /* For interactive progress gauges, always specify a ~1s - timeout, so that the gauge can be updated regularly even - when the data arrives very slowly or stalls. */ - tmout = 0.95; - if (opt.read_timeout) - { - double waittm; - waittm = (wtimer_read (timer) - last_successful_read_tm) / 1000; - if (waittm + tmout > opt.read_timeout) - { - /* Don't let total idle time exceed read timeout. */ - tmout = opt.read_timeout - waittm; - if (tmout < 0) - { - /* We've already exceeded the timeout. */ - ret = -1, errno = ETIMEDOUT; - break; - } - } - } - } + { + /* For interactive progress gauges, always specify a ~1s + timeout, so that the gauge can be updated regularly even + when the data arrives very slowly or stalls. */ + tmout = 0.95; + if (opt.read_timeout) + { + double waittm; + waittm = ptimer_read (timer) - last_successful_read_tm; + if (waittm + tmout > opt.read_timeout) + { + /* Don't let total idle time exceed read timeout. */ + tmout = opt.read_timeout - waittm; + if (tmout < 0) + { + /* We've already exceeded the timeout. */ + ret = -1, errno = ETIMEDOUT; + break; + } + } + } + } ret = fd_read (fd, dlbuf, rdsize, tmout); - if (ret == 0 || (ret < 0 && errno != ETIMEDOUT)) - break; /* read error */ - else if (ret < 0) - ret = 0; /* read timeout */ + if (progress_interactive && ret < 0 && errno == ETIMEDOUT) + ret = 0; /* interactive timeout, handled above */ + else if (ret <= 0) + break; /* EOF or read error */ if (progress || opt.limit_rate) - { - wtimer_update (timer); - if (ret > 0) - last_successful_read_tm = wtimer_read (timer); - } + { + ptimer_measure (timer); + if (ret > 0) + last_successful_read_tm = ptimer_read (timer); + } if (ret > 0) - { - sum_read += ret; - if (!write_data (out, dlbuf, ret, &skip, &sum_written)) - { - ret = -2; - goto out; - } - } + { + sum_read += ret; + if (!write_data (out, dlbuf, ret, &skip, &sum_written)) + { + ret = -2; + goto out; + } + } if (opt.limit_rate) - limit_bandwidth (ret, timer); + limit_bandwidth (ret, timer); if (progress) - progress_update (progress, ret, wtimer_read (timer)); + progress_update (progress, ret, ptimer_read (timer)); #ifdef WINDOWS if (toread > 0 && !opt.quiet) - ws_percenttitle (100.0 * - (startpos + sum_read) / (startpos + toread)); + ws_percenttitle (100.0 * + (startpos + sum_read) / (startpos + toread)); #endif } if (ret < -1) @@ -323,12 +333,12 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, out: if (progress) - progress_finish (progress, wtimer_read (timer)); + progress_finish (progress, ptimer_read (timer)); if (elapsed) - *elapsed = wtimer_read (timer); + *elapsed = ptimer_read (timer); if (timer) - wtimer_delete (timer); + ptimer_destroy (timer); if (qtyread) *qtyread += sum_read; @@ -338,22 +348,35 @@ fd_read_body (int fd, FILE *out, wgint toread, wgint startpos, return ret; } -/* Read a hunk of data from FD, up until a terminator. The terminator - is whatever the TERMINATOR function determines it to be; for - example, it can be a line of data, or the head of an HTTP response. - The function returns the data read allocated with malloc. - - In case of error, NULL is returned. In case of EOF and no data - read, NULL is returned and errno set to 0. In case of EOF with - data having been read, the data is returned, but it will - (obviously) not contain the terminator. +/* Read a hunk of data from FD, up until a terminator. The hunk is + limited by whatever the TERMINATOR callback chooses as its + terminator. For example, if terminator stops at newline, the hunk + will consist of a line of data; if terminator stops at two + newlines, it can be used to read the head of an HTTP response. + Upon determining the boundary, the function returns the data (up to + the terminator) in malloc-allocated storage. + + In case of read error, NULL is returned. In case of EOF and no + data read, NULL is returned and errno set to 0. In case of having + read some data, but encountering EOF before seeing the terminator, + the data that has been read is returned, but it will (obviously) + not contain the terminator. + + The TERMINATOR function is called with three arguments: the + beginning of the data read so far, the beginning of the current + block of peeked-at data, and the length of the current block. + Depending on its needs, the function is free to choose whether to + analyze all data or just the newly arrived data. If TERMINATOR + returns NULL, it means that the terminator has not been seen. + Otherwise it should return a pointer to the charactre immediately + following the terminator. The idea is to be able to read a line of input, or otherwise a hunk of text, such as the head of an HTTP request, without crossing the boundary, so that the next call to fd_read etc. reads the data after the hunk. To achieve that, this function does the following: - 1. Peek at available data. + 1. Peek at incoming data. 2. Determine whether the peeked data, along with the previously read data, includes the terminator. @@ -381,9 +404,9 @@ fd_read_hunk (int fd, hunk_terminator_t terminator, long sizehint, long maxsize) { long bufsize = sizehint; char *hunk = xmalloc (bufsize); - int tail = 0; /* tail position in HUNK */ + int tail = 0; /* tail position in HUNK */ - assert (maxsize >= bufsize); + assert (!maxsize || maxsize >= bufsize); while (1) { @@ -394,90 +417,91 @@ fd_read_hunk (int fd, hunk_terminator_t terminator, long sizehint, long maxsize) pklen = fd_peek (fd, hunk + tail, bufsize - 1 - tail, -1); if (pklen < 0) - { - xfree (hunk); - return NULL; - } - end = terminator (hunk, tail, pklen); + { + xfree (hunk); + return NULL; + } + end = terminator (hunk, hunk + tail, pklen); if (end) - { - /* The data contains the terminator: we'll drain the data up - to the end of the terminator. */ - remain = end - (hunk + tail); - if (remain == 0) - { - /* No more data needs to be read. */ - hunk[tail] = '\0'; - return hunk; - } - if (bufsize - 1 < tail + remain) - { - bufsize = tail + remain + 1; - hunk = xrealloc (hunk, bufsize); - } - } + { + /* The data contains the terminator: we'll drain the data up + to the end of the terminator. */ + remain = end - (hunk + tail); + assert (remain >= 0); + if (remain == 0) + { + /* No more data needs to be read. */ + hunk[tail] = '\0'; + return hunk; + } + if (bufsize - 1 < tail + remain) + { + bufsize = tail + remain + 1; + hunk = xrealloc (hunk, bufsize); + } + } else - /* No terminator: simply read the data we know is (or should - be) available. */ - remain = pklen; + /* No terminator: simply read the data we know is (or should + be) available. */ + remain = pklen; /* Now, read the data. Note that we make no assumptions about - how much data we'll get. (Some TCP stacks are notorious for - read returning less data than the previous MSG_PEEK.) */ + how much data we'll get. (Some TCP stacks are notorious for + read returning less data than the previous MSG_PEEK.) */ rdlen = fd_read (fd, hunk + tail, remain, 0); if (rdlen < 0) - { - xfree_null (hunk); - return NULL; - } + { + xfree_null (hunk); + return NULL; + } tail += rdlen; hunk[tail] = '\0'; if (rdlen == 0) - { - if (tail == 0) - { - /* EOF without anything having been read */ - xfree (hunk); - errno = 0; - return NULL; - } - else - /* EOF seen: return the data we've read. */ - return hunk; - } + { + if (tail == 0) + { + /* EOF without anything having been read */ + xfree (hunk); + errno = 0; + return NULL; + } + else + /* EOF seen: return the data we've read. */ + return hunk; + } if (end && rdlen == remain) - /* The terminator was seen and the remaining data drained -- - we got what we came for. */ - return hunk; + /* The terminator was seen and the remaining data drained -- + we got what we came for. */ + return hunk; /* Keep looping until all the data arrives. */ if (tail == bufsize - 1) - { - /* Double the buffer size, but refuse to allocate more than - MAXSIZE bytes. */ - if (maxsize && bufsize >= maxsize) - { - xfree (hunk); - errno = ENOMEM; - return NULL; - } - bufsize <<= 1; - if (maxsize && bufsize > maxsize) - bufsize = maxsize; - hunk = xrealloc (hunk, bufsize); - } + { + /* Double the buffer size, but refuse to allocate more than + MAXSIZE bytes. */ + if (maxsize && bufsize >= maxsize) + { + xfree (hunk); + errno = ENOMEM; + return NULL; + } + bufsize <<= 1; + if (maxsize && bufsize > maxsize) + bufsize = maxsize; + hunk = xrealloc (hunk, bufsize); + } } } static const char * -line_terminator (const char *hunk, int oldlen, int peeklen) +line_terminator (const char *start, const char *peeked, int peeklen) { - const char *p = memchr (hunk + oldlen, '\n', peeklen); + const char *p = memchr (peeked, '\n', peeklen); if (p) - /* p+1 because we want the line to include '\n' */ + /* p+1 because the line must include '\n' */ return p + 1; return NULL; } @@ -501,18 +525,22 @@ fd_read_line (int fd) return fd_read_hunk (fd, line_terminator, 128, FD_READ_LINE_MAX); } -/* Return a printed representation of the download rate, as - appropriate for the speed. If PAD is non-zero, strings will be - padded to the width of 7 characters (xxxx.xx). */ -char * -retr_rate (wgint bytes, double msecs, int pad) +/* Return a printed representation of the download rate, along with + the units appropriate for the download speed. */ + +const char * +retr_rate (wgint bytes, double secs) { static char res[20]; static const char *rate_names[] = {"B/s", "KB/s", "MB/s", "GB/s" }; - int units = 0; + int units; - double dlrate = calc_rate (bytes, msecs, &units); - sprintf (res, pad ? "%7.2f %s" : "%.2f %s", dlrate, rate_names[units]); + double dlrate = calc_rate (bytes, secs, &units); + /* Use more digits for smaller numbers (regardless of unit used), + e.g. "1022", "247", "12.5", "2.38". */ + sprintf (res, "%.*f %s", + dlrate >= 99.95 ? 0 : dlrate >= 9.995 ? 1 : 2, + dlrate, rate_names[units]); return res; } @@ -524,21 +552,23 @@ retr_rate (wgint bytes, double msecs, int pad) UNITS is zero for B/s, one for KB/s, two for MB/s, and three for GB/s. */ + double -calc_rate (wgint bytes, double msecs, int *units) +calc_rate (wgint bytes, double secs, int *units) { double dlrate; - assert (msecs >= 0); + assert (secs >= 0); assert (bytes >= 0); - if (msecs == 0) + if (secs == 0) /* If elapsed time is exactly zero, it means we're under the - granularity of the timer. This often happens on systems that - use time() for the timer. */ - msecs = wtimer_granularity (); + resolution of the timer. This can easily happen on systems + that use time() for the timer. Since the interval lies between + 0 and the timer's resolution, assume half the resolution. */ + secs = ptimer_resolution () / 2.0; - dlrate = (double)1000 * bytes / msecs; + dlrate = bytes / secs; if (dlrate < 1024.0) *units = 0; else if (dlrate < 1024.0 * 1024.0) @@ -552,31 +582,25 @@ calc_rate (wgint bytes, double msecs, int *units) return dlrate; } -/* Maximum number of allowed redirections. 20 was chosen as a - "reasonable" value, which is low enough to not cause havoc, yet - high enough to guarantee that normal retrievals will not be hurt by - the check. */ - -#define MAX_REDIRECTIONS 20 - -#define SUSPEND_POST_DATA do { \ - post_data_suspended = 1; \ - saved_post_data = opt.post_data; \ - saved_post_file_name = opt.post_file_name; \ - opt.post_data = NULL; \ - opt.post_file_name = NULL; \ + +#define SUSPEND_POST_DATA do { \ + post_data_suspended = true; \ + saved_post_data = opt.post_data; \ + saved_post_file_name = opt.post_file_name; \ + opt.post_data = NULL; \ + opt.post_file_name = NULL; \ } while (0) -#define RESTORE_POST_DATA do { \ - if (post_data_suspended) \ - { \ - opt.post_data = saved_post_data; \ - opt.post_file_name = saved_post_file_name; \ - post_data_suspended = 0; \ - } \ +#define RESTORE_POST_DATA do { \ + if (post_data_suspended) \ + { \ + opt.post_data = saved_post_data; \ + opt.post_file_name = saved_post_file_name; \ + post_data_suspended = false; \ + } \ } while (0) -static char *getproxy PARAMS ((struct url *)); +static char *getproxy (struct url *); /* Retrieve the given URL. Decides which loop to call -- HTTP, FTP, FTP, proxy, etc. */ @@ -585,19 +609,22 @@ static char *getproxy PARAMS ((struct url *)); multiple points. */ uerr_t -retrieve_url (const char *origurl, char **file, char **newloc, - const char *refurl, int *dt) +retrieve_url (struct url * orig_parsed, const char *origurl, char **file, + char **newloc, const char *refurl, int *dt, bool recursive, + struct iri *iri) { uerr_t result; char *url; - int location_changed, dummy; + bool location_changed; + bool iri_fallbacked = 0; + int dummy; char *mynewloc, *proxy; - struct url *u, *proxy_url; - int up_error_code; /* url parse error code */ + struct url *u = orig_parsed, *proxy_url; + int up_error_code; /* url parse error code */ char *local_file; int redirection_count = 0; - int post_data_suspended = 0; + bool post_data_suspended = false; char *saved_post_data = NULL; char *saved_post_file_name = NULL; @@ -613,18 +640,11 @@ retrieve_url (const char *origurl, char **file, char **newloc, if (file) *file = NULL; - u = url_parse (url, &up_error_code); - if (!u) - { - logprintf (LOG_NOTQUIET, "%s: %s.\n", url, url_error (up_error_code)); - xfree (url); - return URLERROR; - } - if (!refurl) refurl = opt.referer; redirected: + /* (also for IRI fallbacking) */ result = NOCONERROR; mynewloc = NULL; @@ -634,24 +654,30 @@ retrieve_url (const char *origurl, char **file, char **newloc, proxy = getproxy (u); if (proxy) { + struct iri *pi = iri_new (); + set_uri_encoding (pi, opt.locale, true); + pi->utf8_encode = false; + /* Parse the proxy URL. */ - proxy_url = url_parse (proxy, &up_error_code); + proxy_url = url_parse (proxy, &up_error_code, NULL, true); if (!proxy_url) - { - logprintf (LOG_NOTQUIET, _("Error parsing proxy URL %s: %s.\n"), - proxy, url_error (up_error_code)); - xfree (url); - RESTORE_POST_DATA; - return PROXERR; - } + { + char *error = url_error (proxy, up_error_code); + logprintf (LOG_NOTQUIET, _("Error parsing proxy URL %s: %s.\n"), + proxy, error); + xfree (url); + xfree (error); + RESTORE_POST_DATA; + return PROXERR; + } if (proxy_url->scheme != SCHEME_HTTP && proxy_url->scheme != u->scheme) - { - logprintf (LOG_NOTQUIET, _("Error in proxy URL %s: Must be HTTP.\n"), proxy); - url_free (proxy_url); - xfree (url); - RESTORE_POST_DATA; - return PROXERR; - } + { + logprintf (LOG_NOTQUIET, _("Error in proxy URL %s: Must be HTTP.\n"), proxy); + url_free (proxy_url); + xfree (url); + RESTORE_POST_DATA; + return PROXERR; + } } if (u->scheme == SCHEME_HTTP @@ -660,28 +686,29 @@ retrieve_url (const char *origurl, char **file, char **newloc, #endif || (proxy_url && proxy_url->scheme == SCHEME_HTTP)) { - result = http_loop (u, &mynewloc, &local_file, refurl, dt, proxy_url); + result = http_loop (u, &mynewloc, &local_file, refurl, dt, proxy_url, iri); } else if (u->scheme == SCHEME_FTP) { - /* If this is a redirection, we must not allow recursive FTP - retrieval, so we save recursion to oldrec, and restore it - later. */ - int oldrec = opt.recursive; + /* If this is a redirection, temporarily turn off opt.ftp_glob + and opt.recursive, both being undesirable when following + redirects. */ + bool oldrec = recursive, glob = opt.ftp_glob; if (redirection_count) - opt.recursive = 0; - result = ftp_loop (u, dt, proxy_url); - opt.recursive = oldrec; + oldrec = glob = false; + + result = ftp_loop (u, dt, proxy_url, recursive, glob); + recursive = oldrec; /* There is a possibility of having HTTP being redirected to - FTP. In these cases we must decide whether the text is HTML - according to the suffix. The HTML suffixes are `.html', - `.htm' and a few others, case-insensitive. */ + FTP. In these cases we must decide whether the text is HTML + according to the suffix. The HTML suffixes are `.html', + `.htm' and a few others, case-insensitive. */ if (redirection_count && local_file && u->scheme == SCHEME_FTP) - { - if (has_html_suffix_p (local_file)) - *dt |= TEXTHTML; - } + { + if (has_html_suffix_p (local_file)) + *dt |= TEXTHTML; + } } if (proxy_url) @@ -699,28 +726,39 @@ retrieve_url (const char *origurl, char **file, char **newloc, assert (mynewloc != NULL); if (local_file) - xfree (local_file); + xfree (local_file); /* The HTTP specs only allow absolute URLs to appear in - redirects, but a ton of boneheaded webservers and CGIs out - there break the rules and use relative URLs, and popular - browsers are lenient about this, so wget should be too. */ + redirects, but a ton of boneheaded webservers and CGIs out + there break the rules and use relative URLs, and popular + browsers are lenient about this, so wget should be too. */ construced_newloc = uri_merge (url, mynewloc); xfree (mynewloc); mynewloc = construced_newloc; + /* Reset UTF-8 encoding state, keep the URI encoding and reset + the content encoding. */ + iri->utf8_encode = opt.enable_iri; + set_content_encoding (iri, NULL); + xfree_null (iri->orig_url); + /* Now, see if this new location makes sense. */ - newloc_parsed = url_parse (mynewloc, &up_error_code); + newloc_parsed = url_parse (mynewloc, &up_error_code, iri, true); if (!newloc_parsed) - { - logprintf (LOG_NOTQUIET, "%s: %s.\n", escnonprint_uri (mynewloc), - url_error (up_error_code)); - url_free (u); - xfree (url); - xfree (mynewloc); - RESTORE_POST_DATA; - return result; - } + { + char *error = url_error (mynewloc, up_error_code); + logprintf (LOG_NOTQUIET, "%s: %s.\n", escnonprint_uri (mynewloc), + error); + if (orig_parsed != u) + { + url_free (u); + } + xfree (url); + xfree (mynewloc); + xfree (error); + RESTORE_POST_DATA; + return result; + } /* Now mynewloc will become newloc_parsed->url, because if the Location contained relative paths like .././something, we @@ -729,44 +767,77 @@ retrieve_url (const char *origurl, char **file, char **newloc, mynewloc = xstrdup (newloc_parsed->url); /* Check for max. number of redirections. */ - if (++redirection_count > MAX_REDIRECTIONS) - { - logprintf (LOG_NOTQUIET, _("%d redirections exceeded.\n"), - MAX_REDIRECTIONS); - url_free (newloc_parsed); - url_free (u); - xfree (url); - xfree (mynewloc); - RESTORE_POST_DATA; - return WRONGCODE; - } + if (++redirection_count > opt.max_redirect) + { + logprintf (LOG_NOTQUIET, _("%d redirections exceeded.\n"), + opt.max_redirect); + url_free (newloc_parsed); + if (orig_parsed != u) + { + url_free (u); + } + xfree (url); + xfree (mynewloc); + RESTORE_POST_DATA; + return WRONGCODE; + } xfree (url); url = mynewloc; - url_free (u); + if (orig_parsed != u) + { + url_free (u); + } u = newloc_parsed; /* If we're being redirected from POST, we don't want to POST - again. Many requests answer POST with a redirection to an - index page; that redirection is clearly a GET. We "suspend" - POST data for the duration of the redirections, and restore - it when we're done. */ + again. Many requests answer POST with a redirection to an + index page; that redirection is clearly a GET. We "suspend" + POST data for the duration of the redirections, and restore + it when we're done. */ if (!post_data_suspended) - SUSPEND_POST_DATA; + SUSPEND_POST_DATA; goto redirected; } - if (local_file) + /* Try to not encode in UTF-8 if fetching failed */ + if (!(*dt & RETROKF) && iri->utf8_encode) { + iri->utf8_encode = false; + if (orig_parsed != u) + { + url_free (u); + } + u = url_parse (origurl, NULL, iri, true); + if (u) + { + DEBUGP (("[IRI fallbacking to non-utf8 for %s\n", quote (url))); + url = xstrdup (u->url); + iri_fallbacked = 1; + goto redirected; + } + else + DEBUGP (("[Couldn't fallback to non-utf8 for %s\n", quote (url))); + } + + if (local_file && *dt & RETROKF) + { + register_download (u->url, local_file); + if (redirection_count && 0 != strcmp (origurl, u->url)) + register_redirection (origurl, u->url); + if (*dt & TEXTHTML) + register_html (u->url, local_file); if (*dt & RETROKF) - { - register_download (u->url, local_file); - if (redirection_count && 0 != strcmp (origurl, u->url)) - register_redirection (origurl, u->url); - if (*dt & TEXTHTML) - register_html (u->url, local_file); - } + { + register_download (u->url, local_file); + if (redirection_count && 0 != strcmp (origurl, u->url)) + register_redirection (origurl, u->url); + if (*dt & TEXTHTML) + register_html (u->url, local_file); + if (*dt & TEXTCSS) + register_css (u->url, local_file); + } } if (file) @@ -774,19 +845,22 @@ retrieve_url (const char *origurl, char **file, char **newloc, else xfree_null (local_file); - url_free (u); + if (orig_parsed != u) + { + url_free (u); + } - if (redirection_count) + if (redirection_count || iri_fallbacked) { if (newloc) - *newloc = url; + *newloc = url; else - xfree (url); + xfree (url); } else { if (newloc) - *newloc = NULL; + *newloc = NULL; xfree (url); } @@ -795,59 +869,132 @@ retrieve_url (const char *origurl, char **file, char **newloc, return result; } -/* Find the URLs in the file and call retrieve_url() for each of - them. If HTML is non-zero, treat the file as HTML, and construct - the URLs accordingly. +/* Find the URLs in the file and call retrieve_url() for each of them. + If HTML is true, treat the file as HTML, and construct the URLs + accordingly. If opt.recursive is set, call retrieve_tree() for each file. */ uerr_t -retrieve_from_file (const char *file, int html, int *count) +retrieve_from_file (const char *file, bool html, int *count) { uerr_t status; struct urlpos *url_list, *cur_url; + struct iri *iri = iri_new(); + + char *input_file = NULL; + const char *url = file; - url_list = (html ? get_urls_html (file, NULL, NULL) - : get_urls_file (file)); status = RETROK; /* Suppose everything is OK. */ *count = 0; /* Reset the URL count. */ + /* sXXXav : Assume filename and links in the file are in the locale */ + set_uri_encoding (iri, opt.locale, true); + set_content_encoding (iri, opt.locale); + + if (url_has_scheme (url)) + { + int dt,url_err; + uerr_t status; + struct url * url_parsed = url_parse(url, &url_err, iri, true); + + if (!url_parsed) + { + char *error = url_error (url, url_err); + logprintf (LOG_NOTQUIET, "%s: %s.\n", url, error); + xfree (error); + return URLERROR; + } + + if (!opt.base_href) + opt.base_href = xstrdup (url); + + status = retrieve_url (url_parsed, url, &input_file, NULL, NULL, &dt, + false, iri); + if (status != RETROK) + return status; + + if (dt & TEXTHTML) + html = true; + + /* If we have a found a content encoding, use it. + * ( == is okay, because we're checking for identical object) */ + if (iri->content_encoding != opt.locale) + set_uri_encoding (iri, iri->content_encoding, false); + + /* Reset UTF-8 encode status */ + iri->utf8_encode = opt.enable_iri; + xfree_null (iri->orig_url); + iri->orig_url = NULL; + } + else + input_file = (char *) file; + + url_list = (html ? get_urls_html (input_file, NULL, NULL, iri) + : get_urls_file (input_file)); + for (cur_url = url_list; cur_url; cur_url = cur_url->next, ++*count) { char *filename = NULL, *new_file = NULL; int dt; + struct iri *tmpiri = iri_dup (iri); + struct url *parsed_url = NULL; if (cur_url->ignore_when_downloading) - continue; + continue; if (opt.quota && total_downloaded_bytes > opt.quota) - { - status = QUOTEXC; - break; - } + { + status = QUOTEXC; + break; + } + + /* Need to reparse the url, since it didn't have iri information. */ + if (opt.enable_iri) + parsed_url = url_parse (cur_url->url->url, NULL, tmpiri, true); + if ((opt.recursive || opt.page_requisites) - && cur_url->url->scheme != SCHEME_FTP) - status = retrieve_tree (cur_url->url->url); + && (cur_url->url->scheme != SCHEME_FTP || getproxy (cur_url->url))) + { + int old_follow_ftp = opt.follow_ftp; + + /* Turn opt.follow_ftp on in case of recursive FTP retrieval */ + if (cur_url->url->scheme == SCHEME_FTP) + opt.follow_ftp = 1; + + status = retrieve_tree (parsed_url ? parsed_url : cur_url->url, + tmpiri); + + opt.follow_ftp = old_follow_ftp; + } else - status = retrieve_url (cur_url->url->url, &filename, &new_file, NULL, &dt); + status = retrieve_url (parsed_url ? parsed_url : cur_url->url, + cur_url->url->url, &filename, + &new_file, NULL, &dt, opt.recursive, tmpiri); + + if (parsed_url) + url_free (parsed_url); if (filename && opt.delete_after && file_exists_p (filename)) - { - DEBUGP (("Removing file due to --delete-after in" - " retrieve_from_file():\n")); - logprintf (LOG_VERBOSE, _("Removing %s.\n"), filename); - if (unlink (filename)) - logprintf (LOG_NOTQUIET, "unlink: %s\n", strerror (errno)); - dt &= ~RETROKF; - } + { + DEBUGP (("\ +Removing file due to --delete-after in retrieve_from_file():\n")); + logprintf (LOG_VERBOSE, _("Removing %s.\n"), filename); + if (unlink (filename)) + logprintf (LOG_NOTQUIET, "unlink: %s\n", strerror (errno)); + dt &= ~RETROKF; + } xfree_null (new_file); xfree_null (filename); + iri_free (tmpiri); } /* Free the linked list of URL-s. */ free_urlpos (url_list); + iri_free (iri); + return status; } @@ -868,41 +1015,41 @@ printwhat (int n1, int n2) void sleep_between_retrievals (int count) { - static int first_retrieval = 1; + static bool first_retrieval = true; if (first_retrieval) { /* Don't sleep before the very first retrieval. */ - first_retrieval = 0; + first_retrieval = false; return; } if (opt.waitretry && count > 1) { /* If opt.waitretry is specified and this is a retry, wait for - COUNT-1 number of seconds, or for opt.waitretry seconds. */ + COUNT-1 number of seconds, or for opt.waitretry seconds. */ if (count <= opt.waitretry) - xsleep (count - 1); + xsleep (count - 1); else - xsleep (opt.waitretry); + xsleep (opt.waitretry); } else if (opt.wait) { if (!opt.random_wait || count > 1) - /* If random-wait is not specified, or if we are sleeping - between retries of the same download, sleep the fixed - interval. */ - xsleep (opt.wait); + /* If random-wait is not specified, or if we are sleeping + between retries of the same download, sleep the fixed + interval. */ + xsleep (opt.wait); else - { - /* Sleep a random amount of time averaging in opt.wait - seconds. The sleeping amount ranges from 0 to - opt.wait*2, inclusive. */ - double waitsecs = 2 * opt.wait * random_float (); - DEBUGP (("sleep_between_retrievals: avg=%f,sleep=%f\n", - opt.wait, waitsecs)); - xsleep (waitsecs); - } + { + /* Sleep a random amount of time averaging in opt.wait + seconds. The sleeping amount ranges from 0.5*opt.wait to + 1.5*opt.wait. */ + double waitsecs = (0.5 + random_float ()) * opt.wait; + DEBUGP (("sleep_between_retrievals: avg=%f,sleep=%f\n", + opt.wait, waitsecs)); + xsleep (waitsecs); + } } } @@ -914,7 +1061,7 @@ free_urlpos (struct urlpos *l) { struct urlpos *next = l->next; if (l->url) - url_free (l->url); + url_free (l->url); xfree_null (l->local_name); xfree (l); l = next; @@ -946,7 +1093,7 @@ rotate_backups(const char *fname) rename(fname, to); } -static int no_proxy_match PARAMS ((const char *, const char **)); +static bool no_proxy_match (const char *, const char **); /* Return the URL of the proxy appropriate for url U. */ @@ -959,7 +1106,7 @@ getproxy (struct url *u) if (!opt.use_proxy) return NULL; - if (!no_proxy_match (u->host, (const char **)opt.no_proxy)) + if (no_proxy_match (u->host, (const char **)opt.no_proxy)) return NULL; switch (u->scheme) @@ -994,12 +1141,37 @@ getproxy (struct url *u) return proxy; } +/* Returns true if URL would be downloaded through a proxy. */ + +bool +url_uses_proxy (struct url * u) +{ + bool ret; + if (!u) + return false; + ret = getproxy (u) != NULL; + return ret; +} + /* Should a host be accessed through proxy, concerning no_proxy? */ -int +static bool no_proxy_match (const char *host, const char **no_proxy) { if (!no_proxy) - return 1; + return false; + else + return sufmatch (no_proxy, host); +} + +/* Set the file parameter to point to the local file string. */ +void +set_local_file (const char **file, const char *default_file) +{ + if (opt.output_document) + { + if (output_stream_regular) + *file = opt.output_document; + } else - return !sufmatch (no_proxy, host); + *file = default_file; }