X-Git-Url: http://sjero.net/git/?a=blobdiff_plain;f=src%2Fretr.c;h=e619f1cbb74b9ef5b1d50131d8d4ad957c5a6684;hb=9228f0bf53d3b42459daeb28372196a007de3014;hp=dabb407e971257a69acb48d0351667655145a67a;hpb=eb0f9de05d27d0c7c41b1300a84dfb3cd49442c9;p=wget diff --git a/src/retr.c b/src/retr.c index dabb407e..e619f1cb 100644 --- a/src/retr.c +++ b/src/retr.c @@ -5,8 +5,8 @@ This file is part of GNU Wget. GNU Wget is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by -the Free Software Foundation; either version 2 of the License, or -(at your option) any later version. +the Free Software Foundation; either version 2 of the License, or (at +your option) any later version. GNU Wget is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of @@ -15,7 +15,17 @@ GNU General Public License for more details. You should have received a copy of the GNU General Public License along with Wget; if not, write to the Free Software -Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. */ +Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA. + +In addition, as a special exception, the Free Software Foundation +gives permission to link the code of its release of Wget with the +OpenSSL project's "OpenSSL" library (or with modified versions of it +that use the same license as the "OpenSSL" library), and distribute +the linked executables. You must obey the GNU General Public License +in all respects for all of the code used other than "OpenSSL". If you +modify this file, you may extend this exception to your version of the +file, but you are not obligated to do so. If you do not wish to do +so, delete this exception statement from your version. */ #include @@ -58,7 +68,7 @@ int global_download_count; static struct { long bytes; - long dltime; + double dltime; } limit_data; static void @@ -74,26 +84,26 @@ limit_bandwidth_reset (void) TIMER the timer, and ADJUSTMENT the previous. */ static void -limit_bandwidth (long bytes, long delta) +limit_bandwidth (long bytes, double delta) { - long expected; + double expected; limit_data.bytes += bytes; limit_data.dltime += delta; - expected = (long)(1000.0 * limit_data.bytes / opt.limit_rate); + expected = 1000.0 * limit_data.bytes / opt.limit_rate; if (expected > limit_data.dltime) { - long slp = expected - limit_data.dltime; + double slp = expected - limit_data.dltime; if (slp < 200) { - DEBUGP (("deferring a %ld ms sleep (%ld/%ld) until later.\n", + DEBUGP (("deferring a %.2f ms sleep (%ld/%.2f).\n", slp, limit_data.bytes, limit_data.dltime)); return; } - DEBUGP (("sleeping %ld ms\n", slp)); - usleep (1000 * slp); + DEBUGP (("sleeping %.2f ms\n", slp)); + usleep ((unsigned long) (1000 * slp)); } limit_data.bytes = 0; @@ -125,13 +135,13 @@ limit_bandwidth (long bytes, long delta) from fd immediately, flush or discard the buffer. */ int get_contents (int fd, FILE *fp, long *len, long restval, long expected, - struct rbuf *rbuf, int use_expected, long *elapsed) + struct rbuf *rbuf, int use_expected, double *elapsed) { int res = 0; - static char c[8192]; + static char c[16384]; void *progress = NULL; struct wget_timer *timer = wtimer_allocate (); - long dltime = 0, last_dltime = 0; + double dltime = 0, last_dltime = 0; *len = restval; @@ -154,7 +164,7 @@ get_contents (int fd, FILE *fp, long *len, long restval, long expected, res = -2; goto out; } - if (opt.verbose) + if (progress) progress_update (progress, sz, 0); } @@ -202,7 +212,7 @@ get_contents (int fd, FILE *fp, long *len, long restval, long expected, last_dltime = dltime; } - if (opt.verbose) + if (progress) progress_update (progress, res, dltime); *len += res; } @@ -213,7 +223,7 @@ get_contents (int fd, FILE *fp, long *len, long restval, long expected, res = -1; out: - if (opt.verbose) + if (progress) progress_finish (progress, dltime); if (elapsed) *elapsed = dltime; @@ -226,7 +236,7 @@ get_contents (int fd, FILE *fp, long *len, long restval, long expected, appropriate for the speed. If PAD is non-zero, strings will be padded to the width of 7 characters (xxxx.xx). */ char * -retr_rate (long bytes, long msecs, int pad) +retr_rate (long bytes, double msecs, int pad) { static char res[20]; static char *rate_names[] = {"B/s", "KB/s", "MB/s", "GB/s" }; @@ -246,7 +256,7 @@ retr_rate (long bytes, long msecs, int pad) UNITS is zero for B/s, one for KB/s, two for MB/s, and three for GB/s. */ double -calc_rate (long bytes, long msecs, int *units) +calc_rate (long bytes, double msecs, int *units) { double dlrate; @@ -254,9 +264,9 @@ calc_rate (long bytes, long msecs, int *units) assert (bytes >= 0); if (msecs == 0) - /* If elapsed time is 0, it means we're under the granularity of - the timer. This often happens on systems that use time() for - the timer. */ + /* If elapsed time is exactly zero, it means we're under the + granularity of the timer. This often happens on systems that + use time() for the timer. */ msecs = wtimer_granularity (); dlrate = (double)1000 * bytes / msecs; @@ -274,30 +284,6 @@ calc_rate (long bytes, long msecs, int *units) return dlrate; } -static int -register_redirections_mapper (void *key, void *value, void *arg) -{ - const char *redirected_from = (const char *)key; - const char *redirected_to = (const char *)arg; - if (0 != strcmp (redirected_from, redirected_to)) - register_redirection (redirected_from, redirected_to); - return 0; -} - -/* Register the redirections that lead to the successful download of - this URL. This is necessary so that the link converter can convert - redirected URLs to the local file. */ - -static void -register_all_redirections (struct hash_table *redirections, const char *final) -{ - hash_table_map (redirections, register_redirections_mapper, (void *)final); -} - -#define USE_PROXY_P(u) (opt.use_proxy && getproxy((u)->scheme) \ - && no_proxy_match((u)->host, \ - (const char **)opt.no_proxy)) - /* Maximum number of allowed redirections. 20 was chosen as a "reasonable" value, which is low enough to not cause havoc, yet high enough to guarantee that normal retrievals will not be hurt by @@ -305,9 +291,29 @@ register_all_redirections (struct hash_table *redirections, const char *final) #define MAX_REDIRECTIONS 20 +#define SUSPEND_POST_DATA do { \ + post_data_suspended = 1; \ + saved_post_data = opt.post_data; \ + saved_post_file_name = opt.post_file_name; \ + opt.post_data = NULL; \ + opt.post_file_name = NULL; \ +} while (0) + +#define RESTORE_POST_DATA do { \ + if (post_data_suspended) \ + { \ + opt.post_data = saved_post_data; \ + opt.post_file_name = saved_post_file_name; \ + post_data_suspended = 0; \ + } \ +} while (0) + /* Retrieve the given URL. Decides which loop to call -- HTTP, FTP, FTP, proxy, etc. */ +/* #### This function should be rewritten so it doesn't return from + multiple points. */ + uerr_t retrieve_url (const char *origurl, char **file, char **newloc, const char *refurl, int *dt) @@ -315,14 +321,16 @@ retrieve_url (const char *origurl, char **file, char **newloc, uerr_t result; char *url; int location_changed, dummy; - int use_proxy; char *mynewloc, *proxy; - struct url *u; + struct url *u, *proxy_url; int up_error_code; /* url parse error code */ char *local_file; - struct hash_table *redirections = NULL; int redirection_count = 0; + int post_data_suspended = 0; + char *saved_post_data = NULL; + char *saved_post_file_name = NULL; + /* If dt is NULL, just ignore it. */ if (!dt) dt = &dummy; @@ -336,8 +344,6 @@ retrieve_url (const char *origurl, char **file, char **newloc, if (!u) { logprintf (LOG_NOTQUIET, "%s: %s.\n", url, url_error (up_error_code)); - if (redirections) - string_set_free (redirections); xfree (url); return URLERROR; } @@ -350,55 +356,38 @@ retrieve_url (const char *origurl, char **file, char **newloc, result = NOCONERROR; mynewloc = NULL; local_file = NULL; + proxy_url = NULL; - use_proxy = USE_PROXY_P (u); - if (use_proxy) + proxy = getproxy (u); + if (proxy) { - struct url *proxy_url; - - /* Get the proxy server for the current scheme. */ - proxy = getproxy (u->scheme); - if (!proxy) - { - logputs (LOG_NOTQUIET, _("Could not find proxy host.\n")); - url_free (u); - if (redirections) - string_set_free (redirections); - xfree (url); - return PROXERR; - } - /* Parse the proxy URL. */ proxy_url = url_parse (proxy, &up_error_code); if (!proxy_url) { logprintf (LOG_NOTQUIET, _("Error parsing proxy URL %s: %s.\n"), proxy, url_error (up_error_code)); - if (redirections) - string_set_free (redirections); xfree (url); + RESTORE_POST_DATA; return PROXERR; } - if (proxy_url->scheme != SCHEME_HTTP) + if (proxy_url->scheme != SCHEME_HTTP && proxy_url->scheme != u->scheme) { logprintf (LOG_NOTQUIET, _("Error in proxy URL %s: Must be HTTP.\n"), proxy); url_free (proxy_url); - if (redirections) - string_set_free (redirections); xfree (url); + RESTORE_POST_DATA; return PROXERR; } - - result = http_loop (u, &mynewloc, &local_file, refurl, dt, proxy_url); - url_free (proxy_url); } - else if (u->scheme == SCHEME_HTTP + + if (u->scheme == SCHEME_HTTP #ifdef HAVE_SSL || u->scheme == SCHEME_HTTPS #endif - ) + || (proxy_url && proxy_url->scheme == SCHEME_HTTP)) { - result = http_loop (u, &mynewloc, &local_file, refurl, dt, NULL); + result = http_loop (u, &mynewloc, &local_file, refurl, dt, proxy_url); } else if (u->scheme == SCHEME_FTP) { @@ -406,22 +395,28 @@ retrieve_url (const char *origurl, char **file, char **newloc, retrieval, so we save recursion to oldrec, and restore it later. */ int oldrec = opt.recursive; - if (redirections) + if (redirection_count) opt.recursive = 0; - result = ftp_loop (u, dt); + result = ftp_loop (u, dt, proxy_url); opt.recursive = oldrec; /* There is a possibility of having HTTP being redirected to FTP. In these cases we must decide whether the text is HTML - according to the suffix. The HTML suffixes are `.html' and - `.htm', case-insensitive. */ - if (redirections && local_file && u->scheme == SCHEME_FTP) + according to the suffix. The HTML suffixes are `.html', + `.htm' and a few others, case-insensitive. */ + if (redirection_count && local_file && u->scheme == SCHEME_FTP) { - char *suf = suffix (local_file); - if (suf && (!strcasecmp (suf, "html") || !strcasecmp (suf, "htm"))) + if (has_html_suffix_p (local_file)) *dt |= TEXTHTML; } } + + if (proxy_url) + { + url_free (proxy_url); + proxy_url = NULL; + } + location_changed = (result == NEWLOCATION); if (location_changed) { @@ -448,10 +443,9 @@ retrieve_url (const char *origurl, char **file, char **newloc, logprintf (LOG_NOTQUIET, "%s: %s.\n", mynewloc, url_error (up_error_code)); url_free (u); - if (redirections) - string_set_free (redirections); xfree (url); xfree (mynewloc); + RESTORE_POST_DATA; return result; } @@ -461,49 +455,32 @@ retrieve_url (const char *origurl, char **file, char **newloc, xfree (mynewloc); mynewloc = xstrdup (newloc_parsed->url); - if (!redirections) - { - redirections = make_string_hash_table (0); - /* Add current URL immediately so we can detect it as soon - as possible in case of a cycle. */ - string_set_add (redirections, u->url); - } - - /* The new location is OK. Check for max. number of - redirections. */ + /* Check for max. number of redirections. */ if (++redirection_count > MAX_REDIRECTIONS) { logprintf (LOG_NOTQUIET, _("%d redirections exceeded.\n"), MAX_REDIRECTIONS); url_free (newloc_parsed); url_free (u); - if (redirections) - string_set_free (redirections); xfree (url); xfree (mynewloc); + RESTORE_POST_DATA; return WRONGCODE; } - /*Check for redirection cycle by - peeking through the history of redirections. */ - if (string_set_contains (redirections, newloc_parsed->url)) - { - logprintf (LOG_NOTQUIET, _("%s: Redirection cycle detected.\n"), - mynewloc); - url_free (newloc_parsed); - url_free (u); - if (redirections) - string_set_free (redirections); - xfree (url); - xfree (mynewloc); - return WRONGCODE; - } - string_set_add (redirections, newloc_parsed->url); - xfree (url); url = mynewloc; url_free (u); u = newloc_parsed; + + /* If we're being redirected from POST, we don't want to POST + again. Many requests answer POST with a redirection to an + index page; that redirection is clearly a GET. We "suspend" + POST data for the duration of the redirections, and restore + it when we're done. */ + if (!post_data_suspended) + SUSPEND_POST_DATA; + goto redirected; } @@ -512,8 +489,8 @@ retrieve_url (const char *origurl, char **file, char **newloc, if (*dt & RETROKF) { register_download (u->url, local_file); - if (redirections) - register_all_redirections (redirections, u->url); + if (redirection_count && 0 != strcmp (origurl, u->url)) + register_redirection (origurl, u->url); if (*dt & TEXTHTML) register_html (u->url, local_file); } @@ -526,9 +503,8 @@ retrieve_url (const char *origurl, char **file, char **newloc, url_free (u); - if (redirections) + if (redirection_count) { - string_set_free (redirections); if (newloc) *newloc = url; else @@ -542,6 +518,7 @@ retrieve_url (const char *origurl, char **file, char **newloc, } ++global_download_count; + RESTORE_POST_DATA; return result; }