+2001-12-18 Hrvoje Niksic <hniksic@arsdigita.com>
+
+ * recur.c (retrieve_tree): Make a copy of file obtained from
+ dl_url_file_map because the code calls xfree(file) later.
+
2001-12-18 Hrvoje Niksic <hniksic@arsdigita.com>
* recur.c (register_html): Maintain a hash table of HTML files
the second time. */
if (dl_url_file_map && hash_table_contains (dl_url_file_map, url))
{
- file = hash_table_get (dl_url_file_map, url);
+ file = xstrdup (hash_table_get (dl_url_file_map, url));
DEBUGP (("Already downloaded \"%s\", reusing it from \"%s\".\n",
url, file));
- /* #### This check might be horribly slow when downloading
- sites with a huge number of HTML docs. Use a hash table
- instead! Thankfully, it gets tripped only when you use
- `wget -r URL1 URL2 ...', as explained above. */
-
if (string_set_contains (downloaded_html_set, file))
descend = 1;
}