]> sjero.net Git - wget/blobdiff - src/options.h
[svn] Committed C. Frankel's SSL patch.
[wget] / src / options.h
index cd53f8e5fafee2bf4267bf50a750e8aa53a6f900..e080adb03000c7609c06dcc07ed57ab31f5770d2 100644 (file)
@@ -64,6 +64,9 @@ struct options
   char **domains;              /* See host.c */
   char **exclude_domains;
 
+  char **follow_tags;           /* List of HTML tags to recursively follow. */
+  char **ignore_tags;           /* List of HTML tags to ignore if recursing. */
+
   int follow_ftp;              /* Are FTP URL-s followed in recursive
                                   retrieving? */
   int retr_symlinks;           /* Whether we retrieve symlinks in
@@ -83,10 +86,11 @@ struct options
   char *http_user;             /* HTTP user. */
   char *http_passwd;           /* HTTP password. */
   char *user_header;           /* User-defined header(s). */
+  int http_keep_alive;         /* whether we use keep-alive */
 
   int use_proxy;               /* Do we use proxy? */
   int proxy_cache;             /* Do we load from proxy cache? */
-  char *http_proxy, *ftp_proxy;
+  char *http_proxy, *ftp_proxy, *https_proxy;
   char **no_proxy;
   char *base_href;
   char *proxy_user; /*oli*/
@@ -95,12 +99,14 @@ struct options
   long timeout;                        /* The value of read timeout in
                                   seconds. */
 #endif
-  long wait;                   /* The wait period between retries. */
+  long wait;                   /* The wait period between retrievals. */
+  long waitretry;              /* The wait period between retries. - HEH */
   int use_robots;              /* Do we heed robots.txt? */
 
   long quota;                  /* Maximum number of bytes to
                                   retrieve. */
-  long downloaded;             /* How much we downloaded already. */
+  VERY_LONG_TYPE downloaded;   /* How much we downloaded already. */
+  int downloaded_overflow;     /* Whether the above overflowed. */
   int numurls;                 /* Number of successfully downloaded
                                   URLs */
 
@@ -120,6 +126,9 @@ struct options
   char *useragent;             /* Naughty User-Agent, which can be
                                   set to something other than
                                   Wget. */
+  char *referer;               /* Naughty Referer, which can be
+                                  set to something other than
+                                  NULL. */
   int convert_links;           /* Will the links be converted
                                   locally? */
   int remove_listing;          /* Do we remove .listing files
@@ -134,6 +143,13 @@ struct options
 
   int delete_after;            /* Whether the files will be deleted
                                   after download. */
+
+  int html_extension;          /* Use ".html" extension on all text/html? */
+
+  int page_requisites;         /* Whether we need to download all files
+                                  necessary to display a page properly. */
+
+  struct sockaddr_in *bind_address; /* What local IP address to bind to. */
 };
 
 #ifndef OPTIONS_DEFINED_HERE