Wget Usage Details, learning and learning

Source: Internet
Author: User
Tags glob ftp protocol

 

 

Use of wget

1) Support for resumable download (2) Both FTP and HTTP download (3) support for proxy server (4) easy to set; 5) small programs, completely free of charge;

Command Format:

Wget [parameter list] [target software and web site]

1. startup parameters

This type of parameter mainly provides some basic information about the software;

-V, -- version: displays the software version number and exits;
-H, -- help: displays the software help information;
-E, -- execute = command to execute a ". wgetrc" command

Each of the above functions has two parameters, which are the same as the long function and can be used. Note that the-e parameter is used to execute a. wgettrc command. The. wgettrc command is actually a list of parameters, which can be written together directly.

2. File Processing Parameters

This type of parameter defines the output mode of the software log file;

-O, -- output-file = file: Save the software output information to the file;
-A, -- append-output = file: append the software output information to the file;
-D, -- debug displays the output information;
-Q, -- Quiet does not display output information;
-I, -- input-file = file: Get the URL from the file;

The above parameters are useful to attackers. Let's take a look at the specific usage;

Example 1: Download the Home Page 192.168.1.168 and display the download information
Wget-dhttp: // 192.168.1.168

Example 2: Download the Home Page 192.168.1.168 without displaying any information
Wget-qhttp: // 192.168.1.168

Example 3: download all the files of the link contained in filelist.txt
Wget-I filelist.txt

Wget-NP-m-L5 http://jpstone.bokee.com // do not download other site content linked to this site, Level 5 directory structure
3. Download Parameters

The download parameter defines the number of download repetitions and the file name to be saved;

-T, -- tries = number indicates the number of downloads (0 indicates infinite times)
-O -- output-document = file: The downloaded file is saved as another file name.
-NC, -- no-clobber do not overwrite existing files
-N, -- timestamping only downloads new files than local
-T, -- timeout = seconds
-Y, -- proxy = On/Off disable proxy

For example, download the first page of 192.168.1.168and Save the input information in the download process to the test.htm file.
Wget-O test.htm http: // 192.168.1.168

4. Directory Parameters

Directory parameters mainly set the correspondence between the directory for saving the downloaded file and the directory of the original file (Server File;

-Nd -- no-Directories
-X, -- force-directories force Directory Creation
Maybe we don't know much about the directory. Let's look at an example.

For example, download the home page of 192.168.1.168 and keep the website structure
Wget-xhttp: // 192.168.1.168

5. HTTP parameters

Set HTTP parameters to attributes related to HTTP downloads;

-- Http-user = User: Set the HTTP user
-- Http-passwd = pass: Set the HTTP Password
-- Proxy-user = User: Set proxy user
-- Proxy-passwd = pass: sets the proxy password.

The above parameters mainly set the user and password for HTTP and proxy;

6. recursive parameter settings

When downloading a directory of a website or website, we need to know the download level, and these parameters can be set;
-R, -- Recursive download the entire website and directory (use it with caution)
-L, -- level = Number download level

Example: Download the entire website
Wget-rhttp: // 192.168.1.168

7. Recursive allow and deny option Parameters

When downloading a website, some files can be downloaded as quickly as possible, such as clips and sounds, which can be set here;

-A, -- accept = list acceptable file types
-R, -- reject = file type rejected by list
-D, -- domains = list acceptable domain names
-- Exclude-domains = List rejected Domain Name
-L, -- relative download link
-- Follow-FTP: only download the FTP Link
-H, -- span-hosts can download external hosts
-I, -- include-directories = List Directory
-X, -- exclude-directories = List reject directory

How to set the proxy server used by wget
Wget can use the user setting file ". wgetrc" to read many settings. Here we mainly use this file
Set the proxy server. The ". wgetrc" file in the user's main directory starts.
Function. For example, if the "root" user wants to use ". wgetrc" to set the proxy server, start with "/root/. wgetrc ".
The following shows the content of a ". wgetrc" file. You can refer to this example to compile your own "wgetrc" file:
HTTP-proxy = 111.111.111.111: 8080
FTP-proxy = 111.111.111.111: 8080
The two lines indicate that the Proxy Server IP address is 111.111.111.111 and the port number is 80. Specify
The proxy server used by the HTTP protocol. The second line specifies the proxy server used by the FTP protocol.

Wget User Guide
Wget is a free tool for automatically downloading files from the network. It supports HTTP, https, and FTP protocols and can use HTTP proxy.

The so-called automatic download means that wget can be executed in the background after the user exits the system. This means that you can log on to the system, start a wget download task, and exit the system. wget will be executed in the background until the task is completed, compared with most other browsers, users need to participate in downloading a large amount of data, which saves a lot of trouble.

Wget allows you to track links on the HTML page and download them to create the local version of the remote server, completely recreating the directory structure of the original site. This is often called recursive download ". During recursive download, wget complies with the robot exclusion standard (/robots.txt). wget can convert the link to a local file while downloading to facilitate offline browsing.

Wget is very stable. It has strong adaptability to unstable networks with narrow bandwidth. if the download fails due to network reasons, wget keeps trying until the entire file is downloaded. If the server interrupts the download process, it will be again connected to the server to continue the download from the stopped place. This is very useful for downloading large files from servers with limited connection time.

Common wget usage
Wget format

Usage: wget [Option]... [url]... use wget for site image:
Wget-r-p-NP-khttp: // dsec.pku.edu.cn /~ Us ..
# Or
Wget-mhttp: // www.tldp.org/ldp/ AB ...download a downloaded file on an unstable network to download it in idle time.
Wget-T 0-W 31-chttp: // dsec.pku.edu.cn/bbc...-O down. log &
# Or read the list of files to be downloaded from filelist
Wget-T 0-W 31-C-bftp: // dsec.pku.edu.cn/linu .. -I filelist.txt-O down. log & the above Code can also be used for downloading when the network is relatively idle. My usage is: in Mozilla, copy the URL link that cannot be downloaded at that time to the memory and paste it into the filelist.txt file. Execute the second code above before going out of the system at night.

Download via proxy
Wget-y on-p-khttps: // sourceforge.net/pr...proxy can be set in the environment variable or wgetrcfile

# Set proxy in Environment Variables
Export proxy = http: // 211.90.168.94: 8080/
# In ~ /. Set proxy in wgetrc
Http_proxy = http://proxy.yoyodyne.com :..
Ftp_proxy = http://proxy.yoyodyne.com:... various options category list
Start
-V, -- version: displays the wget version and exits.
-H, -- help print syntax help
-B, -- after the background is started, it is transferred to the background for execution.
-E, -- execute = command: Execute the command in the '. wgetrc "format. For the wgetrc format, see/etc/wgetrc or ~ /. Wgetrc record and input file
-O, -- output-file = file: Write the record to the file.
-A, -- append-output = file: append the record to the file.
-D, -- debug print debugging output
-Q, -- Quiet quiet mode (no output)
-V, -- verbose lengthy mode (this is the default setting)
-NV, -- Non-verbose turn off the lengthy mode, but not the quiet mode
-I, -- input-file = file: the URL that appears when the file is downloaded.
-F, -- force-HTML treats the input file as an HTML file
-B, -- base = URL uses the URL as the prefix of the relative link in the file specified by the-f-I Parameter
-- Sslcertfile = file: Optional client certificate
-- Sslcertkey = Keyfile the Keyfile of the client certificate is optional.
-- EGD-file = file: Specifies the file name download for the EGD socket.
-- Bind-address = address specifies the local address (host name or IP address, used when there are multiple local IP addresses or names)
-T, -- tries = Number sets the maximum number of attempts to connect (0 indicates no limit ).
-O -- output-document = file: Write the document to the file.
-NC, -- no-clobber do not overwrite existing files or use the. # prefix
-C, -- continue, and then download the files that have not been downloaded
-- Progress = type: set the process bar flag
-N, -- timestamping
-S, -- server-response print the Server Response
-- Spider does not download anything
-T, -- timeout = seconds: set the number of seconds for response timeout.
-W, -- Wait = seconds: the interval between two attempts is seconds.
-- Waitretry = waiting for 1... seconds between reconnections
-- Random-Wait waits for 0 seconds between downloads... 2 * Wait
-Y, -- proxy = On/Off open or close the proxy
-Q, -- quota = Number: Set the download capacity limit.
-- Limit-rate = Rate: Specifies the directory for downloading rates.
-Nd -- no-Directories
-X, -- force-directories force Directory Creation
-NH, -- no-host-directories do not create the host directory
-P, -- directory-Prefix = Prefix: save the file to the directory prefix /...
-- Cut-dirs = Number ignore the HTTP option of the remote directory at the number layer
-- Http-user = User: Set the HTTP user name to user.
-- Http-passwd = pass: Set the HTTP password to pass.
-C, -- cache = On/Off allow/Do Not Allow server-side data caching (generally allow ).
-E, -- HTML-Extension: Save all text/html files with the. html Extension
-- Ignore-length ignore the 'content-length "header field
-- Header = string insert string in Headers
-- Proxy-user = User: Set the proxy username to user.
-- Proxy-passwd = pass: Set the proxy password to pass
-- Referer = the URL contains the 'Referer': URL "header in the HTTP request.
-S, -- save-headers Save the HTTP header to the file
-U, -- User-Agent = agent: Set the proxy name to agent instead of wget/version.
-- No-http-keep-alive disable the HTTP activity Link (permanent link ).
-- Cookies = off do not use cookies.
-- Load-Cookies = file: load the cookie from the file before starting the session
-- Save-Cookies = File Save cookies to file FTP after the session ends Option
-Nr, -- Dont-Remove-listing does not remove the '. Listing "file.
-G, -- glob = On/Off enable or disable the globbing mechanism of the file name
-- Passive-FTP uses passive transmission mode (default ).
-- Active-FTP Active Transmission Mode
-- Retr-symlinks recursively downloads the link to a file instead of a directory during recursion.
-R, -- Recursive recursive download-use with caution!
-L, -- level = maximum recursive depth of number (inf or 0 indicates infinity ).
-- Delete-after: Partial Deletion of objects after completion
-K, -- convert-links converts non-relative links to relative links
-K, -- backup-converted: Back up the file to X. orig before converting file x
-M, -- mirror is equivalent to-r-N-l INF-Nr.
-P, -- page-requisites download show the inclusion and non-inclusion (accept/reject) of all images in the HTML file)
-A, -- accept = List a semicolon-separated list of accepted extensions
-R, -- reject = List a semicolon-separated list of unacceptable extensions
-D, -- domains = List the list of accepted domains separated by semicolons
-- Exclude-domains = List a semicolon-separated list of unacceptable Domains
-- Follow-FTP: Tracking FTP links in HTML documents
-- Follow-tags = List a semicolon-separated list of HTML tags to be tracked
-G, -- ignore-tags = List semicolon-separated list of ignored HTML tags
-H, -- span-hosts is recursively transferred to the external host
-L, -- relative only traces relative links
-I, -- include-directories = List list of allowed Directories
-X, -- exclude-directories = List list of excluded Directories
-NP, -- no-parent should not be traced back to the parent directory

Wget tips

Source: Linux technical backbone site

Wget is used as follows:
Wget [parameter list] URL
First, we will introduce the main parameters of wget:
Expose-B: Run wget in the background. The record file is written in the "wget-log" file in the current directory;
Failed-T [nuber of times]: number of attempts. How many attempts will be made when wget cannot establish a connection with the server?
. For example, "-T
120 "indicates 120 attempts. When this parameter is set to "0", it is specified that the attempt is infinite Multiple times until the connection is successful.
The setting is very useful. When the recipient's server is suddenly shut down or the network is suddenly interrupted, it can continue after it returns to normal.
Download files that have not been uploaded;
Upload-C: resumable upload, which is also a very useful setting, especially when downloading large files, if
If the connection is accidentally interrupted, the connection will be resumed from the last point, instead of starting from scratch.
The remote server also supports resumable data transfer. Generally, the Web/FTP Server Based on Unix/Linux
Supports resumable data transfer;
Timeout-T [number of seconds]: time-out period. It specifies how long the connection will be interrupted if the remote server does not respond.
To start the next attempt. For example, "-T
120 "indicates that if the remote server does not send data after 120 seconds, it will try again. If the network speed ratio
Faster. You can set a shorter time. On the contrary, you can set a longer time. Generally, it cannot exceed 900.
It should be no less than 60, which is generally set to around 120;
Seconds-W [number of seconds]: the number of seconds to wait between two attempts. For example, "-W 100" indicates two attempts.
Wait 100 seconds;
Slave-y on/off: connection through/not through the proxy server;
Restrict-Q [byetes]: Limit the maximum size of the downloaded file. For example, "-q2k" indicates that the total size cannot exceed 2 K.
Byte. "-q3m" indicates a maximum of 3 M bytes. If no number is added, it indicates that the length is a single byte.
For example, "-q200" indicates a maximum of 200 bytes;
Skip-Nd: Do not download the directory structure. heap all files downloaded from all specified directories on the server to the current directory;
Reverse-X: opposite to "-nd", create a complete directory structure, for example, "wget-ndhttp: // www.gnu.org "will be created under the current directory" w ..., then follow the server
The actual directory structure is created at the first level until all files are uploaded;
Keep-NH: do not create a directory named by the target host domain name. directly store the directory structure of the target host to the current directory.
Directory;
Response -- http-user = Username
Authorization -- http-passwd = password: if the web server needs to specify the user name and password, use these two parameters;
Login -- proxy-user = Username
Login -- proxy-passwd = password: If the proxy server needs to enter the user name and password, use these two options
;
Slave-R: Create the server directory structure on the local machine;
Depth-L [depth]: the depth of the remote server directory structure, for example, the depth of the "-l 5" Download directory is smaller than or
Directory structure or file within 5;
Mirror-M: The option when you create a site image. If you want to create a site image, use this option to automatically set
Set other appropriate options for site images;
Upload-NP: only download the contents of the specified directory and Its subdirectories of the target site. This is also a very useful option.
Assume that a person's personal homepage has a connection to the personal homepage of another person on the website, and we only want to download
If this option is not set on the personal homepage of this person, or even -- the whole site may be captured, this is obviously
We usually don't want it;
Ü how to set the proxy server used by wget
Wget can use the user setting file ". wgetrc" to read many settings. Here we mainly use this file
Set the proxy server. The ". wgetrc" file in the user's main directory starts.
Function. For example, if the "root" user wants to use ". wgetrc" to set the proxy server, "/root/. wgert"
Function, the following is a ". WGE
The content of the TRC file. You can refer to this example to compile your own "wgetrc" file:
HTTP-proxy = 111.111.111.111: 8080
FTP-proxy = 111.111.111.111: 8080
The two lines indicate that the Proxy Server IP address is 111.111.111.111 and the port number is 80. Specify
The proxy server used by the HTTP protocol. The second line specifies the proxy server used by the FTP protocol.

Wget instance:
Wget is a command line tool used to download files in batches. It supports HTTP and FTP. What is better than other tools? Check the content.

Wget is basically carried by all Linux versions, but how many users are using it? If you have never used it, try it. Windows users can use the gnuwin32 project, wget, and the basic functions are completely consistent. Let's take a few simple examples to see the power of wget.

If we want to download all the files in a directory in FTP, we don't need to use the FTP dumb. You can easily drag a directory by using graphical tools such as cute FTP. For example

Wget-rftp: // 10.8.8.8/movie/

Wait! When I found something wrong, how did I get out a directory named 10.8.8.8 and check it out? It's a movie. Oh, wget recorded both the directory structure and the website title, don't you ?? No problem! For example

Wget-r-ndftp: // 10.8.8.8/movie/

No directories are returned, faint! Why? If you want to do so, let it do so. Otherwise, use

Wget-r-nhftp: // 10.8.8.8/movie/

Well? And movie? OK.

Wget-r-NH -- Cut-dirs = 1ftp: // 10.8.8.8/movie/

What's the purpose? CuteFTP is much easier to use than him. Besides, can you connect to CuteFTP if it is broken? Sorry, you can connect

Wget-c-r-NH -- Cut-dirs = 1ftp: // 10.8.8.8/movie/

But can CuteFTP do the following? For example, many websites currently use Apache to build websites and do not provide FTP services. However, Apache has an indexing function that provides an FTP-like interface. What should I do if there are many files? Because of the HTTP protocol, CuteFTP is powerless, but flash get and other functions such as get all do not know how they process directories. But wget has no problem at all. Do not believe it? Let's take ctan as an example (this is not an appropriate example, ctan has an FTP service). We will download all the macro packages here.

Wget-r-khttp: // www.txia.com/blog

-K indicates converting the connection to a local connection. But now there are the same problems, so add what you need. In addition, you may not need to go down so many layers. For example, we will go to the second layer.

Wget-r-L2-khttp: // www.txia.com/blog

The new problem is that the webpage has a sorting function, which is annoying because it has been repeated many times during download, you can use the-A and-R switches to control the download type and use wildcards.

Wget-r "*. htm *? * "-Khttp: // www.txia.com/blog

This time there is no such web page, right? -R indicates that a file of the matching type is rejected.-A indicates only the accepted file type, such as-"*. GIF will only download GIF images. If multiple GIF images are allowed or not, they can be used separately.

So what should we do if we are using a proxy server? Haha, I'm glad you chose wget. You can use its configuration file and environment variables to use proxies. We recommend that you use environment variables. For example, in bash, we can add the daily proxy to. bash_profile. This is a standard Linux Statement (many software programs use apt-Get, yum, and so on)

Export http_proxy = http: // 10.20.30.40: 8080

Then, proxy is enabled by default. If you need to temporarily disable it, you can use

Wget -- proxy = OFF-r-khttp: // www.txia.com/blog

Of course, you can also write a. wgetrc file, which can be found in/usr/local/etc. There are detailed comments in it, so I won't say much about it.

When downloading a Web page, it is troublesome that some web pages are directed multiple times at the same time. To avoid multiple downloads, we use

Wget-nc-r-khttp: // www.txia.com/blog

This can be avoided. We should add

Wget-nc-NP-r-khttp: // www.txia.com/blog

Avoid downloading files that are not in this directory, and avoid accessing different hosts. Of course, if you want this function to be available, you can use

Wget-nc-NP-H-r-khttp: // www.txia.com/blog

So that we can span between multiple hosts, while-I and-x allow us to track only some directories or not some directories. If something you need in some HTML is not made by this type of thing, you have to use -- follow-tags and -- ignore-tags.

Hey, I have a file that is connected. What should I do? You only need

Wget-I your. File

If yes, It is not complex.

Wget-f-I your. File

Wget User Guide
Wget is a free tool for automatically downloading files from the network. It supports HTTP, https, and FTP protocols and can use HTTP proxy.

The so-called automatic download means that wget can be executed in the background after the user exits the system. This means that you can log on to the system, start a wget download task, and exit the system. wget will be executed in the background until the task is completed, compared with most other browsers, users need to participate in downloading a large amount of data, which saves a lot of trouble.

Wget allows you to track links on the HTML page and download them to create the local version of the remote server, completely recreating the directory structure of the original site. This is often referred to as "recursive download ". During recursive download, wget complies with the robot exclusion standard (/robots.txt). wget can convert the link to a local file while downloading to facilitate offline browsing.

Wget is very stable. It has strong adaptability to unstable networks with narrow bandwidth. if the download fails due to network reasons, wget keeps trying until the entire file is downloaded. If the server interrupts the download process, it will be again connected to the server to continue the download from the stopped place. This is very useful for downloading large files from servers with limited connection time.

Common wget usage
Wget format

Usage: wget [Option]... [url]...

Use wget for site image:
Wget-r-p-NP-khttp: // dsec.pku.edu.cn /~ Us ..
-R indicates that recursive download will download all links, but note that you should not use this parameter separately, because if the website you want to download also has a link to another website, wget will also download things from other websites. Therefore, the-NP parameter must be added, indicating that the link of other websites will not be downloaded. -K indicates modifying the link in the downloaded webpage to a local link. -P: Get all the elements required to display the webpage, such as slice or something.

# Or
Wget-mhttp: // www.tldp.org/ldp/ AB ..

Download a part of the downloaded files from an unstable network and during free time.
Wget-T 0-W 31-chttp: // dsec.pku.edu.cn/bbc...-O down. log &
# Or read the list of files to be downloaded from filelist
Wget-T 0-W 31-C-bftp: // dsec.pku.edu.cn/linu...-I filelist.txt-O down. log &

The above code can also be used for downloading when the network is relatively idle. My usage is: In mozillawill not easily download the urlchain to the memory, then paste it to the filelist.txt file, and execute the second code above before going out of the system at night.

Download via proxy
Wget-y on-p-khttps: // sourceforge.net/pr ..

The proxy can be set in the environment variable or wgetrc file.
# Set proxy in Environment Variables
Export proxy = http: // 211.90.168.94: 8080/
# In ~ /. Set proxy in wgetrc
Http_proxy = http://proxy.yoyodyne.com :..
Ftp_proxy = http://proxy.yoyodyne.com :..

Wget option category list
Start
-V, -- version: displays the wget version and exits.
-H, -- help print syntax help
-B, -- after the background is started, it is transferred to the background for execution.
-E, -- execute = command: Execute the command in the '. wgetrc "format. For the wgetrc format, see/etc/wgetrc or ~ /. Wgetrc

Record and input file
-O, -- output-file = file: Write the record to the file.
-A, -- append-output = file: append the record to the file.
-D, -- debug print debugging output
-Q, -- Quiet quiet mode (no output)
-V, -- verbose lengthy mode (this is the default setting)
-NV, -- Non-verbose turn off the lengthy mode, but not the quiet mode
-I, -- input-file = file: the URL that appears when the file is downloaded.
-F, -- force-HTML treats the input file as an HTML file
-B, -- base = URL uses the URL as the prefix of the relative link in the file specified by the-f-I Parameter
-- Sslcertfile = file: Optional client certificate
-- Sslcertkey = Keyfile the Keyfile of the client certificate is optional.
-- EGD-file = file specifies the file name of the EGD socket.

Download
-- Bind-address = address specifies the local address (host name or IP address, used when there are multiple local IP addresses or names)
-T, -- tries = Number sets the maximum number of attempts to connect (0 indicates no limit ).
-O -- output-document = file: Write the document to the file.
-NC, -- no-clobber do not overwrite existing files or use the. # prefix
-C, -- continue, and then download the files that have not been downloaded
-- Progress = type: set the process bar flag
-N, -- timestamping
-S, -- server-response print the Server Response
-- Spider does not download anything
-T, -- timeout = seconds: set the number of seconds for response timeout.
-W, -- Wait = seconds: the interval between two attempts is seconds.
-- Waitretry = waiting for 1... seconds between reconnections
-- Random-Wait waits for 0 seconds between downloads... 2 * Wait
-Y, -- proxy = On/Off open or close the proxy
-Q, -- quota = Number: Set the download capacity limit.
-- Limit-rate = Rate: Specifies the download rate.

Directory
-Nd -- no-Directories
-X, -- force-directories force Directory Creation
-NH, -- no-host-directories do not create the host directory
-P, -- directory-Prefix = Prefix: save the file to the directory prefix /...
-- Cut-dirs = Number ignore the remote directory of the number layer

HTTP OPTIONS
-- Http-user = User: Set the HTTP user name to user.
-- Http-passwd = pass: Set the HTTP password to pass.
-C, -- cache = On/Off allow/Do Not Allow server-side data caching (generally allow ).
-E, -- HTML-Extension: Save all text/html files with the. html Extension
-- Ignore-length ignore the 'content-length "header field
-- Header = string insert string in Headers
-- Proxy-user = User: Set the proxy username to user.
-- Proxy-passwd = pass: Set the proxy password to pass
-- Referer = the URL contains the 'Referer': URL "header in the HTTP request.
-S, -- save-headers Save the HTTP header to the file
-U, -- User-Agent = agent: Set the proxy name to agent instead of wget/version.
-- No-http-keep-alive disable the HTTP activity Link (permanent link ).
-- Cookies = off do not use cookies.
-- Load-Cookies = file: load the cookie from the file before starting the session
-- Save-Cookies = file: saves cookies to the file after the session ends.

FTP options
-Nr, -- Dont-Remove-listing does not remove the '. Listing "file.
-G, -- glob = On/Off enable or disable the globbing mechanism of the file name
-- Passive-FTP uses passive transmission mode (default ).
-- Active-FTP Active Transmission Mode
-- Retr-symlinks: recursively points the link to a file instead of a directory)

Recursive download
-R, -- Recursive recursive download-use with caution!
-L, -- level = maximum recursive depth of number (inf or 0 indicates infinity ).
-- Delete-after: Partial Deletion of objects after completion
-K, -- convert-links converts non-relative links to relative links
-K, -- backup-converted: Back up the file to X. orig before converting file x
-M, -- mirror is equivalent to-r-N-l INF-Nr.
-P, -- page-requisites download and display all images of HTML files

Include and exclude (accept/reject) in recursive download)
-A, -- accept = List a semicolon-separated list of accepted extensions
-R, -- reject = List a semicolon-separated list of unacceptable extensions
-D, -- domains = List the list of accepted domains separated by semicolons
-- Exclude-domains = List a semicolon-separated list of unacceptable Domains
-- Follow-FTP: Tracking FTP links in HTML documents
-- Follow-tags = List a semicolon-separated list of HTML tags to be tracked
-G, -- ignore-tags = List semicolon-separated list of ignored HTML tags
-H, -- span-hosts is recursively transferred to the external host
-L, -- relative only traces relative links
-I, -- include-directories = List list of allowed Directories
-X, -- exclude-directories = List list of excluded Directories
-NP, -- no-parent should not be traced back to the parent directory

Problem
During recursive download, when the directory contains Chinese characters, the local directory name created by wget will be processed using URL encoding rules. For example, "Skynet firewall" is saved as "% CC % EC % Cd % F8 % B7 % C0 % BB % F0 % C7 % BD", which makes reading inconvenient.

Contact Us

The content source of this page is from Internet, which doesn't represent Alibaba Cloud's opinion; products and services mentioned on that page don't have any relationship with Alibaba Cloud. If the content of the page makes you feel confusing, please write us an email, we will handle the problem within 5 days after receiving your email.

If you find any instances of plagiarism from the community, please send an email to: info-contact@alibabacloud.com and provide relevant evidence. A staff member will contact you within 5 working days.

A Free Trial That Lets You Build Big!

Start building with 50+ products and up to 12 months usage for Elastic Compute Service

  • Sales Support

    1 on 1 presale consultation

  • After-Sales Support

    24/7 Technical Support 6 Free Tickets per Quarter Faster Response

  • Alibaba Cloud offers highly flexible support services tailored to meet your exact needs.