Network Security Internet Technology Development Database Servers Mobile Phone Android Software Apple Software Computer Software News IT Information

In addition to Weibo, there is also WeChat

Please pay attention

WeChat public account

Shulou

What is the wget command in the linux command

2025-04-05 Update From: SLTechnology News&Howtos shulou NAV: SLTechnology News&Howtos > Development >

Share

Shulou(Shulou.com)06/01 Report--

In this issue, the editor will bring you about what the wget command in the linux command is. The article is rich in content and analyzes and describes it from a professional point of view. I hope you can get something after reading this article.

For network administrators to often download some software or restore backups from a remote server to a local server, using the Linux wget tool is a good choice.

1. Download a single file using wget

The following example is to download a file from the network and save it in the current directory

Wget http: ``/ / cn``.wordpress.org`` / wordpress- 3``.1-zh_CN.zip displays a progress bar during the download process, including (download completion percentage, bytes already downloaded, current download speed, remaining download time).

2. Use wget-O to download and save under a different file name.

By default, wget commands with the last character following the "/", and the file name is usually incorrect for dynamically linked downloads. Error: the following example downloads a file and saves it with the name download.php?id=1080

Wget http: ``/ / www.w``.centos.bz`` / download``? ``id`` = 1 even if the downloaded file is in zip format, it still uses the download.php?id=1080 command. Correct: to solve this problem, we can use the parameter-O to specify a file name:

Wget-O wordpress.zip http: ``/ / www.w``.centos.bz`` / download`.php? ``id`` = 10803, download with wget-limit-rate speed limit

When you execute wget, it takes up all possible broadband downloads by default. But the speed limit is necessary when you are ready to download a large file and you need to download other files.

Wget-limit-rate=300k http: ``/ / cn``.wordpress.org`` / wordpress- 3``.1-zh_CN.zip4, use wget-c breakpoint to resume upload

Restart the download interrupted file using wget-c:

Wget-c http: ``/ / cn``.wordpress.org`` / wordpress- 3``.1-zh_CN.zip is very helpful for us to download large files suddenly due to network and other reasons, we can continue to download instead of downloading a file again. You can use the-c parameter when you need to continue interrupting downloads.

5. Use wget-b to download in the background

For downloading very large files, we can use the parameter-b for background download.

Wget-b http: ``/ / cn``.wordpress.org`` / wordpress- 3``.1-zh_CN.zipContinuing in background, pid 1840. Output will be written to `wget-log'.

You can use the following command to check the download progress

Tail-f wget-log

6. Download the name of camouflage agent

Some websites can reject your download request by judging that the proxy name is not a browser. However, you can disguise it with the-user-agent parameter.

Wget-user-agent= "Mozilla/5.0 (Windows; U; Windows NT 6.1; en-US) AppleWebKit/534.16 (KHTML, like Gecko) Chrome/10.0.648.204 Safari/534.16" download link

7. Use wget-spider to test the download link

When you plan to download regularly, you should test whether the download link is valid at the scheduled time. We can add the-spider parameter to check.

Wget-spider URL will display if the download link is correct

Wget-spider URL ``Spider mode enabled. Check ``if``remote ``file`` exists. ``HTTP request sent, awaiting response... 200 OK ``Length: unspecified [text`` / html``] ``Remote ``file`` exists and could contain further links, ``but recursion is disabled-not retrieving. This ensures that the download will take place at the scheduled time, but when you give the wrong link, the following error will be displayed

Wget-spider url ``Spider mode enabled. Check ``if``remote ``file`` exists. ``HTTP request sent, awaiting response... 404 Not Found ``Remote ``file`` does not exist-broken Linklink links! You can use the spider parameter in the following situations:

Check the interval before regular download to check whether the website is available to check the dead links on the page of the website.

8. Use wget-tries to increase the number of retries

If there is a problem with the network or downloading a large file may also fail. By default, wget retries the connection to download the file 20 times. If necessary, you can use-tries to increase the number of retries.

Wget-tries=40 URL9, download multiple files using wget-I

First, save a download link file

Cat > filelist.txt url1 url2 url3 url4 then use this file and parameters-I to download

Wget-I filelist.txt10, using wget-mirror mirroring website

The following example is to download the entire site locally.

Wget-mirror-p-convert-links-P. / LOCAL URL-miror: download from account opening image-p: download all files that are displayed normally for the html page-convert-links: after download, convert to local link-P. / LOCAL: save all files and directories to the local specified directory

11. Use wget-reject to filter downloads in the specified format

You want to download a website, but you do not want to download pictures, you can use the following command.

Wget-reject=gif url12, use wget-o to save download information to log file

You do not want the download information to be displayed directly on the terminal but in a log file, you can use the following command:

Wget-o download.log URL13, use wget-Q to limit the total download file size

When you want to download more than 5m files and exit the download, you can use the following command:

Wget-Q5m-I filelist.txt Note: this parameter does not work for a single file download and is only valid for recursive downloads.

14. Use wget-r-A to download files in the specified format

You can use this feature in the following situations

Download all the pictures of a website download all the videos of a website download all the PDF files of a website

Wget-r-A.pdf url15, download using wget FTP

You can use wget to download the ftp link. Download using wget anonymous ftp

Wget ``ftp``-url ftp download authenticated with wget username and password

Wget-``ftp``-user=USERNAME-``ftp``-password=PASSWORD urlwget is an open source software developed under Linux by Hrvoje Niksic and later migrated to various platforms, including Windows. It has the following functions and features:

(1) support breakpoint download function; this is also the biggest selling point of network ants and FlashGet at that time, now, Wget can also use this function, those users whose network is not very good can rest assured; (2) support both FTP and HTTP download methods; although most software can be downloaded using HTTP, sometimes it is still necessary to use FTP to download software; (3) support proxy servers For systems with high security intensity, they generally do not expose their systems directly on the Internet, so supporting agents is a necessary function for downloading software; (4) setting up is convenient and simple; maybe, users who are accustomed to the graphical interface are no longer used to the command line, but the command line actually has more advantages in setting up. At least, the mouse can be clicked many times, and don't worry about whether it is wrong to click the mouse. (5) programs are small and completely free; small programs can be ignored, because now the hard disk is too big; completely free has to be considered, even though there are many so-called free software on the Internet, but the advertisements of these software are not what we like.

Although wget is powerful, it is relatively easy to use. The basic syntax is: wget [parameter list] URL. Let's illustrate the use of wget with specific examples.

1. Download the entire http or ftp site.

The command wget http: ``/ place`.your.url`` / here downloads the home page of http://place.your.url/here. Using-x forces the creation of exactly the same directory on the server, and if you use the-nd parameter, everything downloaded on the server is added to the local current directory.

Wget-r http: ``/ / place`.your.url`` / here this command will download all the directories and files on the server recursively, essentially downloading the entire website. This command must be used carefully, because when downloading, all addresses pointed to by the downloaded site will be downloaded, so if this site references other sites, then the referenced site will also be downloaded! For this reason, this parameter is not commonly used. You can use the-l number parameter to specify the download level. For example, if you download only two layers, use-l 2.

If you want to create a mirror site, you can use the-m parameter, for example: wget-m http://place.your.url/here. Then wget will automatically determine the appropriate parameters to create a mirror site. At this point, wget logs in to the server, reads robots.txt, and executes as specified by robots.txt.

2. Continue transmission at breakpoint. When the file is particularly large or the network is particularly slow, often a file has not finished downloading, the connection has been cut off, at this time the need for a breakpoint to resume. The breakpoint continuation of wget is automatic. You only need to use the-c parameter, for example:

Wget-c http: ``/ / the ``.url.of`` / incomplete/file requires the server to support breakpoint continuation using breakpoint continuation. The-t parameter indicates the number of retries, for example, if you need to retry 100 times, write-t 100, and if set to-t 0, it means an infinite number of retries until the connection is successful. The-T parameter indicates the timeout wait time, such as-T 120, which means that waiting for 120 seconds to connect is considered a timeout.

3. Batch download.

If you have multiple files to download, you can generate a file, write a line of URL for each file, for example, generate the file download.txt, and then use the command: wget-I download.txt so that every URL listed in download.txt will be downloaded. (download the file if the list is a file, and download the home page if the list is a website)

4. Selective download.

You can specify that wget download only one class of files, or nothing at all. For example:

Wget-m-reject=gif http: ``/ / target`.web.site`` / subdirectory means to download http://target.web.site/subdirectory, but ignore the gif file. -File types that are acceptable to accept=LIST, and file types that reject=LIST rejects.

5. Password and authentication.

Wget can only handle websites that use usernames / passwords to restrict access. You can use two parameters:-http-user=USER to set HTTP users-http-passwd=PASS to set HTTP passwords. For websites that need certificates for authentication, you can only use other download tools, such as curl.

6. Use the proxy server to download.

If the user's network needs to go through a proxy server, you can have wget download files through the proxy server. At this point, you need to create a .wgetrc file in the current user's directory. The proxy server can be set in the file: http-proxy = 111.111.111.111http ftp-proxy 8080 represents the proxy server of http and 8080 of ftp, respectively. If the proxy server requires a password, use:-proxy-user=USER to set the proxy user-proxy-passwd=PASS to set the proxy password. Use the parameter-proxy=on/off to use or close the agent. Wget also has many useful features that users need to mine.

Appendix:

Command format: wget [Parameter list] [URL of target software and web page]

-VMagneWhile version displays the software version number and then exits;-HMAE talk help displays the software help information;-emam executewriter command executes a ".wgetrc" command

-the software output information is saved to the file;-the software output information is appended to the file;-dmam talk debug displays the output information;-qmam horse does not display the output information; and-imam InputLay filekeeper file obtains the URL from the file

-whether to download the number of times (0 means infinite)-O-output-document=FILE download file is saved as another file name-nc,-no-clobber do not overwrite existing files-Njue timestamping only downloads newer files than local ones-time-out timeout setup timeout is set for Tmai timeouts seducts-YMH proxytimestamp off shuts down the agent

-nd,-no-directories does not set up a directory.-XQuin force force to set up a directory

-http-user=USER setting HTTP user-http-passwd=PASS setting HTTP password-proxy-user=USER setting proxy user-proxy-passwd=PASS setting proxy password

-rmaine quotient recursive downloads the entire website and directory (use with care)-lmai level benefit numerer download level

-File types that can be accepted by Amam quotation list-File types rejected by Rmam quotation list-Domain names that can be accepted by Dmam quotes list-Domain names rejected by exclude-domains=LIST-related links for downloading relative downloads by follow-ftp-FTP links only downloaded by follow-ftp-hosts can be downloaded from outside hosts-directories allowed by Imam hands dedirectoriesLIST-directories rejected by XMHI hands dedirectoriesList

The Chinese document name is encoded under normal circumstances, but it is normal at-cut-dirs

Wget-r-np-nH-``cut``-`dirs`` = 3 ``ftp``: ``/ / host/test/ test .txt

Wget-r-np-nH-nd ``ftp``: ``/ / host/test/``% B4% FA% B8% D5.txt ``wget "``ftp``: ``/ / host/test/ ``*"``% B4%FA%B8%D5.txt due to unknown reasons, perhaps to avoid special file names, wget will automatically use encode_string to process the part of the grabbed file name, so the patch will treat the encode_string as "% 3A". Restore to ":" with decode_string, and apply it to the directory and file name section. Decode_string is a built-in function of wget.

Wget-t0-c-nH-x-np-b-m-P ``/ home/sunny/NOD32view/`` http: ``/ / downloads1`.kaspersky-labs.com`` / bases/``-o tool is small in size but complete in function. It supports breakpoint download function, supports both FTP and HTTP download methods, and supports proxy server and easy to set up.

What are the versions of Linux? the versions of Linux are Deepin, UbuntuKylin, Manjaro, LinuxMint, Ubuntu and so on. Among them, Deepin is one of the best-developed Linux distributions in China; UbuntuKylin is a derivative release based on Ubuntu; Manjaro is a Linux release based on Arch; LinuxMint's default Cinnamon desktop is similar to Windows XP's easy-to-use; Ubuntu is the Linux operating system based on desktop applications.

This is what the wget command in the linux command shared by the editor is. If you happen to have similar doubts, please refer to the above analysis to understand. If you want to know more about it, you are welcome to follow the industry information channel.

Welcome to subscribe "Shulou Technology Information " to get latest news, interesting things and hot topics in the IT industry, and controls the hottest and latest Internet news, technology news and IT industry trends.

Views: 0

*The comments in the above article only represent the author's personal views and do not represent the views and positions of this website. If you have more insights, please feel free to contribute and share.

Share To

Development

Wechat

© 2024 shulou.com SLNews company. All rights reserved.

12
Report