There are numerous ways to download a file from a URL via the command line on Linux, and two of the best tools for the job are wget and curl. In this guide, we'll show you how to use both commands to perform the task.

In this tutorial you will learn:
  • Wget vs cURL
  • How to download a file with wget
  • How to download a file with cURL
  • Bash script download examples
Downloading a file from URL via command line on Linux
Downloading a file from URL via command line on Linux
Software Requirements and Linux Command Line Conventions
Category Requirements, Conventions or Software Version Used
System Linux (any distribution)
Software Wget, cURL
Other Privileged access to your Linux system as root or via the sudo command.
Conventions # - requires given linux commands to be executed with root privileges either directly as a root user or by use of sudo command
$ - requires given linux commands to be executed as a regular non-privileged user

Wget vs cURL

Looking for Linux Systems Analyst !

The UAF Geophysical Institute, is looking for an experienced Linux Systems Analyst to join their team of research cyber infrastructure analysts and engineers. LOCATION: Fairbanks, Alaska, USA


Sometimes people get confused over the differences between wget and curl, but actually, it's pretty simple. The confusion stems from both tools being capable of downloading files from the command line. But apart from this overlap in functionality, the commands are totally different. They are both used for different (albeit similar) things.

But we're here to learn about downloading a file from the command line. So, which tool is better for the job? Each tool is usually installed by default on any Linux distribution, so it mostly boils down to user preference.

Wget may have a hair of an advantage because it's a little more straight forward and simple to use. Wget can also download recursively. But curl supports far more protocols outside of FTP and HTTP, and also supports uploading data. As you can tell, they each have their advantages. Regardless of which one you decide to use, you'll be able to follow along on your own system with our example commands below.

How to download a file with wget

Wget makes file downloads painless and easy. The base syntax for downloading a file is very simple:

$ wget
Download progress shown by the wget command
Download progress shown by the wget command

Despite lacking a GUI, wget gives us plenty information about our download, including the progress of the download, transfer speed, and estimated time of completion. The bit of output near the top of the terminal is just wget's attempt to connect to the server to download the file. That output can be useful for troubleshooting when you're having problems downloading a file.

Without supplying any extra parameters in the command, wget will save the downloaded file to whatever directory your terminal is currently set to. If you want to specify where the file should be saved, you can use the -O (output) option in the command.

$ wget -O /path/to/dir/file.tar
Wget allows us to specify where to save a file
Wget allows us to specify where to save a file

To see more examples of wget and learn what else it's capable of, check out our full guide on wget.

How to download a file with curl

Curl is another great utility for downloading files from a URL. By default, curl will download a file to standard output. This might be alright if you're downloading a plain text file or if you are piping the curl command to another tool. But if you're just downloading a file to your PC, you don't want curl to send a bunch of garbled text to your terminal, so you should use the -o (output) option in the command.

curl -o /path/to/dir/file.tar
Download progress shown by the curl command
Download progress shown by the curl command

The output here is similar to wget's where we're shown the current download rate, estimated time of completion, etc. To see more examples of curl and learn what else its capable of, check out our full guide on curl.

Bash script download examples

Wget and curl are very easy to include in a bash script. In these examples, we'll look at how to use either command to download a list of URLs in a text document.

First, let's make a download bash script for wget. You'll need two files - one called which contains our bash script, and one called urls.txt which contains our list of URLs to files that we want to download. Each URL needs to be on its own line.



while read url; do
    wget $url
done < urls.txt

And inside of urls.txt, put your list of files:

This script will loop through our URLs file and execute the wget command for each line. We've kept this script very basic, but you can add as many parameters to the wget command as you'd like.

After you've compiled a list of URLs into urls.txt and pasted the above code into with nano or your favorite text editor, give the file execute permissions and run the script:

$ chmod +x
$ ./

For curl, you can follow the exact same instructions as above, but replace the wget command with curl, like so:


while read url; do
    curl $url -O
done < urls.txt

Notice that we've also appended the -O (note: the O is capitalized) option to our curl command, so curl will download the URLs as files and with their respective names. Without this option, curl would download the files to standard output.


In this guide, we learned about two command line tools that can be used for downloading a URL on Linux: wget and curl. Both are perfect for the job and can perform the task equally well. Be sure to check out their respective full length guides on our site to learn about what else these powerful tools can do.

Submit your RESUME, create a JOB ALERT or subscribe to RSS feed.
Subscribe to NEWSLETTER and receive latest news, jobs, career advice and tutorials.
Get extra help by visiting our LINUX FORUM or simply use comments below.

Comments and Discussions