Download a Set of URLs with GNU Wget
I had a list of URLs that I wanted to download and it was a pain to do it manually. So end up writing a simple shell script and downloading all of them using GNU Wget. Here’s the shell script (modified the one at http://www.linuxquestions.org/questions/programming-9/shell-script-that-read-each-line-separatly-364259/).
# Set the field seperator to a newline
# Loop through the file
for line in `cat file.txt`;do