Assuming you have a file urls.txt
that has, for each row, the name of file to save and the url to fetch, space separated. You can then use the following to download the urls.
parallel -j8 --colsep " " "wget -q -O {1} {2}" < urls.txt
parallel --eta -j4 --colsep " " "wget -q -N -t 1 -T 10 -O {1} {2}" < urls.txt
-q
for quiet
-N
for timestamping
-t
for number of retries
-T
for timeout limit.
wget
info page here
See here for notes on how to prep these images for model training.