wget command is very helpful to download the files by remotely, many functionalities available in wget command. On this post describe to download the entire website from the server.

There are two simple ways, one is 

# wget --limit-rate=200k --no-clobber --convert-links --random-wait -r -p -E -e robots=off -U mozilla http://www.anyanydomainname.com

Run the command in background, use '&' character

# wget --limit-rate=200k --no-clobber --convert-links --random-wait -r -p -E -e robots=off -U mozilla http://www.anyanydomainname.com &

--limit-rate=200k: Set download Limit  to 200 Kb /sec

--no-clobber: do not overwrite any existing files

--random-wait: Random waits between download 

-r: Recursive - downloads full website

-p: downloads everything even pictures

-E: gets the right extension of the file.

-e robots=off: act like we are not a robot.

--convert-links: convert links so that they work locally, off-line

Download a Specific File Only

Do you want to filter the specific file extensions (for example  .PDF, .JPEG), use below command,

To filter for specific file extensions:

# wget -A pdf,jpeg -m -p -E -k -K -np http://anyanydomainname.com/

--2016-07-11 11:35:46--  http://www.anydomainname.com/
Resolving www.anydomainname.com (www.anydomainname.com)...
Connecting to www.anydomainname.com (www.anydomainname.com)||:80... connected.
HTTP request sent, awaiting response... 200 OK
Length: unspecified [text/html]
Saving to: ‘www.anydomainname.com/index.html’

www.anydomainname.com/index.html             [   <=>                                                                     ]  51.75K  68.6KB/s   in 0.8s