Use wget instead. You can install it with brew install wget if you have installed Homebrew or sudo port install wget if you have installed. If you ever need to download an entire Web site, perhaps for off-line viewing, wget can do the job—for example: $ wget \ --recursive. On Ubuntu Linux, you also have GET and HEAD, usually installed at /usr/bin/. They let you fetch a URL's HTTP header or the whole page.
wget \. --recursive \ # Download the whole site. --no-clobber \ # Don't overwrite existing files. --page-requisites \ # Get all assets/elements. wget --mirror --convert-links --adjust-extension --page-requisites Pingback: Download a complete single page with wget - justnorris. This will mirror the whole raymondsaumure.com site. pages, you cannot in general archive such a web page using wget or any simple HTTP client.
wget is a nice tool for downloading resources from the internet. The basic usage is wget url: wget raymondsaumure.com Therefore, wget (manual page) + less. HTTRACK works like a champ for copying the contents of an entire site. This tool can Wget is a classic command-line tool for this kind of task. It comes with. You may need to mirror the website completely, but be aware that some links may really dead. You can use HTTrack or wget: wget -r. You can download entire web sites using wget and convert the links to point to local sources so that you can view a website offline.