31

Is there any free application for downloading an entire site installable on Mac OS X 10.6?

alex
  • 17,684
  • 7
  • 54
  • 77
Am1rr3zA
  • 5,353
  • 10
  • 45
  • 53

8 Answers8

54

You can use wget with it's --mirror switch.

wget --mirror –w 2 –p --HTML-extension –-convert-links –P /home/user/sitecopy/

man page for additional switches here.

For OSX, you can easily install wget(and other command line tools) using brew.

If using the command line is too difficult, then CocoaWget is an OS X GUI for wget. (Version 2.7.0 includes wget 1.11.4 from June 2008, but it works fine.)

John T
  • 163,373
  • 27
  • 341
  • 348
25

I've always loved the name of this one: SiteSucker.

UPDATE: Versions 2.5 and above are not free any more. You may still be able to download earlier versions from their website.

4

HTTrack: http://www.httrack.com.

Found in macOS Homebrew.

Provides ports to Windows, Linux-es, and macOS. Command-line utility on (seemingly?) all OSes, GUI options on some.

Johnny Utahh
  • 581
  • 7
  • 23
MicTech
  • 10,328
  • 5
  • 45
  • 55
3

SiteSuuker has already been recommended and it does a decent job for most websites.

I also find DeepVacuum to be a handy and simple tool with some useful "presets".

Screenshot is attached below.

-

Simple interface with "presets"

PKHunter
  • 215
  • 2
  • 8
2

http://epicware.com/webgrabber.html

I use this on leopard, not sure if it will work on snow leopard, but worth a try

1

A1 Website Download for Mac

It has presets for various common site download tasks and many options for those who wish to configure in detail. Includes UI + CLI support.

Starts as a 30 days trial after which is turns into "free mode" (still suitable for small websites under 500 pages)

Tom
  • 409
  • 4
  • 11
  • 24
1

pavuk is by far the best option ... It is command line but has an X-Windows GUI if you install this from the Installation Disk or download. Perhaps someone could write a Aqua shell for it.

pavuk will even find links in external javascript files that are referenced and point these to the local distribution if you use the -mode sync or -mode mirror options.

It is available through the os x ports project, install port and type

port install pavuk

Lots of options (a forest of options).

Simon Sheehan
  • 9,114
  • 12
  • 52
  • 71
-4

Use curl, it's installed by default in OS X. wget isn't, at least not on my machine, (Leopard).

Typing:

curl http://www.thewebsite.com/ > dump.html

Will download to the file, dump.html in your current folder

Fred
  • 136
  • 2
  • Main problem with that is that that's downloading **the homepage**, not the entire website. – Phoshi Nov 08 '09 at 21:58
  • Well, look at the man page – Fred Nov 08 '09 at 21:59
  • 2
    Last I checked, `curl` doesn't do recursive downloads (that is, it can't follow hyperlinks to download linked resources like other web pages). Thus, you can't really mirror a whole website with it. – Lawrence Velázquez Nov 09 '09 at 00:05
  • Well, then do a quick script to get the links, we are in command line land right? Otherwise, just use a tool with a graphical front end. – Fred Nov 09 '09 at 00:54
  • A quick script, I dare you ;-) Last time I checked, `curl` also didn't even download the media embedded within that single web page. So: I'd love to see that script that, for a single page, 1) fetches all images etcetera and 2) rewrites the HTML to refer to those local copies... ;-) – Arjan Nov 10 '09 at 16:32
  • 2
    (And its name *is* cURL... I think John T's edit were really improving your answer.) – Arjan Nov 10 '09 at 22:04
  • There is no difference, and it's a erroneous way to write. Although often used in marketing and so on. It's a name, so I suppose the correct way would be to capitalize the first letter. Look up the man page in a terminal and you see what I'm talking about, it's either curl or Curl. But what are you arguing about here, something of substance? I think not. – Fred Nov 10 '09 at 23:04