yup, wget -r www.foobar.com. Of course that gets what a browser would "see" not the source code behind dymanic pages, unless of course it's cold fusion ;) If you want to get source code, for dynamic pages or something else that would depend on the situation. Aaron On Wed, 19 Apr 2000, Garrett Nievin wrote: > I think that you can use wget for that. Have not done it myself. > > > Cheers, > Garrett > > On Wed, 19 Apr 2000, Janina Sajka wrote: > > > Hi: > > > > Anyone know how to auto-retrieve an entire www page hierarchy? > > > > I know software like ncftp can and wuftp can tar up an entire directory > > tree, but the pages I need aren't available over ftp, only http. I'd hate > > to have them by hand one at a time, though. > > > > > > -- > Garrett P. Nievin <gnievin at gmu.edu> > > Non est ad astra mollis e terris via. -- Seneca >