[Top][All Lists]

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Bug-wget] Downloading files with specific extensions

From: Jamal Mazrui
Subject: Re: [Bug-wget] Downloading files with specific extensions
Date: Tue, 27 Dec 2011 17:05:48 +0000

It is, indeed, necessary to parse web pages in order to download files with 
particular extensions.  Those web pages could be read in memory, however, 
without being downloaded to disk, themselves.  So, I, too, wish that program 
option were available (unfortunately, I am not able to do the coding, myself).


-----Original Message-----
From: address@hidden [mailto:address@hidden On Behalf Of Micah Cowan
Sent: Thursday, December 22, 2011 3:46 PM
To: Vikram Narayanan
Cc: address@hidden
Subject: Re: [Bug-wget] Downloading files with specific extensions

(2011?12?22? 08:39), Vikram Narayanan wrote:
> On Thu, Dec 22, 2011 at 7:59 PM, Tony Lewis <address@hidden> wrote:
>> Vikram Narayanan wrote:
>>> Isn't it a waste of bandwidth?
>>> Is it not possible to check only the PDF files without downloading the
>> whole content?
>> wget has to download HTML content in order to discover where the PDF files
>> are located,
> But curl does it with ease. :)
> Why can't wget?

I don't really know what you're referring to here. Curl doesn't do
recursive fetches, so it can't possible download "just the PDF files"
from a site by itself. I'm not sure why you think _any_ program can
manage to find all the PDF files on a site, without first downloading
enough information to find where those PDF files might be.


reply via email to

[Prev in Thread] Current Thread [Next in Thread]