[TriLUG] wget, curl ???

James Jones jc.jones at tuftux.com
Mon Jul 21 22:20:08 EDT 2008


Chess,

That's it!

==========================================================

$ lynx -dump http://www.example.com > log.txt

or if you know what types of files you want listed

$ lynx -dump http://www.example.com | grep ogg > log.txt

==========================================================

Although I used links instead of lynx with the results I wanted

Thanks very much.

jcj

On Mon, Jul 21, 2008 at 10:09 AM, Chess Griffin <chess at chessgriffin.com> wrote:
> James Jones wrote:
>>
>> All,
>>
>> I want to capture a list of files on a website. I don't want to
>> download the files. I thought at first that wget would be the best for
>> this, but it appears that it will download the files.
>>
>> What would be the simplest way to achieve my goal?
>>
>> jcj
>
> A list of the files or a list of the links?  If the latter, check out lynx,
> w3m, or one of the other text browsers.  Maybe something like this would
> work:
>
> $ lynx -dump http://www.example.com > log.txt
>
> or if you know what types of files you want listed
>
> $ lynx -dump http://www.example.com | grep ogg > log.txt
>
> You can probably fine-tune the grep to eliminate the junk and capture just
> what you want.
>
> There are probably much better ways to this, but lynx was the first thing
> that came to mind.
>
> --
> Chess Griffin
> GPG Key:  0x0C7558C3
> http://www.chessgriffin.com
>
>
> --
> TriLUG mailing list        : http://www.trilug.org/mailman/listinfo/trilug
> TriLUG FAQ  : http://www.trilug.org/wiki/Frequently_Asked_Questions
>
>



More information about the TriLUG mailing list