[TriLUG] wget, curl ???
James Jones
jc.jones at tuftux.com
Mon Jul 21 22:20:08 EDT 2008
Chess,
That's it!
==========================================================
$ lynx -dump http://www.example.com > log.txt
or if you know what types of files you want listed
$ lynx -dump http://www.example.com | grep ogg > log.txt
==========================================================
Although I used links instead of lynx with the results I wanted
Thanks very much.
jcj
On Mon, Jul 21, 2008 at 10:09 AM, Chess Griffin <chess at chessgriffin.com> wrote:
> James Jones wrote:
>>
>> All,
>>
>> I want to capture a list of files on a website. I don't want to
>> download the files. I thought at first that wget would be the best for
>> this, but it appears that it will download the files.
>>
>> What would be the simplest way to achieve my goal?
>>
>> jcj
>
> A list of the files or a list of the links? If the latter, check out lynx,
> w3m, or one of the other text browsers. Maybe something like this would
> work:
>
> $ lynx -dump http://www.example.com > log.txt
>
> or if you know what types of files you want listed
>
> $ lynx -dump http://www.example.com | grep ogg > log.txt
>
> You can probably fine-tune the grep to eliminate the junk and capture just
> what you want.
>
> There are probably much better ways to this, but lynx was the first thing
> that came to mind.
>
> --
> Chess Griffin
> GPG Key: 0x0C7558C3
> http://www.chessgriffin.com
>
>
> --
> TriLUG mailing list : http://www.trilug.org/mailman/listinfo/trilug
> TriLUG FAQ : http://www.trilug.org/wiki/Frequently_Asked_Questions
>
>
More information about the TriLUG
mailing list