tags:

views:

63

answers:

2

I am looking for a way to scrape URLS from a web page and output it to a text file. E.g if a page contains multiple http://example.com/article I want to grab both these URLS and output it to a text file.

+4  A: 

Have a look at WWW::Mechanize.

Example code:

use strict;
use warnings;
use 5.010;

use WWW::Mechanize;

my $mech = WWW::Mechanize->new();
$mech->get('http://example.com/example');
foreach my $link ($mech->find_all_links()) {
    say $link->url_abs();
}
mscha
+3  A: 

Use HTML::SimpleLinkExtor:

use strict;
use warnings;

use HTML::SimpleLinkExtor;

my $extor = HTML::SimpleLinkExtor->new();
$extor->parse_url('http://example.com/article');
my @links = $extor->absolute_links();
Alan Haggai Alavi