⭐ 欢迎来到虫虫下载站! | 📦 资源下载 📁 资源专辑 ℹ️ 关于我们
⭐ 虫虫下载站

📄 linkextor.pm

📁 美国CMU大学开发的操作系统健壮性评测软件
💻 PM
字号:
package HTML::LinkExtor;=head1 NAMEHTML::LinkExtor - Extract links from an HTML document=head1 SYNOPSIS require HTML::LinkExtor; $p = HTML::LinkExtor->new(\&cb, "http://www.sn.no/"); sub cb {     my($tag, %links) = @_;     print "$tag @{[%links]}\n"; } $p->parse_file("index.html");=head1 DESCRIPTIONThe I<HTML::LinkExtor> is an HTML parser that extract links from anHTML document.  The I<HTML::LinkExtor> is a subclass ofI<HTML::Parser>. This means that the document should be given to theparser by calling the $p->parse() or $p->parse_file() methods.=cutrequire HTML::Parser;@ISA = qw(HTML::Parser);$VERSION = sprintf("%d.%02d", q$Revision: 1.1 $ =~ /(\d+)\.(\d+)/);use strict;use vars qw(%LINK_ELEMENT);# Elements that might contain links and the name of the link attribute%LINK_ELEMENT =( body   => 'background', base   => 'href', a      => 'href', img    => [qw(src lowsrc usemap)],   # 'lowsrc' is a Netscape invention form   => 'action', input  => 'src','link'  => 'href',          # need quoting since link is a perl builtin frame  => 'src', applet => [qw(codebase code)], area   => 'href', frame  => 'src',   # Netscape 2.0 extention embed  => 'src',   # used in Netscape 2.0 for Shockwave and things like that);=over 4=item $p = HTML::LinkExtor->new([$callback[, $base]])The constructor takes two optional argument. The first is a referenceto a callback routine. It will be called as links are found. If acallback is not provided, then links are just accumulated internallyand can be retrieved by calling the $p->links() method.The $base is an optional base URL used to absolutize all URLs found.You need to have the I<URI::URL> module installed if you provide$base.The callback is called with the lowercase tag name as first argument,and then all link attributes as separate key/value pairs.  Allnon-link attributes are removed.=cutsub new{    my($class, $cb, $base) = @_;    my $self = $class->SUPER::new;    $self->{extractlink_cb} = $cb;    if ($base) {	require URI::URL;	$self->{extractlink_base} = URI::URL->new($base);    }    $self;}sub start{    my($self, $tag, $attr) = @_;  # $attr is reference to a HASH    return unless exists $LINK_ELEMENT{$tag};    my $base = $self->{extractlink_base};    my $links = $LINK_ELEMENT{$tag};    $links = [$links] unless ref $links;    my @links;    my $a;    for $a (@$links) {	next unless exists $attr->{$a};	push(@links, $a, $base ? URI::URL->new($attr->{$a}, $base)->abs                               : $attr->{$a});    }    return unless @links;    $self->_found_link($tag, @links);}sub _found_link{    my $self = shift;    my $cb = $self->{extractlink_cb};    if ($cb) {	&$cb(@_);    } else {	push(@{$self->{'links'}}, [@_]);    }}=item $p->linksReturns a list of all links found in the document.  The returnedvalues will be anonymous arrays with the follwing elements:  [$tag, $attr => $url1, $attr2 => $url2,...]The $p->links method will also truncate the internal link list.  Thismeans that if the method is called twice without any parsing inbetween then the second call will return an empty list.Also note that $p->links will always be empty if a callback routinewas provided when the I<HTML::LinkExtor> was created.=cutsub links{    my $self = shift;    exists($self->{'links'}) ? @{delete $self->{'links'}} : ();}# We override the parse_file() method so that we can clear the links# before we start with a new file.sub parse_file{    my $self = shift;    delete $self->{'links'};    $self->SUPER::parse_file(@_);}=back=head1 EXAMPLEThis is an example showing how you can extract links from a documentreceived using LWP:  use LWP::UserAgent;  use HTML::LinkExtor;  use URI::URL;  $url = "http://www.sn.no/";  # for instance  $ua = new LWP::UserAgent;  # Set up a callback that collect image links  my @imgs = ();  sub callback {     my($tag, %attr) = @_;     return if $tag ne 'img';  # we only look closer at <img ...>     push(@imgs, values %attr);  }  # Make the parser.  Unfortunately, we don't know the base yet  # (it might be diffent from $url)  $p = HTML::LinkExtor->new(\&callback);  # Request document and parse it as it arrives  $res = $ua->request(HTTP::Request->new(GET => $url),                      sub {$p->parse($_[0])});  # Expand all image URLs to absolute ones  my $base = $res->base;  @imgs = map { $_ = url($_, $base)->abs; } @imgs;  # Print them out  print join("\n", @imgs), "\n";=head1 SEE ALSOL<HTML::Parser>, L<LWP>, L<URI::URL>=head1 COPYRIGHTCopyright 1996-1998 Gisle Aas.This library is free software; you can redistribute it and/ormodify it under the same terms as Perl itself.=cut1;

⌨️ 快捷键说明

复制代码 Ctrl + C
搜索代码 Ctrl + F
全屏模式 F11
切换主题 Ctrl + Shift + D
显示快捷键 ?
增大字号 Ctrl + =
减小字号 Ctrl + -