Say I have a site on http://website.com. I would really like allowing bots to see the home page, but any other page need to blocked as it is pointless to spider. In other words
http://website.com & http://website.com/ should be allowed, but
http://website.com/anything and http://website.com/someendpoint.ashx should be blocked.
Further...
So I'm working on a server where I can't install any modules whatsoever. Yes this makes my job difficult.
To complete the rest of my script I need to do something that I had thought would be fairly straightforward but it seems to be anything but.
I'm trying to open an .html file as one big long string. This is what I've got:
ope...
Hi, what i am trying to do is get the contents of a file from another server. Since im not in tune with perl, nor know its mods and functions iv'e gone about it this way:
my $fileContents;
if( $md5Con =~ m/\.php$/g ) {
my $ftp = Net::FTP->new($DB_ftpserver, Debug => 0) or die "Cannot connect to some.host.name: $@";
$ftp->log...
Ip: 67.195.112.247
Mozilla/5.0 (compatible; Yahoo! Slurp/3.0; http://help.yahoo.com/help/us/ysearch/slurp)
System.Web.HttpException: The file '/1338.aspx' does not exist.
IP : 67.195.112.247
Host : b3091104.crawl.yahoo.net
Country: United States
...
I had a look at the source of Slurp and I would love to understand how does slurp() work:
sub slurp {
local( $/, @ARGV ) = ( wantarray ? $/ : undef, @_ );
return <ARGV>;
}
Where is the file even opened?
...