function DUMB(d)
{
var newWindow = window.open("", "linklist", "width=300,height=300");
newWindow.document.open("text/html");
newWindow.document.writeln("<B>URL:</B> " + d.URL + "<BR>");
newWindow.document.writeln("<B>Title:</B> " + d.title + "<BR><BR>");
newWindow.document.writeln("Number of links in this document:");
newWindow.document.writeln(d.links.length);
newWindow.document.writeln("<OL>");
for(var i=0; i < d.links.length; i++)
{
newWindow.document.writeln("<LI> <A HREF=\"");
newWindow.document.writeln(d.links[i]);
newWindow.document.writeln("\">" + d.links[i] + "</A><BR>");
}
newWindow.document.close();
}
|
DUMB(document); |
As a point of interest, given the content of the web, what is the
likelihood of a small cycle? Very high!
|
|
It is a certified *BAD THING* to send a robot out to someone's web server if you don't know what you are doing! |
# /robots.txt file # robbieRobot can roam freely User-agent: robbieRobot Disallow: # DUMB (Dorin's URL Muncher Bot) can't look at anything # (It is not allowed to examine anything whose relative # URL begins with /, ie. everything! User-agent: DUMB Disallow: / # Every other bot can examine anything except /tmp and /logs User-agent: * Disallow: /tmp Disallow: /logs |