using:
Webinator WWW Site Indexer Version 2.56 (Commercial)
Release: 20010814
We have a page that lists every page in the site, to make the site easier to spider for webinator (as some links are displayed using javascript).
The only problem is, this spider page is getting spidered itself. I've tried using robots.txt in the form of:
Disallow: /dir/spider.html
But this wasn't picked up. I tried using the -x argument but discovered this stopped the page from being read in the first place.
Does anyone have any suggestions? I need to ignore this page but for it to also be read!
Webinator WWW Site Indexer Version 2.56 (Commercial)
Release: 20010814
We have a page that lists every page in the site, to make the site easier to spider for webinator (as some links are displayed using javascript).
The only problem is, this spider page is getting spidered itself. I've tried using robots.txt in the form of:
Disallow: /dir/spider.html
But this wasn't picked up. I tried using the -x argument but discovered this stopped the page from being read in the first place.
Does anyone have any suggestions? I need to ignore this page but for it to also be read!