Forum Moderators: Robert Charlton & goodroi
I am thinking about using the robots.txt and the Google url removal tool to get rid of those pages.
Googlebot supports wildcards so:
User-agent: Googlebot
Disallow: /pictures/index.php?sessionid*
Should work right?
Or would Google also remove the index.php?
Haven't go the faintest idea....
regards,
Bert
Anyway, I have added :
Disallow: /pictures/index.php?sessionid*
And did a removal request....
Takes 5 days....
If Google also removes index.php without the sessionid I will move it all to a new directory and start from scratch.
This site is a mess anyway after proxies etc :)
The index.php with sessionid has been dropped by Google...
with this line:
Disallow: /pictures/index.php?sessionid*
I then added this to the robots.txt:
Disallow: /pictures/*sessionid*
and tested it with the robots.txt analysis tool, and entered an url like this:
[mywidgets.com...]
To test, and the url was correctly filtered out..., No error message...
Ha, that would be easy for the removal tool, but unfortunately this message:
URLs cannot have wild cards in them (e.g. "*"). The following line contains a wild card:
DISALLOW /pictures/*sessionid*
So it seems the "ordinary" bot understands, but the removal tool doesn't....