Forum Moderators: goodroi
A while ago our news blog went of on a tangent and threw up over 400 404 erros.
I have given these error pages a 410 status and blocked them in the robots.txt
However does anyoone know if google will check the urls any way and get the 410 gone or will it just ignore them.
If its just ignore them should i remove the robots txt exclusion and let google spider these urls
Kind Regards
Pete
If they are already indexed, robots.txt will not remove them, though they will eventually 'fall out'.
But if they are pages that should not exist, much better to fix the problem - it could bounce back and hurt you in other ways, especially if other spiders reach them and they end up with incoming links.
A clean site is always the safest option, especially in the long term, when 'new' problems caused by today's detritus may defy diagnosis and waste loads of time and effort.
Thanks for the advice, but are you saying i should remove the robots exclusion and let the 410 gone i have set up kick in.
So that these pages that are 404 errors, get kiled of rather than blocked by robots txt.
Btw waht happened was we added a link to our blog in the format /holidays.htm so the blog software added this link to the end of the posts generating the 404 errors in the format /blog/post/holidays.htm
What a stupid i am....
Pete