Welcome to WebmasterWorld Guest from 18.104.22.168
Myself, I've lost a root URL in one site while trying to remove www and no-www duplicate. But it was almost six months ago. It will be six months next week.
And I can expect the site back a week later - because I just noticed some earlier removed URLs marked as expired.
More precisely, all URLs removed six months and one week ago are marked as expired, while all other are still marked as complete as they used to.
I write this, because I want to share with members waiting for their accidentally removed sites to be back, that six months period appears to work, in the contrary to "90 days" we believed before.
I didn't realise that the robots.txt which I submitted should only have included the files I wanted removed and not all the standard folders I don't want google so search, like /images/. Will Google now ignore more robots.txt file in 6 months 6 days time and index all those folders they shouldn't be searching?
You are supposed to submit the website or pages that you want removed. Then after a few days, the remove URl console = [services.google.com:8882...] will show status "removed". Then about 6+ months later that removal will expire.
I can't address what happens to old pages once the time is up.
I removed my entire site because it was under development. Six months and 4 days later, 37 pages (most of the site) appeared with the site: command; 93 backlinks were shown with link:, but nothing would rank well for any term (even very specific terms). The next day my site showed a toolbar PR4 for the home page and TBPR3 for most of the interior pages.
So since the site didn't have hardly any pages and, no inbound links at the time I removed the site, I would have to say that, YES the site appears to have progressed in the index during the removal period (just didn't appear in the SERPs during that period).
I'm guessing my site exited the removal period just like a sandboxed site (all the classic symptoms).
For more information on site/page removal: [google.com...] but this thread has more information than I have seen anywhere else, thanks mostly to Wizard.
Hope that helps a bit.
"Get your web server to respond with a 410 (Gone) for any requests for that page"
All those dross pages I've got in .htaccess as 301 redirects to a new or existing page, so that in the unlikely event someone would find one in their google SERPS and click on it, they would be redirected to a valid existing page.
Isn't that the correct way to go, and aren't the search engines supposed to understand that that is a permanent redirect to a new page, follow the new link and delete the old. It's only Google which seems to keep all these old links for posterity.
I submitted it here.. [services.google.com:8882...] Is that's what's referred as the Google Console? That page states.. "Remove pages or subdirectories using a robots.txt file."
Eh? Why will it remove my Google robots.txt file. Well it hasn't and it seems to be working as it's supposed to. We must be at cross-purposes, I suspect.
I guess I misunderstood your post. I thought you said you submitted your robots.text file to the Remove URL Console.
Yes, the link you posted is also known as the the URL console and the automatic URL removal system [services.google.com].