Forum Moderators: open
This time, it's mostly independent websites that were outside our quality guidelines, but that corrected whatever it was (hidden text, hidden links, etc.) and asked to be re-included. Many sites will return to our index within 24 hours. A few sites that used to be way past our guidelines will have to wait for a complete crawl/index cycle before they return completely.
I've said it before, but ultimately each webmaster is responsible for what happens on their own pages. Don't let someone else convince you to put hidden links, hidden links, or other tricks on your domains unless you know what problems it can cause. If you think your site may have penalties for violating our guidelines at one time, but the site has now corrected any problems, you can send an email to webmaster@google.com with the subject line "reinclusion request". Wait to email us until you're sure your site complies with our quality guidelines at
[google.com...]
One of my resolutions for this year is to continue to communicate and to increase communication with webmasters when they've run into problems. The process we're putting into place for these reinclusion requests should help make that easier for webmasters who have made mistakes.
Hope this little GoogleGram helps,
GoogleGuy (topping 900 posts! Woohoo!)
Hey GoogleGuy
Hope you spot/read this post. I was on google.com and reading information about robots.txt files on this specific page: [google.com...]
Under "Technical Guidelines" the link for [robotstxt.org...] is broken. Thought I'd share so it could get fixed.
Bradley
EDIT: What I'm trying to say is go here: [google.com...] and then click on the link to spot the problem.
[robotstxt.org...]
Good note Bradley -- we're all only human! :)
[edited by: fathom at 4:32 am (utc) on Mar. 6, 2003]
Obviously the most anticipated update in the history of Google watching. The last week has seen over 10,000 message posted in our Google forum about the impending monthly update. For the first time we can remember, Google gives a bit of heads up about the impending update.
That's excellent copy GG. (mainpage)
I love that last line which suggests this is now the routine! LOL :)
I understand that but the first "real" ODP update which worked completely since September was dated in mid February. (I saw 19-Feb-2003 but some say there was one a week earlier.)
That seems quite late in the Google crawl, calculate, and update cycle (to me.) I'm glad to hear that they were able to incorporate one of these newer ODP dumps. Radical structure changes have been made since September.
I've submitted my new DN to google and requested many sites to direct their inbound links to my new site, including the links in yahoo's open directory.
Before dissalowing googlebot from crawling either of my sites I wanted to see which site would get the highest ranking and then dissalow googlebot from the lower ranking site.
After having read googleguy's good news, I checked the status of my banned site and it is ranked a level 5 again!
The new domain name is not listed at the moment but I'm afraid it could be listed tomorrow.
My worst nightmare would be to see BOTH sites come up in serps and have someone rat on me and then have google ban BOTH my sites.
I MUST avoid this nightmare. But it may be too late to dissalow googlebot from the new domain name.
I feel as if I'm between a rock and a hard place...or in quick sand. Any suggestion?
BTW, what is the command in robots.txt to only ban googlebot?
User-agent: *
Disallow: /
This will disallow ALL SPIDERS from indexing any of the domain (however, if you have multiple subdomains you will need to create a robots.txt for each subdomain).
If you wish to just disallow GoogleBot, and let the other spiders have access then you simply change the user agent to:
User-agent: GoogleBot
Disallow: /
*** For more information on how to "Remove Content from Google's Index" simply do a search on Google for: google remove
Follow their directions carefully... note from what I have read of this webpage you will have to first put your robots.txt in your root directory stating the information to ignore, then wait for your site to be listed, then follow the directions to remove it.
Hope that helps...
- canuck
"...Check for broken links..."
I love the fact that they have a broken link on the same page!
[#1787748]
[#1769996]
[#1769996]
I can fill the wall with this numbers
Hercules
Since my files are in .html and not .php, I'm using an SSI command as this example:
HTML file: <!--#include file="myfile.php" -->
PHP file: <?include("http://www.MyOldDomainName.com/myfile.htm")?>
All my files at MyNewDomainName.com bring up the files from MyOldDomainName.com
Is it possible to use a redirect in this case?
Many websites have links to dozens different files in MyOldDomainName.com and folks access my site through their bookmark.
I read the following on a post at [webmasterworld.com...] :
----------------------------------------------
redirectpermanent / [newurl.com...] Now, that would send everyone to the front page of your new location. If they had a specific page of the old site bookmarked, it might be somewhat annoying to be sent to the front page of the new location instead.
If the site at the new URL is an exact duplicate of the old one, you might want to put up a page-by-page redirect .htaccess at the old site, so everyone would be sent where they wanted to go...
redirectpermanent /oldpage1.htm [newurl.com...]
redirectpermanent /oldpage2.htm [newurl.com...]
I've done the same thing with mod_rewrite more complex but will allow old bookmarks and links to work ... (which should be hunted down).
also see:
[webmasterworld.com...]
[webmasterworld.com...]
rewriteEngine on
rewriteBase /
rewriteCond %{HTTP_HOST} olddomain.com$ # only required if on same webspace & dir
rewriteRule ^(.+) [newdomain.co.uk...] [L,R=301]
--------------------------------------------------
So my question is: will any of the above redirect command work on MyNewDomainName.com which uses PHP to pull files from MyOldDomainName.com and if yes, which one would be best, considering that I have inbound links to dozens of different files.
Yes, it's always stressful.
For me, it's a chase to keep cleaning old but critical sites. These were created in the old days, when cross linking wasn't seen as a problem. I also wasn't quite so mature then, and didn't consider the value of a social contract (between search engine and webmaster).
My sites are not 'wham bam' selling sites. I add value for the searcher by adding stacks of free information and background. I have done the same with the old sites, and feel with certainty that they deserve top ranking, on the merit of their content. I have therefore fulfilled my half of the 'social contract' and feel ethically sound.
The problem is that the linking structures are a real mess for some of these. I therefore spend time every month trying to clear that up, so that I don't trip an over-zealous spam filter.
The problem is staying ahead of the filters. I will continue to clear up anything that could be considered in any way dodgy (a definition that consistantly becomes stricter) and hope that as Google's filters evolve I don't slip behind the frontier in a given month.
It's hard re-visiting this stuff all the time and of course stressful. Even more so when rumours abound that the next update is going to be an anti-spam update. I think I have evolved from tuna to dolphin... but of course the slightest hint of tuna could get me netted!
Roll on the update therefore, then I can breathe easily again... or panic! And Google... take it easy with us dolphins!