homepage Welcome to WebmasterWorld Guest from 54.242.231.109
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Become a Pro Member

Home / Forums Index / Google / Google SEO News and Discussion
Forum Library, Charter, Moderators: Robert Charlton & aakk9999 & brotherhood of lan & goodroi

Google SEO News and Discussion Forum

    
Making a comback after being penalized by Google
Who's made it back - Who hasn't?
12scott34




msg:763852
 10:59 pm on May 24, 2005 (gmt 0)

Hey all,
I own a company that is 100% pure content and sell great products. I do nothing shady and made sure all links pointing at my site are quality links.

Here is the kicker! I can't even get indexed by Google... I think they put some kind of penalty of my site and I have no idea why. I'm doing everything I can to make a comback but no luck. I feel like MCHammer trying to make a comback right now : )

Anyone had this issue and was able to get indexed? I'm feeling alone out here people. WOuld love to hear from people who have had this happen to them. Would love even more if someone had some advice for me.

SCott

 

BeeDeeDubbleU




msg:763853
 11:20 am on May 25, 2005 (gmt 0)

Scott, one of my sites was dropped at the end of 2003 and it did not get back until about May 2004. During this period I worked at the site improving and adding content, checking links and generally making sure that it was clean. When it did get back it quickly moved up the rankings and it now holds lots of number ones and very high positions in my niche so it was all worthwhile.

Having said that, this is not an ecommerce site per se. I use it sell my services as a consultant, etc. It is also recognised as an authority site in my industry so it should not have been dropped in the first place. I am afraid that apart from the above I cannot help you with this. In fact another of my sites disappeared last weekend so I am now facing the same process with it.

mtishetsky




msg:763854
 11:31 am on May 25, 2005 (gmt 0)

Scott, were you indexed at all or you cannot get indexed for the first time? Are you sure you don't forbid G to crawl your site? I mean robots.txt, meta robots=index,follow and so on.

BTW, your site may be hosted on IP which was used for sending spam or maybe you are linking to someone banned. If a banned site links to you - it's not a problem because you cannot control who links to you. But if YOU link to such a site then it can hurt a lot.

I did come back from G ban. I used search spamming and got banned for about half a year (if not more). But the next day after ban I removed all prohibited content from my site and wrote a begful email to G team. They told they will review my site and remove ban - and they did it. Not too soon of course, but still. Now I keep my site as legal as it is possible.

jaffstar




msg:763855
 11:34 am on May 25, 2005 (gmt 0)

I have a site that has not been indexed in G for about 6 months. I have links to it from some pr5 sites.

This is a brand new site, never been indexed.

Maybe I should wait till the next PR update.

guitaristinus




msg:763856
 11:57 am on May 25, 2005 (gmt 0)

I have about three sites that have been banned from Google for several years or more. I don't think they'll ever make it back in and I don't think they should be.

I emailed Google a few times asking if/why I was banned and got auto response along the lines of "we don't guarantee sites will be indexed."

The following from mtishetsky should give you hope if you have a site worthy of Google's index.
I did come back from G ban. I used search spamming and got banned for about half a year (if not more). But the next day after ban I removed all prohibited content from my site and wrote a begful email to G team. They told they will review my site and remove ban - and they did it. Not too soon of course, but still. Now I keep my site as legal as it is possible.

Some sites get dropped because of algo changes and then make a come back when the algo changes again. Nothing to do about this.

If descriptions of your products are the same as other sites, then you may have a duplicate penalty. Do a search for a dozen words (in quotes) of content and see if Google has them already indexed with another site.

12scott34




msg:763857
 6:25 pm on May 25, 2005 (gmt 0)

Hi all,
Thanks for posting here. I definitely don't feel alone anymore : )

To answer a few questions. My site has never made it into the Google index. Not even once. The only people I link to are the top authoratative sites in my market.

One of you said that you emailed Google a ton and eventually had success. What specific email did you send to? I've seen a ton of them out there so it would be cool to know the exact one to send to.

Anyone ever worked with a company called Proceed Interactive? They say they have connections at Google and could probably get me in the index. They also want about 5 grand out of me to do so. Any thoughts on this?

12scott34




msg:763858
 6:28 pm on May 25, 2005 (gmt 0)

One more thing. One of you mentioned

Are you sure you don't forbid G to crawl your site? I mean robots.txt, meta robots=index,follow and so on.

These are the exact MetaTags I use on my site that relate to robots.

<meta name="robots" content="all=index,follow">
<meta name="revisit-after" content="15 days">

These wouldn't hurt me would they? Could this be my problem?

jimbeetle




msg:763859
 7:02 pm on May 25, 2005 (gmt 0)

Welcome to WW, 12scott34.

The "revisit-after" meta won't hurt, but it won't help either as it's not used by any SE. See: [webmasterworld.com...]

However, your "robots" meta can hurt as the syntax is incorrect. Not sure what a bot would do when running across this, probably get pretty confused if nothing else.

For what you to do, the correct syntax would be:

<meta name="robots" content="all">

--or--

<meta name="robots" content="index,follow">

However, using the tag to allow all bots is not really necessary as most bots consider the absence of an explicit "robots" meta to be the same as "index,follow". Most uses of the robots meta nowadays are to disallow bots using the noindex and nofollow directives.

See [robotstxt.org...] for more on the robots meta. (And, if you've already been there, yep, the explanation in the paragraph "What to put into the Robots META tag" is a bit confusing with examples like "ALL=INDEX,FOLLOW").

For your purposes I think it best just to strip out both the revisit-after and robots metas and see what happens.

12scott34




msg:763860
 7:10 pm on May 25, 2005 (gmt 0)

Hi jimbeetle

Thanks so much for the info. I'm going to strip these off my site and see what happens. Hopefuly this is what was causing us not to get indexed. Maybe all we were doing was confusing the bot.

I'll definitely let you know what happens.

Thanks for the info again!

Scott

privacyman




msg:763861
 8:12 pm on May 25, 2005 (gmt 0)

Here's the another issue....

Even though a person may have a valid robots.txt file which does not block Google or any other desired S/E.... in the case of Google, they refuse to provide a list of IP numbers that their googlebot robots operate upon.... their feeble comment implies that they have too many IP numbers or that they change. That is no excuse, such a list could be the "currently used IP numbers" and that list could change with whenever they change the IP number that any particular googlebot robot is working upon.

In the case of my own site and a few others that I manage, "because bad bots don't always obey the robots.txt file" it was decided to use the better control to prevent bad bots, email harvesters, code thiefs, or other unwanted.... the better control via the .htaccess file (that's dot htaccess). With it, unwanted "referrer" strings are used, unwanted User Agents (bot names etc, including some bad plugins, website copiers), and individual IP numbers are used as well as ranges of IP numbers or entire CIDR groups.

As a result using the dot htaccess control file all of the bad bots are normally kept away from our sites, gives us a more accurate count of actual visitors too. By blocking entire ranges or CIDR groups we keep out most "unwanteds".... that includes some Known Servers or Hosts that "knowingly" give hosting or dialup service to "known spammers" or the like, and it includes countries "known for" being major sources of spam (eg many countries in Asia.... our content is not meant for that audience as we do not do business with Asia).

The BAD PART is that where we have blocked individual IP numbers, ranges or groups, or entire CIDR's, without knowing what IP numbers that "googlebot" operates from (or any other critical indexing bot they may use) it could be that we might have blocked one of their indexing bots.

Google and others do specifiy what their bot name(s) are, eg, googlebot and google-image (think that's correct), but do they have any critical "unnamed bot" that they use for indexing. THEY DO MENTION about having a double-check system to discover cloaking such that googlebot might visit a site, look at a page, then their "other" bot by no name or another name will look at the same page to see if content is different, thus discovering "cloaking" for different bots.

WHAT HAPPENS if one has accidentally blocked one of google's unnamed bots (which might seem to be a harvester or an unwanted or might be in an unwanted group)?

Google or other bots with "unnamed" bots that do double checking would not necessarily have to reveal the exact IP number of their cloak-checking-bots which would kill their means of cloak-checking, but if they said that they have bots working in a CIDR or range that contained dialups or other services then at least we would know Not to block that range.

I had also heard that if a person uses the "noarchive" meta tag in individual pages that a page or site could completely be dropped, likewise I heard that if one blocks some of the image harvesting bots that a page or site could be dropped. Don't know if either of these are true or not.

As for any bot looking for images, I don't feel that any images should be ever be harvested, often later they are infringed upon by the creation of thumbnails, which is an unauthorized duplicate or miniaturization, created without specific authorization and without specification of copyright and no control against anyone Copying and Using that thumbnail version. The original owner could not Sue a user of that thumbnail, but technically the original owner could Sue the thumbnail maker for infringement. THUS, no image should be grabbed and be archived.... ideally it would be better for having a submit page just for images where such authorization could be given for cached copies And/Or for thumbnails creation, but only for a particular S/E.

If anyone has any more specific info about IP numbers that are actually used for the indexing bots would be appreciated. If the indexing bot is ONLY named "googlebot" and does not involve another critical unnamed bot and we would also know the IP number or range to allow then proper indexing of sites/pages could be done, otherwise it could be possible that google or other good bots could be blocked (unintentionally) just to keep out some other bad ones.

Global Options:
 top home search open messages active posts  
 

Home / Forums Index / Google / Google SEO News and Discussion
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved