Forum Moderators: open
Many here feel that the spam report is useless (even when you mention GG and WebmasterWorld) as well as emailing Google with questions or concerns.
Three full months into the New Year and I wonder how Google is doing. From my limited experience, not well. I emailed GG directly about a spammer, which he confirmed receiving, and the spammer still is there.
What is the opinion/experience of everyone else?
I still have lots of respect for Google. This is not an "anger" based thread, but one of disappointment and growing concern in light of Brett's Entropy Part Two thread which most people missed.
What's everyone else think? And what does GG have to say?
I definately missed it.
I've come to the conclusion that I have to get along with my 'neighbor on the hill,' and hope prospective customers will stop by my poor neighborhood occassionally...........
You can put in all of the spam reports you want and name drop WebmasterWorld, GG, and everyone who is supposed to make a difference, but it's a bitter waste of time.
I'm not siding with Google, either. In my opinion, this is a clear disgrace to what was once a great search engine.
Having said that, I'd be the first to salute Google and GoogleGuy if (and when) they actually take the spam reports seriously.
It's almost April of 2003, and Google still has a major job to do.
I find it ironic and somewhat sad that the one aspect that made them what they are today is the same thing that's causing fierce dissent. This one aspect? Relevant results, which Google cannot have when cheap, low-end spam techniques still reign as king.
Perhaps I'd hold Google in a lesser degree of contempt if the spam problem was one of something more high tech, such as cloaking or whatever that'd require large, lengthy algo adjustments.
In this case, it's amateur night and they are winning.
Not because they’re good, but because Google has become lazy.
At least we have GG here at WebmasterWorld to converse with, and I personally hope GG steps up to the plate. It’d be fantastic to correct this problem through open discussion and debate instead of having the lawyers, media, and other SE’s deal with it on an unmerciful "big business" platform.
I have a glimmer of hope that Google will fix the problem so we can all get back to square one, whether you're an SEO or Surfer Joe.
Say, a separate system, that does more intense review of the site? This would allow the regular crawlers to do their jobs at faster but shallower, so the abuse checking system(s) can do more detailed/deeper but slower review of a reported site(s).
This is still open for abuse - I could report someone several times, from various sources. Some protection can be built in, like drop all other abuse reports if the domain is already in the queue.
Repeat reporting of the same domain from the same address (be it IP and/or e-mail) could be reviewed as "reverse-abuse".
Could accept reporting only from fixed IPs. Could time delay, i.e. cool-down, say 15 days forcing some IPs to expire, and if the report could not be "reverified" by the submitter, then drop it. No throwaway e-mail could be used (yahoo.com/hotmail.com/etc.)
Showing how much (not necessarily what) is in the abuse review queue would also "placate" some of us as to the process is running.
All in all - some code could be thrown at the problem to reduce the overhead in hand-verifying each and every site that is reported in an irrate e-mail. Even if just the reporting would be through a non-free-form form (i.e. radio/checkbox, pick-list only, no texbox to describe the issue), would definitely accelerate the resolution process.
Whatever... I stop ranting now.
Google has defined what is spam and what isn't.
Google's SERPs, at this very moment, are flooded with sites that are in violation of these Google defined spam provisions.
Said sites as a whole are not using proverbial ray guys here, either. All of the spam is low tech and can be found by a twelve year old kid, much less a multi-million dollar algo.
Bottom line: If Google never took an anti-spam stance both in their TOS and here at WebmasterWorld where the so called enemy sleeps, we wouldn't be having this discussion today.
And, we'd also be discussiong an entirely different SE.
Google is being called out on the carpet for not enforcing the very rules we're supposed to conduct ourselves by.
If you can't handle it, then get out of the way so others who want to make a difference can. Then again, if our detractors are employing spam techniques, they have nothing to argue about, do they?
That term competes for relevancy in a page full of valid but irrelevant hits, e.g. jobs, resumes, articles about abc consultants.
I have filled out a spam report and a Dissatisfied report at least twice. There were two such sites hogging the top 4 spots. One was banned and I now rank #1 and #2.
This Freshbot rider employs these practices:
* Human visitors with Flash get whisked away to a product page served by a software vendor from another domain.
* Crawlers and others get gibberish text including the phrase "abc consultant" and links to other pages that lead to the same kind of cloaked page with the Flash ad.
* Re-saving the page daily so as to get a fresh timestamp and thus exploiting the Freshbot effect to be always in the first 10.
As the form complaints didn't work, I tried the advice here of writing to GG. Hard to know who replied, but they advised me as follows:
"Thank you for your note. We appreciate your help in maintaining the quality of our index. Please submit this report at [google.com...] .
Regards,
The Google Team"
I had written asking them to address the algorithm, since it is reasonable that they don't have the time to tackle individual problems in an obscure category. e.g. if crawled copy is identical to cache copy, then do not give a freshbot boost. It is possible that the two data sets are on different machines and this could slow things down.
I have replied saying that I have filled out a spam report at least twice. If I have to fill it out again, it could mean that the older ones were just deleted.
I will give it a rest for a while, but if others are seeing examples of Freshbot Riding, please report the practice.
I mentioned this in another post, and it is something many don't get. If through search engine trickery I get a site about purple penguins to the top for "blue widget sales" THAT is spam because someone searching with that doesn't care about purple penguins. However, if you sell blue widgets, the searcher doesn't care if your "spammy" competitor is #1 on the SERP ahead of you.
I mentioned this in another post, and it is something many don't get. If through search engine trickery I get a site about purple penguins to the top for "blue widget sales" THAT is spam because someone searching with that doesn't care about purple penguins. However, if you sell blue widgets, the searcher doesn't care if your "spammy" competitor is #1 on the SERP ahead of you.
I find this widget terminology confusing because my editorial mind equates it to a tangible object. Let's pick a service example.
Say I am looking for an "astrologer" and nearly all the hits on the page are for such people, even if some are resumes, some are job ads and some are newsletter articles. I have no problem with that.
But my complaint is about a reseller of a horoscope generating software tool that an astrologer might use, or an end user might use. This reseller uses the keyword for the occupation to get a SERP result (replete with the gibberish snippet), which might get a few clickthroughs to the software product. There is nothing here for a person who is looking for the service provider.
I am referring to the misuse of phrases to trick the SEs, not an accidental combination of words.
Other examples:
SEO consultant > points to mass-submission software
Fitness trainer > points to fitness software
Brain surgeon > points to a book "Brain surgery for dummies"
Does Google call this spam?
(My complaint is about riding the Freshbot by updating the time stamp and not changing content - which happens to use the above trickery)
"My two wishes (resolutions? ugh, I hate that word) would be: better communication with webmasters when they've run into problems, and closing up loopholes so that webmasters can worry about good content and not waste time trying to do tricks."
Those were my personal wishes, not Google resolutions. I don't get to decide what Google does each year. :) So the first wish was to improve communication for penalized sites. In the last couple months, I've tried to let people know when penalties were about to expire, etc. We also introduced a process where webmasters that have cleaned up their site can request reinclusion. The process is pretty simple: write to webmaster@google.com with the subject "reinclusion request" and explain what happened and how you cleaned up the site. We have more resources on user support to deal with those requests too. Overall, I feel pretty good about how we're doing on this wish so far this year.
The other resolution was to close loopholes. We've made solid progress on that too, both on things behind the scenes (improving scoring and link analysis) as well as some things such as expired domains that we've communicated to webmasters. When I look back over the last few months, I'm pretty happy with the improvements we've done in both these areas already this year, and I think we'll get even better over time. Maybe I shouldn't have listed how I wanted Google to improve--I don't think most search engines engage in a dialogue like that--but I think it's healthy to get out there with a couple goals to keep things focused. Zapatista, I appreciate the three-month status check. :)
Okay, so now let's talk about the spam report form. :) I want to set expectations about spam reports at the right level. Spam reports give us different views of how to improve our quality. Many reports (e.g. off-topic porn for an innocent search) are addressed manually to get a quicker turnaround. For the most part however, the only way to handle spam in a fair, scalable way is to program a computer do it. We use the data from spam reports to help determine what the biggest issues are, and as training data to make sure that our algorithms work well. We tackle projects in the order that we think will improve our quality the most, but often those scoring changes aren't easily visible to the outside world. One thing about Google is that even when it looks like things are quiet, we're still working to make sure that we improve our quality.
Shew! Hope that provides a little more information and my personal take. Maybe next year I'll play it safe and just resolve to lose a couple pounds, eh? :)
better communication with webmasters when they've run into problems
My problem is that I have nightmares when the update is late ... Could google improve the communication on that? ;)
I've tried to let people know when penalties were about to expire
Many reports (e.g. off-topic porn for an innocent search) are addressed manually to get a quicker turnaround
the only way to handle spam in a fair, scalable way is to program a computer do it
I've got one question though. I have read many people reported this 'hidden div' technique. Will google and/or google algo take into account hidden div at one point? Not that I intend to stuff my site with it if you don't ;)
But it seems to me, here down my ground level, that this 'hidden div' should be fairly straightforward to deal with. Could you (not you GG but Google - well if YOU can do it then feel free:) modify the algo so it says:
'ok div, is it visible, if yes - nothing to do ; if not check whether in this div there isn't h1, h4, links, keyword stuffing, very small text, and if there is, relativize the content (weight less than visible content - or ignore it).
Or may be it takes time because you are trying to get rid of the css hidden spec alltogether? I realize that you have to be able to access the stylesheet and that robot can be disallowed from it ... Damn! seem a lot harder all of the sudden ;)
Leo
I speak for everyone when I say we appreciate GG's presence.
The spam problem seems to still be there but after reading so many posts on this here at WebmasterWorld, I really think we need to redefine what is spam and make it required reading. Many people hit the SPAM ALERT button whenever the competitor is ranking higher than them.
Google doesn't like to issue penalities by hand. I wonder if they could issue "warnings" to wayward webmasters. It's not a hand penality but it would probably have the desired effect. It might work as compromise between those who submit the spam report and Google's goal to not issue hand penalties.
If I got a "warning" from Google, I guarantee it would have the desired affect.
Just An Idea
Zapatista
There are many respected companies around using the wrong SEM/SEO. I am sure that whilst they acknowledge that Google provides them with immense amount of referrals, they have absolutely no idea what strange or hidden things have happened to their site.
A warning from an anonymous source would probably be cast aside as a jealous competitors whining. A polite warning from Google would probably waken up all the right responsible people (problem is Google would have to add a telephone number to make it authentic).
Is there anyway an update could be given by Google as to what constitutes Spam? Also, would you encourage webmasters to report Spam to either you or the form if they find it in other (not related to them) areas?
Each G update, a list of the bad factors which Google has discerned about each site might be generated, this could look like this:
A 'bad technique' listing page by number would be available to site operators -
[code] [description........] [action]
1 - excessive dead links (warning)
2 - no text on page (warning)
3 - linking to bad neighborhood (fatal)
4 - robots.txt prevents crawl (fatal)
etc...
And then a page-by-page listing of what Google found to be the case on the site's pages:
mysite.com: 1,2,5,37,678a
mysite.com/innerpage.htm: 2,36,37
mysite.com/innerpage2.htm: 678a
Any pages which would be '100% OKAY' would not show up.
They could still leave out factors to keep everyone still wondering why their PR dropped (wouldn't want to stop all the fun)
Google could maintain these pages online and for security
purposes only allow the IP that the domain is housed on to retrieve the information.
This would give everybody an equal playing field to know that they have 1) site problems or 2) done something stupid and that's why they are banned -- and allow them to correct things, and become wonderful netizens.
Just a thought.
<i>My problem is that I have nightmares when the update is late ... Could google improve the communication on that? ;) </i>
Why? I'm not sure I have ever read anywhere that there is a set date for Google to update. The problem for Google is that if they set an expectation then people will get riled when that expectation is not met.
Hence there is an expectaction that Google will update/dance at some point but it could be any time.
A small dose of randomness actually helps the topic matter of this thread. If the update/dance time were known then people would work to that, including the spammers etc. The fact that the update/dance time is variable is actually a good thing.