1script - 2:43 pm on Mar 28, 2010 (gmt 0)
I honestly had to look this up first:
All the above granted, but why would one not moderate the UGC and eliminate the spam? Automation is fun---that's google (and they can kill you)---when taking charge is so easy? Websites are not "set it and forget" doohickies.
doohickey: A thing (used in a vague way to refer to something whose name one cannot recall)
But, yes, I sympathize with your frustration and I do remember names of my sites. But since I also want to remember the names of my children, I cannot claim that I review every single post being made. And, as you correctly point out, automation often fails. So, in most cases the problem with this spam is that Googlebot always manages to snap the cache copy of the post before I delete it. I do RSS pings and I also do Twitter posts (automatically) when a new article is posted and Googlebot is usually there within minutes. The spam post would be gone in 1 to 24 hours and this URL will return 410 Gone but they would not return to read it again for months because there are no longer links to those deleted posts. I was even thinking about creating a page with just links to the recently deleted posts. Never did it: it would look just plain silly: a whole page of links that return 410.
And, yes, sometimes I do miss a spam post completely. It can be a busy day and other new posts push it too far back or the subject of the spam is too close to the subject of the site (happens to my sewing site all the time).
So, I think my approach to spam control is reasonable (automated filters, then manual review) but not infallible.
But if anyone knows a mechanism of announcing to Google that the page is removed, I would appreciate if you share. I do use WMT -> Site Configuration -> Crawler Access -> Remove URL in the most egregious cases but it's a manual process and in many cases I just don't have time. Do they have an API for removals, BTW?