I think I may have a dup. content filter applied to my site. Can you explain further what you're referring to when you mention "&filter=0"?
When you do a search on google and come to the end of the results there will be a link and text like this:
"In order to show you the most relevant results, we have omitted some entries very similar to the 35 already displayed.
If you like, you can repeat the search with the omitted results included." If you click onthe link, the &filter=0 gets added to the search url and shows all of the sites that Google didnt show on the first pass - at one time Googleguy confirmed here (I believe around two years ago) that this was an example of the dupe content filter - although I would think that it would also be used for other types of filtering?
Primarily it would knock out the sites that have more than two entries for that keyphrase as Google will only show 2 entries from the same domain in its first pass results.
Thanks for the clarification. My pages do appear, but about 200 ranks below where they previously did.
Hi Billy -
We certainly have it applied to us - dropping us from 50k daily uniques to .... 2k overnight after Feb 2 update.
Most of the "duplicate content" appears to be from sites that are actually posting info and snips from .... our site!
Google keeps telling us "you have no penalty" and I finally realized that they don't consider a filter a penalty. Pretty much same effect though.
Marval - any idea if anybody has had it removed? I do not think the filter is a manual process so removal would need to be from changes to the site.
I wonder when we're going to get someone from Google to acknowledge this is a problem?
I think they know how bad it's become, but (I'm guessing here) they want to continue to use automated processes and those are really subject to spamification by clever sites.
Now that Adsense is so lucrative I think that in the past year or so *hundreds of thousands* of large and small scraper sites have come online. Some are in normal formats (such as directories that post other's content) and automation is never going to detect them as "spam". In fact it's subjective. Our big site has a combination of original content, database stuff, public domain, etc. Few large sites are 100% original, so it must be very hard for the algo.
That said, they should at least find a better site review process - the current emails we get telling us we should just wait for things to shake out, are painfully unhelpful.
Can anyone clarify my doubt please?
Duplicate content for google is what?, having the same content on page word by word or few paragraphs having the same content in a page.
Any Idea how can I know that duplicate content of my page is present on other sites page?
Try copyscape and also just taking unique snips of your info in Google to see if other sites have posted that info.
"filter=0" does this .. according the google API documentation. Have no reason to suspect it does anything else ...
he <filter> parameter causes Google to filter out some of the results for a given search. This is done to enhance the user experience on Google.com, but for your application, you may prefer to turn filtering off in order to get the full set of search results.
When enabled, filtering takes the following actions:
* Near-Duplicate Content Filter = If multiple search results contain identical titles and snippets, then only one of the documents is returned.
* Host Crowding = If multiple results come from the same Web host, then only the first two are returned.
By searching with snips of my content, search result showing other website name first and my website is nowhere in search result.
What you guys suggest rephrasing the content totally for the page on my site will work out?
Hi Rama -
Changing *your own* content is an angle I had not thought about and it's probably worth a try. Most people suggest you contact the domains that took your stuff and tell them to take it down. I think doing BOTH might be a good idea.