I'm working with one site that looks like a classic Panda symptoms (Lost 5-8 places on main keywords on Panda).
The only problem is that the site is full and rich with content. So much so it is hard to ever consider Panda as a concern.
As I get further into the site, I realize that a folder resides on site filled with data files that are the results of surveys. The data is stored in tens of thousands of HTML files that are ridiculously repetitive. If the contents of the folder were taken into account, it would be 80% of the site. Needless to say, locked down the folder and next big Panda flow; it should be back.
The problem is that site has always had a robots.txt directive to block that folder from being indexed from day 1. If Gbot followed the rules, it should have never been crawled.
Has anyone seen this before?
Are our robots.txt files just guidelines (and not rules) for Gbot now?