Forum Moderators: open
#
# This file is used to allow crawlers to index our site.
#
# List of all web robots: [robotstxt.org...]
#
# Check robots.txt at:
# [searchengineworld.com...]
## Details about Googlebot available at: [google.com...]
# The Google search engine can see everything
User-agent: Googlebot
Disallow:# All other robots will be restricted from accessing the Google-specific index pages
User-agent: *
Disallow: /google_indexing/
Whats this all about? Are Macromedia getting backhanders from google or vice-versa or something.
-or-
Are Macromedia just being sensible by using the biggest SE.?
In any case, I think that Google is very suspicious. I hope GoogleGuy will answer some of this questions, unless they have no answer. I won't negate their innocence if thet are not proved guilty; but, even so, too many unanswered questions will make me (and probably others) lose my confidence in google.
> -Can a bot access the private files of my website?
-Is this legal?
You mean can Googlebot crack into your hosting account and read your files? No, you can be pretty sure that Google wouldn't want to (they even adhere to Robots Exclusion Protocol). On the other hand, I suspect that the Google Search Appliance [google.com] could be tweaked to spider whatever parts of your site that you wanted it to, so that may be what Macromedia were thinking of doing.
> Don't were penalized those sites that created different contents for the bots and for the users?
Yes, sometimes cloaking is used to get a page indexed for words that are popular, but not directly related. This annoys search engines, but is very uncommon these days.
To me that robots file looks like a template (hence all the comments) that is now outdated. Like CNN ban a whole bunch of pages and directories that no longer exist because their robots file is so out of date.
One other thing - someone the size of macromedia doesn't need to 'cloak' using robots.txt. They could just do IP/useragent cloaking like everybody else. They could even show a different robots txt to Google if they wanted.
>>I won't negate their innocence if thet are not proved guilty
I don't buy the conspiracy theory. I'm struggling to find any evidence that Google is guilty of anything because of Macromedias robots file. If Google is corrupt or favouring certain sites or companies, I'd expect them to be a little more technically accomplished than this ;)
Why would they mess around with robots files when they could just be giving out PR or any other number of much more effective and much more difficult to detect measures?
Or perhaps the fact that practically every site that uses any Flash whatsoever has a link to macromedia.com? ;)