| 3:38 pm on Sep 23, 2011 (gmt 0)|
Just ignore these types of errors.
I have hundreds and it doesn't affect a site ranking.
They even wrote about it.
Just let go.
| 6:39 pm on Sep 23, 2011 (gmt 0)|
You can redirect URL requests with spaces to strip the spaces.
The same question was asked and answered in the Apache forum only a day or two ago.
| 6:49 pm on Sep 23, 2011 (gmt 0)|
I look at the third party site's page to see if it appears to be on topic and well visited.
If so I 301 redirect the wrong location to the correct URL and then smile.
| 9:59 pm on Sep 23, 2011 (gmt 0)|
Yup. I have hundreds of these. I just ignore them.
| 10:07 pm on Sep 23, 2011 (gmt 0)|
You're potentially missing out on at least some traffic and link benefit.
It's worth checking where those links come from (using WMT or Analytics).
| 4:20 am on Sep 24, 2011 (gmt 0)|
Most of these sites are scrappers that use a broken link to trick Google.
I wouldn't waste a time looking if they have traffic - don't need their traffic. I would save the 301 (and its machine process) for important cases.
| 12:22 pm on Sep 24, 2011 (gmt 0)|
I've had googlebot test for wordpress on non-wordpress sites several times, their favorite page(on my sites) being xmlrpc.php (wordpress remote posting page). That page outputs just on line of text that differs if you allow/disallow remote publishing, it's a nice lightweight page that also tells google if your site is more vulnerable to hacking.
If you have a wordpress site you can visit example.com/xmlrpc.php to see what I mean. if you don't you can do a Google search for "XML-RPC server accepts POST requests only." WITH quotes and see over a million listings for a page that is rarely ever linked to. Google's doing some detective work beyond just crawling your pages, looking for signatures...
"aha, this is a wordpress site, apply known wordpress filters." - GBot
| 8:46 pm on Sep 24, 2011 (gmt 0)|
You mean, google is engaging in behavior that would get an unfamiliar robot banned on the spot? Say it ain't so :(
| 8:54 pm on Sep 24, 2011 (gmt 0)|
It would make sense that Google profiles sites to and records what software they run on. Certain platforms have severe structure, URL, and crawling issues.
By profiling the sites and then having crawling and indexing fixes in place for each platform, Googlebot wouldn't repeatly beat it's head on the wall over and over again.