Welcome to WebmasterWorld Guest from 54.147.212.173

Forum Moderators: Robert Charlton & andy langton & goodroi

Message Too Old, No Replies

Google posting "not found" errors for pages that don't exist

     
6:14 am on Sep 23, 2011 (gmt 0)

New User

joined:Sept 2, 2011
posts:11
votes: 0


In webmaster tools account, I found some urls are going to 404 not found but these are text link on third party site:for

example:

http://www.example.com/ 29-Jun-11
but webmaster treated as a - http://www.example.com/29-Jun-11

n these links are from directories n boomarking type sites.. If the third party site is having some alignment problem and they cant show the complete url in one line then why it is showing in crawl error section? Y it is broken link?

So, my questions are:
1. What can reasonably be done to cause these errors to stop appearing?
2. Any thoughts in terms of how long these errors will continue to appear... months/years?
3. Is there any way to contact Google and have them cut this nonsense out.
4. My site has also been out from google, is 404 error is the reason for it.?..

[edited by: tedster at 2:53 pm (utc) on Sep 23, 2011]
[edit reason] switch to example.com [/edit]

3:38 pm on Sept 23, 2011 (gmt 0)

Preferred Member

5+ Year Member

joined:June 10, 2011
posts: 521
votes: 0


Just ignore these types of errors.
I have hundreds and it doesn't affect a site ranking.
They even wrote about it.
Just let go.
6:39 pm on Sept 23, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member g1smd is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month

joined:July 3, 2002
posts:18903
votes: 0


You can redirect URL requests with spaces to strip the spaces.

The same question was asked and answered in the Apache forum only a day or two ago.
6:49 pm on Sept 23, 2011 (gmt 0)

Senior Member from US 

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month

joined:Sept 21, 2002
posts:745
votes: 9


I look at the third party site's page to see if it appears to be on topic and well visited.

If so I 301 redirect the wrong location to the correct URL and then smile.
9:59 pm on Sept 23, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month

joined:May 29, 2003
posts:707
votes: 10


Yup. I have hundreds of these. I just ignore them.
10:07 pm on Sept 23, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member g1smd is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month

joined:July 3, 2002
posts:18903
votes: 0


You're potentially missing out on at least some traffic and link benefit.

It's worth checking where those links come from (using WMT or Analytics).
4:20 am on Sept 24, 2011 (gmt 0)

Preferred Member

5+ Year Member

joined:June 10, 2011
posts: 521
votes: 0


Most of these sites are scrappers that use a broken link to trick Google.
I wouldn't waste a time looking if they have traffic - don't need their traffic. I would save the 301 (and its machine process) for important cases.
12:22 pm on Sept 24, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member sgt_kickaxe is a WebmasterWorld Top Contributor of All Time 5+ Year Member

joined:Apr 14, 2010
posts:3169
votes: 0


I've had googlebot test for wordpress on non-wordpress sites several times, their favorite page(on my sites) being xmlrpc.php (wordpress remote posting page). That page outputs just on line of text that differs if you allow/disallow remote publishing, it's a nice lightweight page that also tells google if your site is more vulnerable to hacking.

If you have a wordpress site you can visit example.com/xmlrpc.php to see what I mean. if you don't you can do a Google search for "XML-RPC server accepts POST requests only." WITH quotes and see over a million listings for a page that is rarely ever linked to. Google's doing some detective work beyond just crawling your pages, looking for signatures...

"aha, this is a wordpress site, apply known wordpress filters." - GBot
8:46 pm on Sept 24, 2011 (gmt 0)

Senior Member from US 

WebmasterWorld Senior Member lucy24 is a WebmasterWorld Top Contributor of All Time 5+ Year Member Top Contributors Of The Month

joined:Apr 9, 2011
posts:13538
votes: 403


You mean, google is engaging in behavior that would get an unfamiliar robot banned on the spot? Say it ain't so :(
8:54 pm on Sept 24, 2011 (gmt 0)

Senior Member

WebmasterWorld Senior Member g1smd is a WebmasterWorld Top Contributor of All Time 10+ Year Member Top Contributors Of The Month

joined:July 3, 2002
posts:18903
votes: 0


It would make sense that Google profiles sites to and records what software they run on. Certain platforms have severe structure, URL, and crawling issues.

By profiling the sites and then having crawling and indexing fixes in place for each platform, Googlebot wouldn't repeatly beat it's head on the wall over and over again.
 

Join The Conversation

Moderators and Top Contributors

Hot Threads This Week

Featured Threads

Free SEO Tools

Hire Expert Members