Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

How best to indicate a server-side error with minimal SERP impact?

         

jscroft

3:32 pm on May 26, 2010 (gmt 0)

10+ Year Member



We have a family of affiliate sites that are driven by data feeds from a third-party aggregator. We hit the aggregator from the server side during the page construction process, and occasionally the aggregator fails to deliver, meaning we have no data to present on the page.

Of relevance is that we represent many vendors, and each vendor may have thousands of distinct static URLs on a site. So if our aggregator drops a vendor for a few minutes or hours, every one of those pages will produce the same error. Also, new pages will arise and old pages will drop off the map during the normal course of business as a vendor shuffles its feeds.

The question is, how can we indicate these errors with the least possible impact on SERPs?

It's probably also worth noting that the traffic dynamic we typically see is a long period of very little traffic, followed by a VERY fast increase over a few weeks, followed by a crash and then almost nothing again for months. We suspect that the way we handle these errors may be responsible for our traffic crashes, but I'd be interested in other suggestions.

Our initial solution was to deliver an error page (HTTP Response Code 200) that effectively said, "Wups! No data, please try again!" The effect of this was a raft of duplicate content and, eventually, a traffic crash. We served this error page for way too long.

Recently--after the most recent crash--we tried something more sophisticated. Now a server-side data error serves the same old 200 error page to a regular browser, but a 503 (try again in 1 minute) to search spiders. I know it's working because I can see the "Network Unreachable" errors pile up in GWT.

Unfortunately, I am also seeing some de-listing. A month ago the site referenced above had no 503s and 30K pages in the Google index; now we have 266 503 errors and 6,500 pages in the index. So I question whether serving the 503s is a good idea.

Another option would be to serve a 404 to spiders, but since most of these errors are temporary I really don't want Google to de-list the page on the first error.

Eventually we plan to cache data to serve in the event of an error, but we aren't there yet. Any thoughts?

[edited by: tedster at 5:56 pm (utc) on May 26, 2010]

tedster

6:04 pm on May 26, 2010 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



As you've concluded, caching data server-side is probably the best way to avoid SERP disruption.

When there is an error, the HTTP status code is critical, as you've also discovered. It needs to show a true error code, and 503 is the most accurate in your situation. However, SERP disruptions are inevitable in this situation, because Google does not want their end user to get a link that goes to an error page.

I'd say your solution is to get the caching implemented as soon as possible - fast track it as best you can!

jscroft

6:58 pm on May 26, 2010 (gmt 0)

10+ Year Member



Thanks for the response, tedster. I was afraid you were going to say that.

Can anybody speak to the trajectory we are seeing... major traffic ramp-ups, followed by a crash and months of nothing, repeat? Is this pattern consistent with this kind of error?