Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

Is Google re-crawling the web and re-evaluating all links right now?

         

aakk9999

11:17 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



And I mean - is it happening on a big scale right now?

And is the news on "nofollow" attribute handling as well as the new GWT interface announcement and rollout in fact here to distract SEO world on what is really happening?

I might be wrong but what I am seeing is the following:

1) May/June changes in SERPs - just read a long discussion on a separate thread here

2) The new GWT, which not just introduced a new interface many talk about, but in fact changed how inbound (and perhaps outbound) links are counted (e.g. no more 301 inbound links) and perhaps other things?

3) Big discussion about nofollow on internal links + PR sculpting, a discussion which to me seems to be out of proportion to acknowledged influence of page's PR in relationship on how the page ranks in SERPs

On top of this, I am seeing in new GWT that Google suddenly "found" a large number of old links of ours, which were present before we did changes in URL structure and is trying to crawl them. Only noticed because they got reported as errors in crawl. I am not talking here about a link or two, but about hundred or so links using old syntax we abandoned and which were never before in GWT as either internal or external links. These links are nowhere on our site nor are being linked from "wide world", but suddenly there is a surge of them being crawled, whereas these have certainly not being crawled since at least Feb.

It almost seems to me that Google pulled these "old" links up from its own data storage from somewhere and is retrying them.

And further, we just got a surge of Google alerts for few articles we syndicated over 18 months ago and for which we already had Google alerts coming 12 months ago or earlier. So why again the same alert?

Without wanting this to sound like a conspiracy theory, but I am wondering whether the announced change in nofollow handling and the new GWT interface are there to distract from other perhaps major change that is happening right under our noses?

Kind of like world politicians, who, when they want some changes to go unobserved, they make them at the times when there is some other major news that distracts general public.

tedster

11:20 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Just a quick comment - the new WMT interface changes how links are REPORTED, but not how they are scored.

aakk9999

11:28 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Yes this is what it should be, but I am myself not really sure. I am wondering why links pointing to a domain that is page to page 301-ed to a different domain now do not show any more on the "redirected to" domain. Is it just reporting or is there more to it?

Plus old links being re-crawled...

And there is a question already asked on GWT thread about keywords missing from the new GWT - with the comment that they should have the data already, so why it is not there?

Or maybe it is so simple as for Google to release a new version of (GWT) product in such a "buggy" state and hence things are not showing?

Just makes me wonder - too many coincidences maybe.

tedster

11:32 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I mentioned this recently in the thread about the new Webmaster Tools interface [webmasterworld.com]:

JohnMu: ...we built a new backend for gathering links data as part of our new UI and we decided not to follow 301s. The reason we did this was a lot of users were trying to go to the source page that we mention has the URL and were unable to find it since we were showing the final destination URL rather than what was specified on page.

We understand this change has solved issues for some users, but created new problems for others and we are going to look at ways to solve this problem.

Webmaster Help Forums [google.com]

Webmaster Tools gathers links data from the main data set at Google. This comment in its full context makes it pretty clear that it's only the reporting that has changed. Not following 301's at all would be a HUGE change - and a stupid one, too. Google is many things, but they're not stupid.

aakk9999

11:45 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Hi, yes, I saw that in the other thread. So, JohnMu says that the problem is that the GWT users (webmasters) would click on the URL that is reported to have external link to their page, and could not find their link there because the link was to a page on a different domain that redirects to their page.

Wouldn't one normally assume that these users own both domains and that they would know if one of them redirects 301 to the other? This would be much more common scenario then a third party redirecting 301 to your domain hence leaving you oblivious to where the link to your page is.

On the other hand, if 301 is not followed when getting links data, does this mean that anchor text of redirected link will also not be reported to belong to a target domain? And what about passing PR?

Just wondering...

tedster

11:59 pm on Jun 17, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Be careful not to confuse the WebmasterTools team and their function, with the main Search team and the actual ranking algo.

Wouldn't one normally assume that these users own both domains...

There is all manner of funny business on the web. Consider one example - incorrectly configured click counting scripts on external links. There is also a kind of disruptive webmastering that happily experiments with 301 and other redirects, just to see what kind of mayhem they might be able to create.

But getting back to your overall point, there is certainly a lot of churn and (from what I see) spidering right now. Google has had such periods in the past - and apparently we've got one now. It's the beginning of summer and in years past that was also a time when we saw similar things.

aakk9999

12:02 am on Jun 18, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



I see your point, perhaps I was not on these boards long enough to compare with the past

CainIV

7:04 am on Jun 18, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Could definitely be a possibility that the way Google handles and scores nofollow internally on websites will and in fact is already influencing rankings. Perhaps this was part of the link value and re-calulation that seemed to cause some anomalies some webmasters reported recently.

I believe if the changes to this attribute were being folded in, they would be done so in manageable increments as it is quite likely flipping those type of changes over wholesale would cause some very strong changes.

tedster

6:51 pm on Jun 18, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



The nofollow changes were live last year, so I doubt that's what we're seeing now. One idea I have is that the -50 penalties from early in this month (see SERP Changes -June 2009 [webmasterworld.com]) are not upt to par and more data is needed to straighten this new algo tweak out.

dstiles

10:14 pm on Jun 18, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



I had a dozen links in WMT for one site a couple of months ago that have been redirected to different domains for over 8 years. No idea why.

Today I had a hit from googlebot on default.html - I've NEVER had a page named that on ANY of my/customer sites.

I'm quite prepared to believe there are a lot of other instances I haven't seen - I only caught this one because it triggered an attack trap on the page name "default" (a common exploit/scrape attempt).

And... Just checked this month's good-bot log for google.*404 and got 309 of 'em. My sites don't lose pages very often so aakk9999 could be right about a re-scan of old pages.

Mind you, the same scan for msnbot got me about 540 hits and yahoo 430. Not that either is a surprise.

micklearn

4:57 am on Jun 19, 2009 (gmt 0)

10+ Year Member



Apologies to all, since this drifts off-topic a bit...Tedster, I'm curious, how do you express such confidence with your statement "The nofollow changes were live last year...". How could you or anyone know for sure, unless one was employed by Google itself and working on the search/spam/crawl teams there? I just fired an SEO person I had hired a while ago, after reading about and then mentioning the present/past nofollow matter to them. So, the nofollow bit he sold me on, was really a sham all along? -->(Search engines need to be told what pages aren't of importance on a site.)

tedster

5:59 am on Jun 19, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Let's make that more a exact statement: Matt Cutts just blogged that the changes to nofollow were live last year [webmasterworld.com].

Also note, whether a nofollow attribute tells a search engine that the page linked to is "unimportant" is not the change that went live. The change is in how a nofollow link affects the PR vote of other links on the page. I still use nofollow for internal links to utility pages that I don't want to see in sitelinks, for example. It's easier for me than hawking over the sitelinks and deleting them in WMT.

We've got two other [webmasterworld.com] threads [webmasterworld.com] where we are discussing the nofollow - so let's not have a third.

So now we return this thread to it's regularly scheduled topic: Is Google doing a major crawl right now to re-evaluate links?

micklearn

3:45 am on Jun 20, 2009 (gmt 0)

10+ Year Member



Well, what if the threads could be related? So...yes, I think they might have been doing a major crawl, re-evaluating links and I believe the purpose of the crawl could easily be based upon the nofollow attribute. Who can honestly say when the change happened? It may or may not have gone live last year...just my opinion. I've never believed everything I read, even when it's coming from a high ranking employee of Google.

Thanks for your reply, tedster. I read about and had also read the full blog post of Mr. Cutts. I fully understood the changes that Matt discussed, prior to my post here and I appreciate that you shared your knowledge about the matter.

[edited by: tedster at 6:19 am (utc) on June 21, 2009]
[edit reason] member request [/edit]

tangor

4:05 am on Jun 20, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



I see odd things from time to time in results. Google, these days, seems to be like that man with two watches... never knows what time it is (look up the joke if you don't know what that means). When you have SO MUCH data at hand you have to keep testing it, over and over, and that's what I see happening. Merely an observation as I'm not that anal re: SERP results for several personal sites... then again, on the sites I manage for clients which I don't talk about I can be a real bear. :)

Then again, there's the packrat syndrome: I got it all, I want it all, and I want it all the time, if only I have the time to check to see if I have what I've forgotten I have.

At what point does information glut break the machine? Not sure, but I suspect it's not that far away given the MFA and scrappers, etc. out there. And the gamers, too.

tedster

4:21 am on Jun 20, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I'm not convinced that Google would need more crawling data to re-evaluate links. They've already got tons of crawl data and the link map of the web that generates, right? So all they need to do is change the algo that evaluates that data in a new way. In fact they do that all the time on experimental data sets before the new version of an algo tweak goes into live production.

One thing they "could be" doing is getting more current data to see how webmasters are responding to the announcements about PR sculpting, nofollow, javascript links and so on - "Are webmasters going crazy with changes?"

But even that seems not so likely to me. Most webmasters do not follow Google's or Matt Cutts announcements - many major corporate web teams (and certainly smaller site owners) don't know who Matt Cutts is. Google goes through cycles of heavier spidering and then lighter spidering throughout the year, depending on plans and needs that we'll never know about. This is more likely to be just that kind of business-as-usual, I think.

g1smd

2:07 am on Jun 21, 2009 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Just to comment on the not following redirects thing: I believe the situation there is when you look at a page with many "Tiny" or "Short" URLs on it, and don't know which one is linking to your content. Such "Tiny" URLs are, of course, redirects (hopefully of the 301 type, not the 302 type).

Love2Blog

2:07 am on Jul 7, 2009 (gmt 0)

10+ Year Member



I too feel like they are re-evaluating links and maybe devaluing certain link profiles, which may explain my drop in SERPS, as I have lots of links from article marketing. And I too am getting a ton of Google Alerts like a 300% increase from usual with old links coming in.

Also of interest is suddenly the Google blog widget that shows incoming links to blogs in the Wordpress admin dashboard, which has not worked for some time in not showing any links for many users, suddenly lit up on all my blogs (6) and all the links I have been getting are suddenly coming in like crazy.