Forum Moderators: open

Message Too Old, No Replies

"refreshing" the website to fool new content flags

update all pages to make googlebot think theres new stuff

         

Frank_Rizzo

12:27 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Does googlebot any sort of scouting to see if sites have cahnged contents?

What I'm thinking is that googlebot visits, sees that pages have not changed and then gets the heck out of there. However, if pages have changed, then a deep crawl required flag is set and performed later.

If this is true then is it a good idea to 'refresh' all webpages in a site. This is a simple case of like, changing the name of a .gif file which is on each page, or removing a trailing space after <html> etc. You then upload the complete site and to the crawlers, it looks as if your content has changed because of date stamp and file size.

vitaplease

12:40 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



If you are referring to the Fresh! stamps or date stamps Google gives for certain sites, I am not so sure Google is as (un)sophisticated as that.

In my case, Google is giving that honour to the two highest pagerank pages of my site (index PR6 and menu-type page PR5). These two pages tend not to be the most fresh or most changed or updated pages of my site at all.

For the rest a high pagerank tends to help most for deep crawling.
Even better, get the internal pages some high quality inbound external links.

Net_Wizard

1:12 pm on Jul 1, 2002 (gmt 0)



Deep crawl is done once a month.

Fresh stamp, as from my experience, seems to be tied up with PR6 and above.
-updates are limited only/mostly to the index page, sometimes to the next level directory, it doesn't go further than that, regardless if you have new contents in your inner pages.
-if Google deemed that your site deserves to be visited daily, it will grab your home page regardless of timestamp(ie, news sites - where partial contents changes dynamically using SSI, yet page timestamp remains the same.

ciml

4:11 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I'm not sure I've noticed deeper full crawling on more regularly updated sites. The Fresh listings follow Open Directory and Yahoo! entries, even if the link is quite deep and the page has only a little PR.

olias

4:22 pm on Jul 1, 2002 (gmt 0)

10+ Year Member



I have a dynamic site with a homepage that changes regularly through a server side include, I found that actually uploading my default.asp each day helped me to get a 'fresh' tag every other day.

I've now stopped changing it and I'm hardly ever fresh. So I think that timestamp may be a factor.

But the main point is that you will not force a full crawl of your site by making slight changes.

nancyb

4:43 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



well, my site is a lowly PR2 and has been since it recovered (hah!) from PR0 - back last year it ranged between PR4-PR6. I get 'freshed' continually, usually a day or two after I change something on the home page. Every day, she comes a crawlin' and gets at least a couple pages somewhere on the site. I get a deep crawl after the update (like everyone else) and then another deep crawl later in the month about two days after every one else reports that the deep crawl has started. My listings in DMOZ and Y! haven't changed in at least two years.

And, the really stange thing - the whole time I was at PR0 (Dec-Apr) I was 'freshed' whenever I changed anything on the home page. As far as I can see, by my own site and reading about others' fresh tag, there doesn't seem to be a standard for the 'fresh' although I'm sure there is one.

vitaplease

5:14 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



The Fresh listings follow Open Directory and Yahoo! entries, even if the link is quite deep and the page has only a little PR.

Ciml, thats interesting, are you sure thats the case for every ODP/Yahoo listing?
Have you seen that with sites with say five different pages all listed in DMOZ and all five pages getting the Fresh stamp? (hard to check at this moment because the date stamps often alterate to be presented every other day).

It would be yet another case of Google taking sides with ODP and to a limited extent Yahoo.

WebGuerrilla

5:17 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member




The first place to start if you are trying to get on the fresh list is using Brett's Headers Tool [webmasterworld.com] to see if your server returns a valid Last Modified header.

If you are running dynamic content, then there won't be any LM header. And in therory, the page will still get delivered to Googlebot when it makes an "If Modified request."

However, I've found that the spidering seems to improve a bit if you actually send a date.

ciml

5:38 pm on Jul 1, 2002 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I can't say it happens to all ODP or Yahoo! listed pages, vitaplease, but the deep linked, not very prominent pages that I host and that are listed in ODP/Yahoo! get daily crawls.

This does not mean that they get date stamps, though. One covers a period from 1302AD to 1329AD, so it doesn't need to get updated often;).

<added>I've just noticed; that page does get the date stamp, while others that have last modified headers are crawled daily but don't get the date stamp. (It just happens that the site I mentioned has SSI without the x-bit-hack. Bad Calum...)</added>