Forum Moderators: Robert Charlton & goodroi
Google is coming out with a new tag called “unavailable_after” which will allow people to tell Google when a particular page will no longer be available for crawling. For instance, if you have a special offer on your site that expires on a particular date, you might want to use the unavailable_after tag to let Google know when to stop indexing it. Or perhaps you write articles that are free for a particular amount of time, but then get moved to a paid-subscription area of your site.[highrankings.com...]
>> Google is starting to crawl the supplemental index more often... <<
Hmm. Googleguy already said that - about a year to 18 months ago.
Yes, and "unavailable after" may just indicate to Google not to bother re-visiting the page to re-index, but does it really, really mean that it'll disappear from the index? Or will it simply languish in the Supplemental index for months to a year?
Added:
Oh yeah, the candle burning brightly on that cake is how to get your site into Google:
1) Get links.
2) Do up and submit a site map.
Adding to that making sure that all pages within a site can be reached by a crawlable link, am I mistaken, or aren't there still people who do all those things saying that they still can't get all their pages indexed?
[edited by: Marcia at 8:27 pm (utc) on July 13, 2007]
Google is coming out with a new tag called "unavailable_after".
Well, I guess they've reached the point where they can now do that and many will follow.
It sure would be nice if ALL the crawlers would just follow one protocol, possibly something already in existence with a slight modification? All these proprietary tags are going to cause issues. I can see it now, This Site Best Viewed in Google.
410 Gone
a way to tell Google in advance of when a page goes away, or is no longer relevant, exactly when it is going to be gone.
It's also a good way to tell Google that whoever is responsible for the site attends SEO events or frequents SEO forums, because the average Charlie won't know and probably won't care.
But Dan Crow really is Google's director of crawl systems (see wikipedia), he really did speak to the marketing group in Rhode Island (there are even photos) and this information is now reported by at least one other first hand attendee (Eric Lander).
I'm more skeptical of some of the spam-control spin directed to the SEO community from various Google vehicles than I am of this purely technical "sneak preview".
Well, I guess they've reached the point where they can now do that and many will follow.
Let's hope so. Someone has to be a leader.
The "unavailable_after" tag sounds like an excellent idea, because it's good for both the search engine and site owners. If other search engines want to take advantage of it, great; if not, no harm done.
If other search engines want to take advantage of it, great; if not, no harm done.
I'll disagree with that. There is a lot of harm done. Google ends up indexing all that unavailable_after stuff after its been regurgitated and stripped of that metadata. Same with the nofollow attribute. They are at most, band-aid fixes to a much deeper issue.
But Dan Crow really is Google's director of crawl systems
some skepticism is certainly healthy -- it prevents rumors and misinformation, and that's a good thing.
"how to get your site into Google".
And until there's a formal announcement, we won't know the exact syntax for this meta-tag anyway and cannot implement it even if want to.
"getting their site into Google."
It gives me mental indigestion to see a Google rep's good intentions being "served up" this way.
For example, Crow discussed the ability to add in meta tags for non-traditional web media such as Flash, videos, etc. Still, he certainly added caution in that these things were months (or more) off -- but were being developed.
[edited by: Robert_Charlton at 3:19 am (utc) on July 16, 2007]
[edit reason] edited out response to moved post [/edit]
The Expires entity-header field gives the date/time after which the response is considered stale. A stale cache entry may not normally be returned by a cache (either a proxy cache or a user agent cache) unless it is first validated with the origin server (or with an intermediate cache that has a fresh copy of the entity). [..]
The presence of an Expires field does not imply that the original resource will change or cease to exist at, before, or after that time.
[w3.org...]
As an example, the unavailable_after or expire tag has been supported by NetInsert since 2000, along with other kinds of useful tags such as revisit, news, and hover.
I mean, what's the actual incentive for a webmaster to add an extra tag to a page in order to tell Google to decrease traffic to their site?
Is achieving lower amounts of free traffic a high-demand area for SEO experts these days?
This "proposed" tag doesn't seem to pass the common sense and stupidity tests.
Otherwise, the content will become available only from the google cache, surrounded by ads of course, and the excuse is going to be that you said it would not be available from your site after date yyyy-mm-dd. But as a public service, we have cached it in perpetuity.
After all, "not available after" is not the same as "not cached after".
no-cache + 410 is the perfect combination.
It exists today, and is recognised universally.
So why foist off an oval wheel that only fits fords?
"google where serp quality is job 1"
Can someone tell me why a webmaster would want to use this "unavailable_after" tag instead of just taking the page down and replacing it with a 301 to a different page they want related traffic for?
Maybe because they genuinely don't want a page to be available after a certain date, and they don't want to wait for the old URL to be crawled and processed?
Example: A manufacturer's rebate page with a July 30 expiration date, or an e-commerce site with a sale that ends on August 10.
Just because everyone may not need or use an "unavailable_after" meta statement doesn't mean nobody would find it useful.