The idea that one is creating duplicate content arises from the view that a Link=URL=Document=Page.
When is a query a query and not "a page reference"?
Only when a query is instantiated as plain text in a search box?
Isn't the problem Google's difficulty gleaning meaning from links themselves? Discerning when a link is a trigger designed to return a certain "view of data" versus "(entirely) new data"?
What about Google advancing its analysis of links themselves? Why stop at "no follow"? Why not "archive by author dataview"? Argh. Not my job.
As an aside, perhaps we should all emulate Google, forgoing website design by designing "websites" around a datastore, a search box and a bit of text? Perhaps then we could then better spend our time and resources on stocking the shelves in the datastore - adding to the data, information, content - instead of constantly moving the cans, boxes and "product" around on the shelves in an effort to please the bot.
Ack. Ignore me. Buckworks no doubt grows rich by knowing how to make the bot happy and respecting that, whilst I fiddle about.
Yes, I know, put away the fiddle. Amen.
[edited by: Webwork at 2:05 pm (utc) on Oct. 24, 2008]