Forum Moderators: phranque
I am interested in creating a bot that will convert search queries into static html pages. The pages do not need to be persistent, but I would like them to be indexable by a search engine should a bot arrive.
For instance, suppose someone searched the term "blue widgets". I'd like to then create a page like "/category/blue-widgets.html". From there I could include information relevant to blue-widgets.
Does anyone have any idea how this could be done or where I can learn some more about it?
Thanks,
Chris
I'm not sure how you'd do this - they don't accept instructions. Unless you're talking about a SE on your own site in which case all you need to do is to write the code the way you want it.
But I think what you might be in search of is the htaccess file (Apache servers). There are several threads here on WebmasterWorld regarding how to make dynamic pages look static etc... Search for htaccess or mod rewrite.
This has been discussed here before and some have used the technique with much success. I don't remember the technology used, although it was discussed in the last thread I read about this. I don't know if it was PHP or some other script. Good luck finding the threads though...I don't even know what you'd search for here at WW.
I might add that other here have considered it a spammy technique. And if I saw a competitor doing it I would send in the spam report faster than certain politicians jump on vulnerable oil reserves in certain countries.
Kevin
I am going to reconsider the honesty thing.
Kevin
So, where does the content come from? Do you use bot of your own to collect content from other sites and then strip, cleanse, reorganize, reformat, and then deliver the new page?
"Looking for small purple widgets? Try our little blue widgets."
And at least ink has an excellent natural language system to detect garbled keywords so you couldn't just steal content with a bot and rearrange.
Alas, I can think of no automated way to provide any semi-unique stolen content. Anyone have any ideas around here?
I guess the best thing would be to steal content and leave as is. Then you could strip out all of the hrefs and have the link to your real content in a table above with lots o' white space. Then just contend with the occasional angry email and remove the page then or rearrange.
BTW, this is just all out of curiosity as I am one of those spam-cop types that think if it unnaturally helps your rankings...don't do it.
Kevin
[edited by: khuntley at 12:51 am (utc) on July 28, 2003]
That is in fact exactly what I'd like to do. Namely have a search on my own site that creates the "static" and SE listable page.
I just noticed someone doing exactly this and I've got to admit, it's really got me intrigued on how it's done.
John, thanks for the advice, I will try that out and see how it goes.
Chris
However, I can guarantee you that if the site is in any category where those at the top of the serps are making any money, when you start approaching them they will delight in going to extreme lengths to remove you from the engine altogether. And such an artificial traffic generating system is the perfect ammo for them.
Sure you might make a little money for a little while, but if the site isn't a throw-away in the long term I wouldn't do it. I'd say you would last 3-4 months with this system in place.
Kevin
Chris
There are scripts and programs available to build the pages first which are spidered by the search engines to get visitors to your site and then your site can be searched for more results
What I've been thinking about doing is simply:
1) Saving search queries into a file.
2) Scheduling an hourly parse of that file to find the most popularly searched terms.
3) Generating x number of top search pages based upon a given template and wiping the log file.
What I'm most worried about is efficiency, if I start getting hundreds of thousands of entries, what's going to happen to my scheduled task and also the size of the log file.
One yould always do a "Recent Searches" list on teh home page listign the last 10 searches.... since in my case searches are in the form of domain.com/search/searchterms Googles eats that up...
Don't see how that would be in any way "illegal".
SN
If you are going to the trouble of creating an index you might as well use a directory script and build a directory, then you can target the pages for your keywords?