homepage Welcome to WebmasterWorld Guest from 54.243.17.133
register, free tools, login, search, pro membership, help, library, announcements, recent posts, open posts,
Pubcon Platinum Sponsor 2014
Visit PubCon.com
Home / Forums Index / Marketing and Biz Dev / SEM Research Topics
Forum Library, Charter, Moderators: phranque

SEM Research Topics Forum

    
New Site Over 40k Pages, SEO Problem?
linsys




msg:3356165
 8:30 pm on Jun 1, 2007 (gmt 0)

I have read that if a site comes out with too many pages initially it could be flagged a spam site by the SEs. I'm creating a new site which will have about 40,000 pages of content, the content is all real estate related getting the content from the local MLS in my state.

I was wondering if the SEs will flag the site as spam for offering 40K pages right off the bat.

I have never launched a site this big with this many pages before. The largest sites I have, have about 10k pages indexed, but they took about one year to make..

Having a site with over 40k pages from day one is new to me, want to make sure I go about it properly so I don't get flagged as a spam site.

 

LifeinAsia




msg:3356194
 8:49 pm on Jun 1, 2007 (gmt 0)

Since you're getting the content from the local MLS (along with dozens, if not hundreds, of other local sites), you main problem will probably be be duplicate content, regardless of the number of pages.

BananaFish




msg:3356370
 1:01 am on Jun 2, 2007 (gmt 0)

Listings are not content. And if it's a brand new site, I'm sure you won't have to worry about all the listings being indexed. Maybe just concentrate on the SEO of the actual core content.

ashish21cool




msg:3356509
 5:09 am on Jun 2, 2007 (gmt 0)

I think you shud go ahead with adding a fixed number of pages every week. This shud help u in getting a gradual rise in front of SE's and not like a spam.

But make it sure its no lots of pages in a week and next week not a single page.

Best Regards

linsys




msg:3362632
 3:08 am on Jun 9, 2007 (gmt 0)

How are listings not content? Each listing will be a separate page.. confused as to why you think that listings are not content, the few sites that actually import the entire MLS database, have almost every listing indexed with about 40k pages indexed in Google.

Terabytes




msg:3362652
 4:47 am on Jun 9, 2007 (gmt 0)

I have to agree that the listings will be considered content to that site...

I also have to agree that it will be considered duplicate content once indexed...an exact copy of the original data...

what is to be gained by rehashing the same information that's already indexed by a buh-zillion other sites...perhaps I'm missing a concept here...dunno

Tera

linsys




msg:3369563
 4:37 pm on Jun 16, 2007 (gmt 0)

Well there is a lot to be gained. Mainly customers.

1) For the Search "State Real Estate" the top like 5 sites ALL have the same MLS listed, all duplicate content.

2) Customers can signup to receive email alerts about the latest results for "City #*$!".

Number 2 is the biggest reason. This site is obviously for a real estate company which means people need to be able to search the MLS from the site and sign up for new listings via email. There isn't much around this.

Although there is a risk of duplicate content, it doesn't seem to be effecting the other top sites for "State Real Estate" or "Main City Real Estate".

I really just wanted to know if adding 40k pages over night was going to kill me or if I should try and throttle the results.

Also, since the site will update daily I don't think it will give Google a good chance to penalize the site for duplicate content thats probably why it works so well for the other competitors.

pageoneresults




msg:3369572
 4:51 pm on Jun 16, 2007 (gmt 0)

I really just wanted to know if adding 40k pages over night was going to kill me or if I should try and throttle the results.

Lets look at this from Google and a PageRankô perspective. If the root domain is sitting at PR3/PR4, adding 40,000 pages is going to suck the life right out of that site.

I would definitely throttle back the release of new pages. I'd really focus on click paths and "core content" first. Get the upper level click paths indexed first and then once you have some PR to work with, release more.

The problem I see as others have noted is that you are basically regurgitating the same content that millions of other agents are using. Those pages have little to no value unless of course they have been set up differently than everyone else has done it. ;)

Competing in the Real Estate sector in itself is a major task. I'd have to imagine that it is one of the most heavily filtered set of SERPs.

What will probably happen (if you release 40,000 pages at once) is those new pages will get indexed and many will end up in the Supplemental index shortly thereafter. Then you'll have an unnaturally high ratio of pages in the Supplemental index which in turn may have an effect on the overall performance of those pages not in Supplemental.

linsys




msg:3369790
 12:58 am on Jun 17, 2007 (gmt 0)

See I was actually thinking the same thing, untill I discovered that in my state and probably MOST states Realty websites that display the MLS are using an iFrame or pulling the listings from another site, which means 90% of the sites in my area don't host the entire MLS them selves.

I'm not sure how it works in all states but I know where I am at the local MLS requires a huge questionnaire including hardware specs, firewall specs, how you protect against scrapers etc.. there are probably 10 sites that have the entire MLS hosted on their domain that I could find, NONE of them have pages which are considered supplemental, all pages, and homes in the index are fully indexed with Google atleast.

pageoneresults




msg:3370138
 12:17 pm on Jun 17, 2007 (gmt 0)

I'm not sure how it works in all states but I know where I am at the local MLS requires a huge questionnaire including hardware specs, firewall specs, how you protect against scrapers etc.. there are probably 10 sites that have the entire MLS hosted on their domain that I could find, NONE of them have pages which are considered supplemental, all pages, and homes in the index are fully indexed with Google atleast.

Hmmm, well I guess you have your answer then. Go ahead and dump 40,000 pages into the index. Let us know how things progress for you.

Did you perform any research on those 10 sites? Are they authorities in their space?

What will probably happen (if you release 40,000 pages at once) is those new pages will get indexed and many will end up in the Supplemental index shortly thereafter. Then you'll have an unnaturally high ratio of pages in the Supplemental index which in turn may have an effect on the overall performance of those pages not in Supplemental.

Global Options:
 top home search open messages active posts  
 

Home / Forums Index / Marketing and Biz Dev / SEM Research Topics
rss feed

All trademarks and copyrights held by respective owners. Member comments are owned by the poster.
Home ¦ Free Tools ¦ Terms of Service ¦ Privacy Policy ¦ Report Problem ¦ About ¦ Library ¦ Newsletter
WebmasterWorld is a Developer Shed Community owned by Jim Boykin.
© Webmaster World 1996-2014 all rights reserved