Forum Moderators: open

Message Too Old, No Replies

Sandboxed Sites - Back Together?

Do they come out together or one by one?

         

McMohan

10:09 am on Nov 20, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Most of the new sites that I work with are still in the sandbox. Was just curios to know, if all the sanboxed sites come out of the sandbox during one fine major updation or one by one, over the rolling updates?

That is to say, should one be checking to see if the sites are out of the sandbox regularly or only when they know there is a major Google update? :)

Thanks

Mc

BeeDeeDubbleU

4:08 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



The bottom line is that Google has dictated that all sites in the future will not rank well unless they behave like a normal site would behave and unless they are well considered by the Internet population.

Interesting theory.

But ...

... wait a minute!

How come "normal" sites that have been introduced since February are also missing?

Pimpernel

4:49 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



"Normal" by the criteria that you apply to your old sites, which is the mistake everyone is making. The fact is that the old sites would be gone as well if google had the ability to downgrade them. But it doesn't have the data so it can't apply the algorithm. The new "normal" sites are the ones that are performing perfectly well in the SERPS and were created / registered since February. If you look hard enough you will find them.

Bottom line, for the large majority of new sites it is going to be a long hard haul to get from the bottom of the SERPS to the top, not like the old days when you could get ranked in a week.

wanna_learn

4:52 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



Pimpernel,
Why dont you make it easier by throwing example of so called only 5 such Normal sites performing well on compititive KWS?

airpal

5:05 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



This is so ridiculous, there's no such thing as a "normal" or a "spammy" site! Does anybody else have more feedback regarding specific results they have seen, so we can try and solve the REAL sandbox matter once and for all?

BeeDeeDubbleU

7:00 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



You want to know how Google ranks pages?

Have a look at [google.co.uk...]

Here's an excerpt ...

Google uses PageRank™ to examine the entire link structure of the web and determine which pages are most important. It then conducts hypertext-matching analysis to determine which pages are relevant to the specific search being conducted. By combining overall importance and query-specific relevance, Google is able to put the most relevant and reliable results first.

Oh yeah? So is Google saying that virtually no sites that have been introduced during the last nine months provide relevant or reliable information?

There is no mention of new sites or new pages being treated differently from those that are established. Isn't Google's mission to deliver SERPs that are all based on their algo and all sites being treated equally?

If this situation is deliberate then, if not actually lying, they are being very economical with the truth.

steveb

7:57 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



"Just very difficult and seriously anti-spam."

Talk about backward. The spammiest tactics are what beat the sandbox (<<<<<giving up in the jargon wars).

Jane_Doe

7:59 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



>>> It doesn't take more than nine months to build a new index. Does it?

I think with projects like that the hard part and really time consuming part is expanding fields in all of the places an index may be used - all the reports, temporary files, files that get transferred to external companies, screen layouts where the field is used etc. The more business partners you have who have to change all of their systems, screen layouts and reports to accept a new size field, the more complex the project becomes.

Namaste

8:03 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Pimpernel, how do you explain what I said earlier, that even new pages that don't fall within the keyword categories of an existing (well listed) sites are sandboxed.

airpal

9:05 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



Pimpernel, how do you explain what I said earlier, that even new pages that don't fall within the keyword categories of an existing (well listed) sites are sandboxed.

You will be extremely hard-pressed to find somebody who will agree with you that new pages on old sites are sandboxed at all. I have launched numerous pages on an old site that were ranking very well within days. Those pages had hundreds of completely different "keyword categories".

UK_Web_Guy

9:18 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



airpal

It's impossible to generalise full stop - what you've expereinced is one thing, what another experiences is totally different.

I've got sandboxing of new pages on old sites.

It's not just specific keywords that Google seem to be using to determine what is sandboxed and what isn't - no one has figured out what they are using yet - hence why this type of thread appears every few weeks and hence why they become so long.

DerekH

9:22 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



I've tried to stay out of this "sandbox" debate, partly because other parts of the English-speaking globe have no idea what a sandbox is, and partly because it's been fun to watch people slagging each other off merely for the apparent sin of having different points of view.

Am I correct in assuming that the sandbox theory is either "in" or "out", according to one's preference? And that pages (or sites - I'm not sure which) - are either in this box-thingy or not?

Well, when I step back, I find that the whole concept of Google's PageRank, its SERPS, its toolbar and everything else is based on "better than" or "not as good as" - real quantifiable measures (whether you agree with the measure or not).

I personally find it rather sad that we're wasting time discussing the sandbox like it was some sort of portcullis - you're in or you're out of the Google Castle.

Do others really believe that the massive matrix calculations that define PR are then going to be adjusted by a coin-toss? In or out? Heads you win, tails you're on page 100?

I've a Masters degree in Mathematics, but I'm finding that I'm turning into a philosopher in this debate, trying to understand what I see, rather than making up "in or out" theories that are quite, quite childish.
We would, I think, be better served by trying to make sense of the contrasting and contrary things we're seeing here, instead of heaping coals onto some vast fire.
The fact that we ARE seeing constrasting and contrary things is what we should be grasping - not that we don't think we ARE seeing them.

----
OK - I've had my rant and I feel better now.
Sorry about that - us Brits tend to keep our emotions bottled up far longer than is good for us...

Why not skip over this and read the next post instead....
DerekH

mark1615

9:45 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



Whether or not one agrees with Pimpernel's theory it appears internally consistent:

Page are what is ranked not sites
Old sites cannot be subject to the same part of the algo that he thinks new sites are because the data wasn't kept prior to the algo change.
Links are tracked and aged.

One question on this though, a new page on an old site would still seem to be subject to the algo - no? Yet many people have experience that suggests this is not true.

Likewise the other commonly observed attribute that sites/pages in the so-called SB can still rank well for obscure 3+ word combinations. What does DerekH think about this with his background in higher mathematics?

And one other thought: The so-called anti-spam tactics employed by G are to fight a "problem" they largely invented. The basic premise of the G algo, we are lead to believe, is that links are votes. Well, then webmasters go out and get links. And anchor text in links is important, but the fact is that truly natural anchor text very often is totally unrelated to the keyoword and is thus devalued (we think) by G. So now in response to webmasters actions to get links - G (again, we think) takes action to combat aggressive linking - which causes this problem because of course, new pages and new sites have new links. This has resulted in G becoming unarguably stale.

DerekH

10:24 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



mark1615 wrote
Likewise the other commonly observed attribute that sites/pages in the so-called SB can still rank well for obscure 3+ word combinations. What does DerekH think about this with his background in higher mathematics?

Well, I'm not sure that this is anything more than the way the reverse or inverse index that looks things up is updated and made current.
After all, in addition to the algorithms that decide results, there is the data that is fed to those algorithms. With some pages on one of my sites indexed yesterday, and some not visited since last February, the spread of currency of the data is massive. Who can say what effect the age of the last visit of one of your competitor's pages has on the weight that page is ascribed.

For a long long time I've seen my sites rank really well for one keyword and not for another, and yet for the pair to beat sites that beat me on both searches.
I don't regard that as anything more than "something" in my site doing well for an obscure combination of keywords, any more than I regard the fact that a site doing well for an obscure search means anything more important than the fact that other sites don't.

My god what a sentence that was!
What I meant is that it's easy to do well in an obscure search. That's what obscure means.

And what I didn't say was that I don't actually have a view one way or the other about the sandbox. Some of my pages have done well, some have been wiped out; but the last thing I think is that it's something quite so black and white.

Anyway - you shouldn't as me to justify my rant <grin> - it was just something I needed to get off my chest...
DerekH

bak70

10:27 pm on Nov 24, 2004 (gmt 0)

10+ Year Member



This might be controversial.

The sand box does not exist.

Google updates roughly every three months.
Im talking deep update.

If you have enough seo in time for the update you move
if not you stay.

I have 100 s of sites between my partner and I .
we have seen this happen many times.

If your site has not moved in 6 months than you havent done enough seo or you are doing it wrong.

If you have time to complain on this board chances are you havent done enough.

Powdork

11:21 pm on Nov 24, 2004 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



THERE IS NO SANDBOX!

Google introduced new algorithms in February and these algorithms are tough, tough ant-spam algorithms. They are based on lots of factors like:

How quickly the links were amassed
Quality of links
How quickly pages were increased
Etc etc


You start off saying there is no sandbox, and then describe one of the popular theories as to its existence. If this terminology works better for you, thats fine. Wherever it says 'sandbox', just replace it with 'tough anti spam algorithms' and you should be ok. AFAIK Google has always kept the date a link first appeared. This is not something new.

When you search for a restaurant by its name and city and google does not return the restaurant's website even though it is indexed, just because the site's links aren't aged to perfection, it doesn't matter what its called. It's a reason to leave Google.
People aren't leaving Google in droves because every other aspect of the search engine is far superior to the competition. But each day the 'tough anti spam algorithms' continue, this aspect becomes more noticeable.

Google updates roughly every three months.
Im talking deep update.
bak70 if you think we've had a deep update in the last nine months, you may be in for a shocker soon. At least I hope so.
This 472 message thread spans 32 pages: 472