Forum Moderators: Robert Charlton & goodroi

Message Too Old, No Replies

Where to redirect these extra dynamic links to remove them from google

         

fsmobilez

10:23 am on Dec 25, 2008 (gmt 0)

10+ Year Member



There are some urls of my site which were crawled by google and i dont want them in google any more

e.g

if the url is

1stlink:www.example.com/post/post.php?post_id=12345&thread_id=&jtype=emailed

and i add permanent 301 redirect in .httaccess and url changes to

2ndlink:www.example.com/post/post.php?post_id=12345&thread_id=

means this string is removed from end of url

>>=&jtype=emailed<<

it means google will remove first link

i want to know will google add the 2nd link in search engine ?

which i dont want....

tedster

1:32 pm on Dec 25, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



Yes, Google might well index that second url, unless you give instructions not to with a meta noindex robots tag.

fsmobilez

1:38 pm on Dec 25, 2008 (gmt 0)

10+ Year Member



AS the site is dynamic script based site

So what i do .

i have two suggestions in mind and need ur help

if i redirect all urls to my site main page

Or

redirect it to 404 page

If u can help me what to do?

tedster

1:54 pm on Dec 25, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



A 301 redirect to your home page could be a major problem over time. Your home page is not a new location for the content in the original url, but that's what the redirect would be saying. Over time, you could pile up a whole bunch of duplicate urls.

A 301 redirect to a page that actual returns a 404 http status (and not just the 404 message) would be better.

g1smd

3:51 pm on Dec 25, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



You should modify the scripting that runs the site so that pages you do not want to be indexed include a meta robots noindex tag.

The logic to do this is very simple, and is based on detecting the requested URL for that page view and a simple IF statement.

Robots.txt can also be used to stop the crawling of certain types of URL. Google supports URL wildcards, as do some other search engines.

fsmobilez

4:54 pm on Dec 25, 2008 (gmt 0)

10+ Year Member



Well i used robots.txt and it actually worked but there was a confusion

Google removed the content of the site but url of the site was still there

and i was in doubt that might google consider it as duplicate or it effects my rankings bcoz many unwanted urls crawled by google

and even after robots used , urls were still in google search and some one on another forum told me that google will never removed these urls which are indexed by google even u block them using robots.

g1smd

6:00 pm on Dec 25, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



Google shows URLs that are blocked by robots.txt rules as URL-only entries in the SERPs.

Using the meta robots noindex tag ensures that nothing shows up. Edit your scripting for that.

fsmobilez

6:28 am on Dec 26, 2008 (gmt 0)

10+ Year Member



Well i can only add the rule in robots.txt file and cant add noindex tag in the script

So should i go for that?

and will google remove my urls which will be in robots.txt file

Quadrille

5:20 pm on Dec 26, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



No.

That will not guarantee to solve your problem.

The noindex tag is what you need; is there no way you can revise your script?

Else you could ask Google to remove the pages :)

fsmobilez

8:44 am on Dec 27, 2008 (gmt 0)

10+ Year Member



How to contact google regarding this issue

Key_Master

8:57 am on Dec 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member



An internal redirect (using mod_rewrite) to a blank page that spits out a X-Robots-Tag noindex header would work.

fsmobilez

9:08 am on Dec 27, 2008 (gmt 0)

10+ Year Member



U mean i redirect all that pages to one page which have No index tag.

RIght now i have reidrected all that pages to a 404 page.

g1smd

11:55 am on Dec 27, 2008 (gmt 0)

WebmasterWorld Senior Member 10+ Year Member Top Contributors Of The Month



When you display a 404 error there should be no redirect involved.

The error page text and the 404 HTTP Status Code should be returned for the originally requested URL.

The browser URL bar should not change to some other URL. If it does, then it is likely that your site returns "302 Found" for the originally requested URL. That is not the right way to do this.

[spelling]

[edited by: g1smd at 12:21 pm (utc) on Dec. 27, 2008]

fsmobilez

12:07 pm on Dec 27, 2008 (gmt 0)

10+ Year Member



So tell me any easy solution what to do

and also if u explain this what tdtester said

>>A 301 redirect to a page that actual returns a 404 http status (and not just the 404 message) would be better. <<