We have added multiple new sections in our site, the past few months.
We have not seen these pages indexed in Google.
I did a site: search on Google and was amazed to see that our new pages are listed with http: and https: (both versions)
We use the https: version, once users go into our shopping cart (secure part of our site) to complete thier order. (we have an ecommerce site).
Being that our site is set up using relevant links and not absolute, once a user goes back from the shopping cart to the main site the url will begin with https:// instead of the [....]
We have blocked the bots from indexing the https through robots.txt, but that does not seem to help. (Perhaps the reason to this is: because we have blocked it, after we noticed Google is indexing these pages. I guess once they have entered our secure part of the site we can't get rid of the bots, they keep on coming back.)
We are considering a sub domain for our shopping cart and blocking the sub domain in robots.txt and hopefully this will solve the issue. (please correct me if I am wrong)
Being that Google has many https pages indexed from our site, what would you suggest me to do with the https pages. If we drop it, an error page will come up for all these pages, is it OK with Google to have so many error pages?
Or perhaps, we should consider a different solution to address this issue?