Forum Moderators: Robert Charlton & goodroi
Google Will Seek Out HTTPS Pages By Default
....we'd like to announce that we're adjusting our indexing system to look for more HTTPS pages. Specifically, we’ll start crawling HTTPS equivalents of HTTP pages, even when the former are not linked to from any page. When two URLs from the same domain appear to have the same content but are served over different protocol schemes, we’ll typically choose to index the HTTPS URL if:
It doesn’t contain insecure dependencies. It isn’t blocked from crawling by robots.txt. It doesn’t redirect users to or through an insecure HTTP page. It doesn’t have a rel="canonical" link to the HTTP page. It doesn’t contain a noindex robots meta tag. The sitemaps lists the HTTPS URL, or doesn’t list the HTTP version of the URL The server has a valid TLS certificate.
Googlebot Will Seek Out HTTPS Pages By Default [googlewebmastercentral.blogspot.com]
Imagine if someone [has] a few hundred links to a proper URI on you site but with some curse words in query string. That rule says URI found and moved to a new location, then the linking page gets crawled and you server redirects it.
make HTTPS *the* standard
get rid of the certificates altogether
Now, can anyone explain why google would look for https URLs that are linked from nowhere (their own prose)? I can't figure out who benefits.
I don't see it as a bad thing per se, but not unusually, G are going overboard with it and requiring https for a flat html info page is overkill.
You want to do this to make your site more secure/trustworthy and to future-proof it. The nytimes article that EditorialGuy referenced earlier in this thread is worth reading.
Hm, in the case of one of my sites it says (Firefox) Untrusted Connection, then asks me if I want to continue. Explorer says Certificate Error, and also offers to continue.
RewriteEngine on
RewriteCond %{HTTP_HOST} ^sub.domain.com$ [NC]
RewriteRule .* - [F] https with potentially dodgy certificates
https with potentially dodgy certificates