Hi,
I hope this experience could be be useful for other users of this forum. I've got a website with 2,000,000+ URLs, and I migrated from HTTP to HTTPS in early 2019.
I suffered a 50% drop of my visits in the November-2019 update, and browsing the 'Coverage>Excluded' section of the Google Search Console of the HTTP property I've found:
100,000 URLs with "Duplicate without user-selected canonical"
80,000 URLs with "Duplicate, Google chose different canonical than user"
These 'duplicated URLs' do not appear in the Search Console of the HTTPS property.
It just so happens that the last date of "last crawled" of the two groups of 'duplicated URLs' is "Nov 7, 2019", just the day before of the penalization. It seems that Googlebot decided not to crawl any more 'duplicated URLs' of my website.
As a clue (not sure if this is linked or not), I've just found that I made a mistake in the HTTP-to-HTTPS migration of early 2019. Instead of setting a 301 redirection from "http: //whatever" to "https: //whatever", I set a 302. I've just fixed it.
Now I have some questions for the forum:
Do you think that the 302 mistake is linked to the existence of 'duplicated URLs' and the penalization? Or it may be linked to the existence of a real 'thin content' issue?
How could I tell Google to crawl again such 'duplicated URLs'?
Thank you!