Forum Moderators: Robert Charlton & goodroi
[edited by: Robert_Charlton at 12:24 pm (utc) on Sep 26, 2020]
[edit reason] Removed violations of Forum Charter and TOS. Will explain in post below. [/edit]
The specific error [the webmaster] is seeing from Google Search Console is "Duplicate, submitted URL not selected as canonical." He said "The problem is that the two pages are not duplicates and google selected canonical does not match the user selected canonical."
Google's John Mueller came in on Christmas eve and responded:
"Usually this happens when we run across a number of URL patterns on a site that all lead to substantially the same content. If this all happened during a short time, it might be that there was something misconfigured that caused this, and in that case, it'll settle back down over time as our algorithms confirm that these URLs are actually separate. That said, I agree that this looks really confusing, so I also forwarded these to the team to check out, in case there's something on our side which we can do to speed things up for you :). "
"Usually this happens when we run across a number of URL patterns on a site that all lead to substantially the same content."
We are currently working to resolve two separate indexing issues that have impacted some URLs. One is with mobile-indexing. The other is with canonicalization, how we detect and handle duplicate content. In either case, pages might not be indexed....
We are currently working to resolve two separate indexing issues that have impacted some URLs. One is with mobile-indexing. The other is with canonicalization, how we detect and handle duplicate content.
...None of the posts I've seen about the issue gets to the heart of the problem, which is that the CMS's used probably introduce a lot of dupe content, essentially the same articles or article snippets sorted in different ways....
does suggest that you are probably on the edge of dupe content issues
I'll definitely take a look and see what can be fixed. Currently, my subpages are deindexed to improve the site's overall quality, but that will not stop the Googlebots from crawling the pagesI don't think that crawl budget is an issue, if that's what you're thinking when you mention Googlebots.