"non-WWW" vs "WWW" in Google SERPS and Lost Back Link Connection
-
A Screaming Frog report indicates that Google is indexing a client's site for both: www and non-www URLs. To me this means that Google is seeing both URLs as different even though the page content is identical.
The client has not set up a preferred URL in GWMTs. Google says to do a 301 redirect from the non-preferred domain to the preferred version but I believe there is a way to do this in HTTP Access and an easier solution than canonical.
https://support.google.com/webmasters/answer/44231?hl=enGWMTs also shows that over the past few months this client has lost more than half of their backlinks. (But there are no penalties and the client swears they haven't done anything to be blacklisted in this regard.
I'm curious as to whether Google figured out that the entire site was in their index under both "www" and "non-www" and therefore discounted half of the links.
Has anyone seen evidence of Google discounting links (both external and internal) due to duplicate content?
Thanks for your feedback.
Rosemary
-
Don't think so as Google will update its index. If it's not too much work you can do it but not sure it's worthful
-
Thank you for your reply.
Do we still need canonicals if we set the preferred domain in GWMT and update htaccess?
-
Yes that the best you should do and then redefine in GWT the preference for www
-
You can edit you htacess file and do something like this
RewriteEngine on
RewriteCond %{HTTP_HOST} !^www.yourdomain.com$ [NC]
RewriteRule (.*) http://www.yourdomain.com/$1 [R=301]
RewriteRule ^index.php$ / [R=301]This will make sure that no url loads without www if it does it will 301 redirect it to www also you want to make sure that your homepage / index page is not duplicated so, have that one 301 redirect to the root as well.
-
I'm going through the same kind of situation these days. The client I'm providing SEO services was ranking on the 4th page in the serps. The website was having 8k duplicate pages because of www and non-www issue. I just added the canonical tag in the preferred domain and waiting for Google to deindex the non preferred domain.
But I've also noticed that just after creating 14 backlinks on authority sites, the site is ranking on first page in 5 different keywords.
I don't see any down in backlinks in WMT, I'm not sure but I think Google doesn't cut down backlinks due to duplicate pages. Wait for any senior to reply to your question.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google Search Console "Text too small to read" Errors
What are the guidelines / best practices for clearing these errors? Google has some pretty vague documentation on how to handle this sort of error. User behavior metrics in GA are pretty much in line with desktop usage and don't show anything concerning Any input is appreciated! Thanks m3F3uOI
Technical SEO | | Digital_Reach2 -
My Website's Home Page is Missing on Google SERP
Hi All, I have a WordPress website which has about 10-12 pages in total. When I search for the brand name on Google Search, the home page URL isn't appearing on the result pages while the rest of the pages are appearing. There're no issues with the canonicalization or meta titles/descriptions as such. What could possibly the reason behind this aberration? Looking forward to your advice! Cheers
Technical SEO | | ugorayan0 -
Duplicate Content Showing up on Moz Crawl | www. vs. no-www.
Hello Moz Community! I am new to SEO, Moz and this is my first question. My questions; I have a client that is getting flagged for Duplicate Content. He is getting flagged for having two domains that have the same content i.e. www.mysite.com & mysite.com. I read into this and set up a 301 redirect through my hosting site. I evaluated which site had a stronger Page Authority and had the weaker site redirect to the stronger site. However, I am still getting hit for Duplicate pages caused by the www.mysite.com & mysite.com being duplicates. How should I go about resolving this? Is this an example of a Canonical tag needed in the head of the HTML? Any direction is appreciated. Thank You. B/R Will H.
Technical SEO | | MarketingChimp100 -
3,638 mystery inbound links from google.com
Hello, can anybody help me. I my webmaster tools, under "Links to your Site" and "Who links the most" It shows Google as the top inbound link source, with 3,669 out of 4,14, all these links are pointing to one sub page of content. Yet when I click on the source it only shows me one Google Plus page. My traffic hell by 60% when I first noticed these links. we have not been building unnatural links and we have not had any Manual link building warning. This is the site: iunlock.org Two questions:- I have deleted this one linked to page what else can I do to reverse this impact Does anybody know what has happened? See attached WWwnb1W.png
Technical SEO | | Jack4ireland0 -
Missing meta descriptions from Google SERPs
Hullo all, I run an e-commerce website and hence have a lot of product category/sub-category pages to handle. Despite giving each of these category pages meta descriptions, in the Google SERPs, a lot of these descriptions don't show up fully. Rather, only half the text that I'd inputed as my meta desc. shows up; the other half has generic stuff from that page given. I've attached a screen shot to give you an example of what comes up in the SERPs. Could you please tell me what exactly is the problem? Is it a coding issue? Or has Google not crawled that page? Need help asap! Thank you in advance! aE9RKXJ
Technical SEO | | suchde0 -
I have a ton of "duplicated content", "duplicated titles" in my website, solutions?
hi and thanks in advance, I have a Jomsocial site with 1000 users it is highly customized and as a result of the customization we did some of the pages have 5 or more different types of URLS pointing to the same page. Google has indexed 16.000 links already and the cowling report show a lot of duplicated content. this links are important for some of the functionality and are dynamically created and will continue growing, my developers offered my to create rules in robots file so a big part of this links don't get indexed but Google webmaster tools post says the following: "Google no longer recommends blocking crawler access to duplicate content on your website, whether with a robots.txt file or other methods. If search engines can't crawl pages with duplicate content, they can't automatically detect that these URLs point to the same content and will therefore effectively have to treat them as separate, unique pages. A better solution is to allow search engines to crawl these URLs, but mark them as duplicates by using the rel="canonical" link element, the URL parameter handling tool, or 301 redirects. In cases where duplicate content leads to us crawling too much of your website, you can also adjust the crawl rate setting in Webmaster Tools." here is an example of the links: | | http://anxietysocialnet.com/profile/edit-profile/salocharly http://anxietysocialnet.com/salocharly/profile http://anxietysocialnet.com/profile/preferences/salocharly http://anxietysocialnet.com/profile/salocharly http://anxietysocialnet.com/profile/privacy/salocharly http://anxietysocialnet.com/profile/edit-details/salocharly http://anxietysocialnet.com/profile/change-profile-picture/salocharly | | so the question is, is this really that bad?? what are my options? it is really a good solution to set rules in robots so big chunks of the site don't get indexed? is there any other way i can resolve this? Thanks again! Salo
Technical SEO | | Salocharly0 -
How hard is it to rank above a site that has 300 back links from dex?
I am new to SEO and have started my keyword research. One keyword I would like to rank for has a competing site with over 400 back links, 300 of which are from dex. How hard would it be to outrank links coming from dex?
Technical SEO | | marker-3115280 -
.htaccess and www - non www
Recently I have taken over a website and I made a pretty colossal mistake. The site was properly constructed via .htaccess to a www domain. Typically I roll without it and I made a bad assumption that the .htaccess was not previously set correctly because there were hundreds of fundamental mistakes. After a couple of days I noticed the mistake but some of our new (non www) have picked up some solid links etc. So now I feel that I am in a nightmare of creating redirects etc. So should I switch back to WWW or not? Does it matter at this point?
Technical SEO | | mikeusry0