Help With Preferred Domain Settings, 301 and Duplicate Content
-
I've seen some good threads developed on this topic in the Q&A archives, but feel this topic deserves a fresh perspective as many of the discussion were almost 4 years old.
My webmaster tools preferred domain setting is currently non www. I didn't set the preferred domain this way, it was like this when I first started using WM tools.
However, I have built the majority of my links with the www, which I've always viewed as part of the web address.
When I put my site into an SEO Moz campaign it recognized the www version as a subdomain which I thought was strange, but now I realize it's due to the www vs. non www preferred domain distinction.
A look at site:mysite.com shows that Google is indexing both the www and non www version of the site. My site appears healthy in terms of traffic, but my sense is that a few technical SEO items are holding me back from a breakthrough.
QUESTION to the SEOmoz community:
What the hell should I do? Change the preferred domain settings? 301 redirect from non www domain to the www domain?
Google suggests this: "Once you've set your preferred domain, you may want to use a 301 redirect to redirect traffic from your non-preferred domain, so that other search engines and visitors know which version you prefer."
Any insight would be greatly appreciated.
-
The worst thing you can do is nothing.
Above is 5 examples of URLs which COULD all lead to the same page. There are numerous other possibilities as well. If you don't let Google know which version of the page is correct, then you will suffer the consequences of duplicate content.
What happens is Google doesn't know which page is correct. They will pick one of the non-www versions because that is what your Google WMT is set up to do. Meanwhile other versions of the pages are being used.
You are sending your link juice to a page, but it is a complete waste as it is not being considered by Google for SERP. You MUST resolve this issue if you care about SEO at all.
-
Thanks Ryan. So, if most of the links (including all internal links) are built with the www format then it is wise to change preferred domain settings to www and redirect the non www to the www domain?
Am I likely to damage rankings/traffic by doing this? What happens if I just leave it as is?
-
You are welcome to do so. Go to Google WMT, change your current option to the www, then adjust your .htaccess file as Steven suggested.
Also, canonicalize your pages to help ensure this issue can't happen again. Your .htaccess changes will work as long as the file is there, but things happen so it's better to be covered.
-
Guys,
Thanks for the input. I just want to do what is best for traffic and the site. I don't want to do anything that is going to tank my rankings and visitors.
I don't get alot of type in traffic.
www is the main way the links have been built, why not just redirect those to the non www version?
-
As Ryan said, make a decision. The easiest way to make sure either of your decisions sticks is to use an htaccess file and rewrite to your preferred.
If using the www version:
RewriteEngine On
RewriteCond %{HTTP_HOST} ^[0-9]+(.[0-9]+){3} [OR]
RewriteCond %{HTTP_HOST} ^mydomain.com [NC]
RewriteRule ^(.*)$ http://www.mydomain.com/$1 [L,R=301]if using the non www version:
RewriteEngine On
RewriteCond %{HTTP_HOST} ^www.mydomain.com [NC]
RewriteRule ^(.*)$ http://mydomain.com/$1 [L,R=301]A few other questions to keep in mind:
Do you get a lot of type-in traffic?
Do they tend to type the www?
In the SERP it is easier to read the domain name with out the www if looking for a specific domain name. Do you have a brand built where people just say your domain name?
-
You need to make a decision. Do you want your site address to be seen with or without the www?
Try to assess which version of your URL would require the least number of re-directs. You mentioned the links you built mostly include the www. Take a look at all of your links. You may have a higher number of organic links without the www. Evaluate all the links, then make a decision.
Once you make a decision, stick with it. Canonicalize all your pages with the correct version of the URL. Search your site for all internal links and standardize them.
While you are on this project standardize whether you use a "/" on the end of your url as well. www.mysite.com is not the same as www.mysite.com/. I make this suggestion because if you will go through the painful process of standardizing your site for the www issue, you should resolve all issues at once.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Setting up 301 Redirects after acquisition?
Hello! The company that I work for has recently acquired two other companies. I was wondering what the best strategy would be as it relates to redirects / authority. Please help! Thanks
Intermediate & Advanced SEO | | Colin.Accela0 -
Best method for blocking a subdomain with duplicated content
Hello Moz Community Hoping somebody can assist. We have a subdomain, used by our CMS, which is being indexed by Google.
Intermediate & Advanced SEO | | KateWaite
http://www.naturalworldsafaris.com/
https://admin.naturalworldsafaris.com/ The page is the same so we can't add a no-index or no-follow.
I have both set up as separate properties in webmaster tools I understand the best method would be to update the robots.txt with a user disallow for the subdomain - but the robots text is only accessible on the main domain. http://www.naturalworldsafaris.com/robots.txt Will this work if we add the subdomain exclusion to this file? It means it won't be accessible on https://admin.naturalworldsafaris.com/robots.txt (where we can't create a file). Therefore won't be seen within that specific webmaster tools property. I've also asked the developer to add a password protection to the subdomain but this does not look possible. What approach would you recommend?0 -
Help understanding 301 domain redirect
Can anyone help me understand a specific process of a 301 redirecting a domain. Here is what I would like to know.... When you 301 redirect a site, most if not all the links follow to your new site. But how does this process happen? 1.When Google sees the new domain does it simply apply the backlink profile of the old site to the new one? 2. Does it have to re-crawl all the links one by one and apply them to the new domain? 3. or something else?
Intermediate & Advanced SEO | | gazzerman10 -
Will using 301 redirects to reduce duplicate content on a massive scale within a domain hurt the site?
We have a site that is suffering a duplicate content problem. To help resolve this we intend to reduce the amount of landing pages within the site. There are a HUGE amount of pages. We have identified the potential to reduce the pages by half at first by combing the top level directories, as we believe they are semantically similar enough that they no longer warrant being seperated.
Intermediate & Advanced SEO | | Silkstream
For instance: Mobile Phones & Mobile Tablets (Its not mobile devices). We want to remove this directory path and 301 these pages to the others, then rewrite the content to include both phones and tablets on the same landing page. Question: Would a massive amount of 301's (over 100,000) cause any harm to the general health of the website? Would it affect the authority? We are also considering just severing them from the site, leaving them indexed but not crawlable from the site, to try and maintain a smooth transition. We dont want traffic to tank. Has anyone performed anything similar? Id be interested to hear all opinions. Thanks!0 -
Joomla duplicate content
My website report says http://www.enigmacrea.com/diseno-grafico-portafolio-publicidad and http://www.enigmacrea.com/diseno-grafico-portafolio-publicidad?limitstart=0 Has the same content so I have duplicate pages the only problem is the ?limitstart=0 How can I fix this? Thanks in advance
Intermediate & Advanced SEO | | kuavicrea0 -
Copying my Facebook content to website considered duplicate content?
I write career advice on Facebook on a daily basis. On my homepage users can see the most recent 4-5 feeds (using FB social media plugin). I am thinking to create a page on my website where visitors can see all my previous FB feeds. Would this be considered duplicate content if I copy paste the info, but if I use a Facebook social media plugin then it is not considered duplicate content? I am working on increasing content on my website and feel incorporating FB feeds would make sense. thank you
Intermediate & Advanced SEO | | knielsen0 -
Parameters when deciding how to set up cross domain analytics?
Hi, I am planning on setting up cross domain analytics for my website. I have found lots of sources about the code to use, but haven't been able to find something about what parameters to consider. Would appreciate my fellow mozzers helping me to put together a list of all the variations of elements I may want to track to ensure I am setting it up correctly. Thanks!
Intermediate & Advanced SEO | | theLotter0 -
Two Brands One Site (Duplicate Content Issues)
Say your client has a national product, that's known by different brand names in different parts of the country. Unilever owns a mayonnaise sold East of the Rockies as "Hellmanns" and West of the Rockies as "Best Foods". It's marketed the same way, same slogan, graphics, etc... only the logo/brand is different. The websites are near identical with different logos, especially the interior pages. The Hellmanns version of the site has earned slightly more domain authority. Here is an example recipe page for some "WALDORF SALAD WRAPS by Bobby Flay Recipe" http://www.bestfoods.com/recipe_detail.aspx?RecipeID=12497&version=1 http://www.hellmanns.us/recipe_detail.aspx?RecipeID=12497&version=1 Both recipie pages are identical except for one logo. Neither pages ranks very well, neither has earned any backlinks, etc... Oddly the bestfood version does rank better (even though everything is the same, same backlinks, and hellmanns.us having more authority). If you were advising the client, what would you do. You would ideally like the Hellmann version to rank well for East Coast searches, and the Best Foods version for West Coast searches. So do you: Keep both versions with duplicate content, and focus on earning location relevant links. I.E. Earn Yelp reviews from east coast users for Hellmanns and West Coast users for Best foods? Cross Domain Canonical to give more of the link juice to only one brand so that only one of the pages ranks well for non-branded keywords? (but both sites would still rank for their branded keyworkds). No Index one of the brands so that only one version gets in the index and ranks at all. The other brand wouldn't even rank for it's branded keywords. Assume it's not practical to create unique content for each brand (the obvious answer). Note: I don't work for Unilver, but I have a client in a similar position. I lean towards #2, but the social media firm on the account wants to do #1. (obviously some functionally based bias in both our opinions, but we both just want to do what will work best for client). Any thoughts?
Intermediate & Advanced SEO | | crvw0