What's my best strategy for Duplicate Content if only www pages are indexed?
-
The MOZ crawl report for my site shows duplicate content with both www and non-www pages on the site. (Only the www are indexed by Google, however.) Do I still need to use a 301 redirect - even if the non-www are not indexed? Is rel=canonical less preferable, as usual?
Facts:
- the site is built using asp.net
- the homepage has multiple versions which use 'meta refresh' tags to point to 'default.asp'.
- most links already point to www
Current Strategy:
- set the preferred domain to 'www' in Google's Webmaster Tools.
- set the Wordpress blog (which sits in a /blog subdirectory) with rel="canonical" to point to the www version.
- Ask programmer to add 301 redirects from the non-www pages to the www pages.
- Ask programmer to use 301 redirects as opposed to meta refresh tags & point all homepage versions to www.site.org.
Does this strategy make the most sense? (Especially considering the non-indexed but existent non-www pages.)
Thanks!!
-
Very informative - thank you! It seems when I think I have a relatively firm grip on SEO, I stumble upon something new - like the dangerous potential for an infinite loop in a 301 redirect in IIS. (That was Chinese to me two days ago;))
Your response solved my concerns - hopefully it will help somebody else when they face the same problem.
-
Well the reason Google has picked the www version as it's preferred version automatically is most likely because of all the links you mentioned that were already pointing to that iteration of your domain. Google can figure this out on their own. That said, it still sees both of the sites (non www and www) as duplicates of each other. Best practice is to 301 one to another.
I've waged this war with a programmer before so I know how it goes. The one I dealt with didn't think there was any reason and told me all websites work that way. So I asked him to go to http://google.com and tell me how it resolves. Repeat that step with every major brand you can think of until he/she gets the point and that might help you.
They should be able to 301 this one time, no matter whether they're using an Apache or IIS server. This should be a quick fix. If they're unsure how to do it, have them Google "IIS 301 redirects" if it's a Windows server or "htaccess 301" if it's a Linux/Apache server.
-
Thanks for easing my mind Jesse! One thing still confuses me - the fact that the non-www pages are not indexed. They are not disallowed in robots.txt, there are no rel=canonical tags (except for pages in the blog subdirectory), they are not meta-refreshed, and obviously not 301 redirected. Could it be the doing of a sitemap (though I cant find one)? Or did Google simply decide all the www pages were more relevant? Am I missing something here? I don't want to ask the programmer to add a ton of 301 redirects (which I did) only to get a 'DUH!' response;)
FYI - site is asp.net - not sure if that matters, except when they redirect homepage to avoid creating infinite loop. (right?)
Thanks again!
-
This is an incredibly easy topic to address because you've already laid out exactly what needs to happen.
In other words, yes! That strategy is exactly the way you should go.
Good job and good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do I fix 5xx errors on pdf's?
Hi all, I'm having a huge (and very frustrating) issue-- I have 437 pages with 5xx errors and have no idea how to go about fixing them. All of them are links to pdf's, and when I click on the link on the Site Crawl page, it opens just fine. I've been trying to find the answer to how to fix this issue for weeks with no luck at all. I tried linking the pdf's to their parent page, tried renaming the links to be more concise and crawler friendly, and I've searched all types of articles involving pdf's and 5xx errors, all with no luck. Does anyone have any idea why these pages are getting this error and how I can fix them? Huge thanks in advance! Daniela S.
Moz Pro | | laurelleafep0 -
How to deal with auto generated pages on our site that are considered thin content
Hi there, Wondering how to deal w/ about 300+ pages on our site that are autogenerated & considered thin content. Here is an example of those pages: https://app.cobalt.io/ninp0 The pages are auto generated when a new security researcher joins our team & then filled by each researcher with specifics about their personal experience. Additionally, there is a fair amount of dynamic content on these pages that updates with certain activities. These pages are also getting marked as not having a canonical tag on them, however, they are technically different pages just w/ very similar elements. I'm not sure I would want to put a canonical tag on them as some of them have a decent page authority & I think could be contributing to our overall SEO health. Any ideas on how I should deal w/ this group of similar but not identical pages?
Moz Pro | | ChrissyOck0 -
Need Help with www and non-www redirect
Hi everyone, I've been looking around the forum and found some similar topics but none of them have solved my problem. Anyway, I'm new to SEO and found that when I use Open Site Explorer for the www version of my site I see the links directed to my page, but none of my twitter stats show up. When I do the same for my non-www version. I see my twitter stats, but no links. Facebook seems unchanged. Also, the Page Authority is much higher for the www (I assume because of the links) It is my understanding that it doesn't matter which version of the domain is used, I just need to pick one and stick with it (on Google Webmaster Tools, etc.). And that I also need to do a 301 redirect for the version that I'm not using. The issue now is that I have a 301 redirect, and when I type in my url in any version (http, https, www, non-www) it all redirects in the browser without issue. So i know users are getting to the right page, but the Open Site Explorer still sees them as 2 different sites. My concern is that if Open Site Explorer sees it this way then Google may see it this way as well and I could be missing out on potential rankings. I'm currently doing the redirect in the .htaccess file. The redirect looks like this: RewriteEngine On
Moz Pro | | opstart
RewriteCond %{SERVER_PORT} 80
RewriteRule ^(.*)$ https://mydomain.com/$1 [R=301,L] Any Ideas on what I need to do to properly 301 redirect the site so that my preferred version can get credited with the links,twitter, and Page Authority stats? Also, I'm using wordpress and it has the URL set to the non-www version of the site. However, I think I would like to change this to the www version because most of the links are directing to www, and I also think people are more familiar with that format. Anyway, sorry for the long post. Any help would be greatly appreciated. Thanks.0 -
Howcome my target country (Kuwait) don't appear while im trying to create a campaign? www.google.com.kw.
im in kuwait and already getting good results in the local google search engine, but howcome you dont have my local search engine in your country list www.google.com.kw when using your rank tracking tool thank you PGFHRAlGu
Moz Pro | | WeKnowDigital0 -
Batch lookup domain authority on list of URL's?
I found this site the describes how to use excel to batch lookup url's using seomoz api. The only problem is the seomoz api times out and returns 1 if I try dragging the formula down the cells which leaves me copying, waiting 5 seconds and copying again. This is basically as slow as manually looking up each url. Does anyone know a workaround?
Moz Pro | | SirSud1 -
Duplicate page errors
I have 102 duplicate page title errors and 64 duplicate page content errors. They are almost all from the email a friend forms that are on each product of my online store. I looked and the pages are identical except for the product name. Is this a real problem and if so is there a work around or should I see if I can turn off the email a friend option? Thanks for any information you can give me. Cingin Gifts
Moz Pro | | cingingifts0 -
I have another Duplicate page content Question to ask.Why does my blog tags come up as duplicates when my page gets crawled,how do I fix it?
I have a blog linked to my web page.& when rogerbot crawls my website it considers tags for my blog pages duplicate content.is there any way I can fix this? Thanks for your advice.
Moz Pro | | PCTechGuy20120 -
SEOmoz Crawl CSV in Excel: already split by semicolon. Is this Excel's fault or SEOmoz's?
If for example a page title contains a ë the .csv created by the SEOmoz Crawl Test is already split into columns on that point, even though I haven't used Excel's text to columns yet. When I try to do the latter, Excel warns me that I'm overwriting non-empty cells, which of course is something I would rather not do since that would make me lose valuable data. My question is: is this something caused by opening the .csv in Excel, or earlier in the process when this .csv is created?
Moz Pro | | DeptAgency2