Solved Should I consolidate my "www" and "non-www" pages?
-
My page rank for www and non-www is the same. In one keyword instance, my www version performs SO much better.
Wanting to consolidate to one or the other. My question is as to whether all these issues would ultimately resolve to my chosen consolidated domain (i.e. www or non-www) regardless of which one I choose. OR, would it be smart to choose the one where I am already ranking high for this significant keyword phrase?
Thank you in advance for your help.
-
It may be that one version (www or non-www) has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience in practice this type of migration tends to be smooth.
-
Require the www Options +FollowSymLinks RewriteEngine On RewriteBase / RewriteCond %{HTTP_HOST} !^www\.askapache\.com$ [NC] RewriteRule ^(.*)$ https://www.askapache.com/$1 [R=301,L]
-
Yes. I would recommend picking the version (either www or non-www) that has the historical data showing it performs better than the other version. Check the list of indexed pages for each of the versions to compare. Ideally both the www and non-www version of the website will be indexed in Google so it will help you to decide which version makes the most sense to consolidate to.
Once you identify the preferred version, set 301 redirects from the non-preferred URLs to the preferred version of each URL (the one that has more traffic, links, authority, etc.) of the site. This should be done site-wide so that all URLs are either www or non-www, it shouldn’t be a mix of both. In my experience, I’ve found that between 90-99% of the Site’s SEO Authority is preserved when setting a permanent 301 redirect.
-
@meditationbunny Sorry for the slow reply - but yes, I'd expect Page Authority to increase slightly, if the "other" version had any value to it.
For Page Optimization, yes. For example, for my own site I see:
http://tcapper.co.uk redirects to https://www.tcapper.co.uk/. This on-page analysis is for https://www.tcapper.co.uk/.
-
It may be that one version (www or non-www) has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience in practice this type of migration tends to be smooth.
-
@tom-capper
Thank you. Yes, I should be more clear. I am calling it page rank, when I am actually referring to Moz's domain authority and Moz's keyword ranking. Still, I believe you answered my question. Under page optimization, I can see what appear to be duplicate listings of my pages along with different SERP ranking. It was confusing until I realized that one was the www and the other was non-www. I have since added code to my .htaccess file that will send everything to www. Can I expect the page optimization section to now only show www versions of the pages? Also, can I expect page authority to increase because it is no longer a mish-mash and is all headed to the same domain and same pages (i.e. www version)? -
It may be that one version ("www" or "non-www") has more historical links. You say your PageRank for both is the same, but how are you checking that? Google's public PageRank has not been updated in a decade or so.
Either way, I'd generally say that if you pick one version and stick to it (redirect the other, e.g. so every non-www. URL points to its www. equivalent), you should maintain all rankings. There is a theoretical advantage to picking the version with more links, but in my experience, in practice, this type of migration tends to be smooth.
-
This post is deleted! -
This post is deleted! -
This post is deleted!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Pages being flagged in Search Console as having a "no-index" tag, do not have a meta robots tag??
Hi, I am running a technical audit on a site which is causing me a few issues. The site is small and awkwardly built using lots of JS, animations and dynamic URL extensions (bit of a nightmare). I can see that it has only 5 pages being indexed in Google despite having over 25 pages submitted to Google via the sitemap in Search Console. The beta Search Console is telling me that there are 23 Urls marked with a 'noindex' tag, however when i go to view the page source and check the code of these pages, there are no meta robots tags at all - I have also checked the robots.txt file. Also, both Screaming Frog and Deep Crawl tools are failing to pick up these urls so i am a bit of a loss about how to find out whats going on. Inevitably i believe the creative agency who built the site had no idea about general website best practice, and that the dynamic url extensions may have something to do with the no-indexing. Any advice on this would be really appreciated. Are there any other ways of no-indexing pages which the dev / creative team might have implemented by accident? - What am i missing here? Thanks,
Technical SEO | | NickG-1230 -
WMT "Index Status" vs Google search site:mydomain.com
Hi - I'm working for a client with a manual penalty. In their WMT account they have 2 pages indexed.If I search for "site:myclientsdomain.com" I get 175 results which is about right. I'm not sure what to make of the 2 indexed pages - any thoughts would be very appreciated. google-1.png google-2.png
Technical SEO | | JohnBolyard0 -
Schema markup for products is missing "price": Is this bad?
Hey guys, So a current client of mine has an e-commerce shop with a few hundred products. They purposely choose to keep the prices off of their website, which is causing errors in Google Webmaster Tools. Basically the error shows: Error: Structured Data > Product (markup: schema.org) Error type: missing price 208 items with error Is this a huge deal? Or are we allowed to have non-numerical prices for schema ie. "call for quote"
Technical SEO | | tbinga1 -
Effect of 302 redirects from empty parent page to sub page
A client's website has links to their service pages which then redirect (302 through a php "Location:" header) to that service's first sub-page. For example, our-services/service-x redirects to our-services/service-x/about-service-x I can only think this has been done because there is no actual content for the parent page and to maintain some kind of structure for navigation and URLs. Really there's no reason why the 'about-service-x' page can't be removed and its content transferred to the main 'service-x' page. Then the redirects can be removed also - it's not how a 302 should be used for a start. I'm just wondering what kind of effect this current redirection has on SEO, as I know 302s don't pass any link juice? Thanks for your help.
Technical SEO | | driftingbass0 -
Pageing page and seo meta tag questions
Hi if i am using paging in my website there is lots of product in my website now in paging total paging is 1000 pages now what title tag i need to add for every paging page or is there any good way we can tell search engine all page or same ?
Technical SEO | | constructionhelpline0 -
Best action to take for "error" URLs?
My site has many error URLs that Google webmaster has identified as pages without titles. These are URLs such as: www.site.com/page???1234 For these URLs should I: 1. Add them as duplicate canonicals to the correct page (that is being displayed on the error URLs) 2. Add 301 redirect to the correct URL 3. Block the pages in robots.txt Thanks!
Technical SEO | | theLotter0 -
What to do when you want the category page and landing page to be the same thing?
I'm working on structuring some of my content better and I have a dilemma. I'm using wordpress and I have a main category called "Therapy." Under therapy I want to have a few sub categories such as "physical therapy" "speech therapy" "occupational therapy" to separate the content. The url would end up being mysite/speech-therapy. However, those are also phrases I want to create a landing page for. So I'd like to have a page like mysite.com/speech-therapy that I could optimize and help people looking for those terms find some of the most helpful content on our site for those certain words. I know I can't have 2 urls that are the same, but I'm hoping someone can give me some feedback on the best way to about this. Thanks.
Technical SEO | | NoahsDad0 -
Penalization for Duplicate URLs with %29 or "/"
Hi there - Some of our dynamically generated product URLs somehow are showing up in SEOmoz as two different URLs even though they are the same page- one with a %28 and one with a 🙂 e.g., http://www.company.com/ProductX-(-etc/ http://www.company.com/ProductX-(-etc/ Also, some of the URLs are duplicated with a "/" at the end of them. Does Google penalize us for these duplicate URLs? Should we add canonical tags to all of them? Finally, our development team is claiming that they are not generating these pages, and that they are being generated from facebook/pinterest/etc. which doesn't make a whole lot of sense to me. Is that right? Thanks!
Technical SEO | | sfecommerce0