Why is there duplicates of my domain
-
When viewing crawl diagnostics in SEOmoz I can see both "www.website.com" and a truncated version "website.com" is this normal and why is it showing (I do not have duplicates of my site on the server)?
E.g.:
-
Hi All,
I have just altered my Google Webmaster tools in the Config (as per Geoff's advice above) and that appears to be fine for now. However, I have also updated my ".htaccess" file dong a 301 redirect, details below:
Options +FollowSymLinks
RewriteEngine onRewriteCond %{HTTP_HOST} !^www. [NC]
RewriteRule ^ http://www.%{HTTP_HOST}%{REQUEST_URI} [L,R=301]I also have recently changed my pages from .html to .php do I redirect by using this code as I want all old links to now refer to the .php file:
RewriteRule (.*).html$ /$1.php [R=301,L]
Lastly, Thomas on the canonical side of things, is it necessary for me to implement this in the Head?
How can I see if the work completed is correct or check that it works?
Paul.
-
Ok, I have implemented the Google Webmaster redirect. and now I have tried the edit to the .htaccess file trying both options below with no success.
Option 1:
AddHandler php-stable .php
RewriteEngine On
RewriteCond %{HTTP_HOST} !^www.RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]Option 2:
AddHandler php-stable .php
Options +FollowSymLinks
RewriteEngine on
RewriteRule (.*) http://www.klinehimalaya.com/$1 [R=301,L]Neither seem to be working when viewed here:
http://anythinghere.klinehimalaya.com/Please help.
Thanks
-
Hey Mat,
Another question. If I have changed a page address from:
http://www.klinehimalaya.com/packages.html to
http://www.klinehimalaya.com/packages.phpwill I also need to redirect the pages I have changed from .html to .php or is this not needed because I have not changed the file name prior to the prefix?
Thanks.
Paul.
-
You can set the preferred domain just like Geoff mentioned. But basically since you have a www version and a non www version it can be seen as duplicate content. Although the site is only one website and changes will appear on each the URL structure changes for depending on if you use the www or non www and therefore it looks like duplicate content.
-
I have checked out your website we you may not have implemented the 301 redirect correctly. In addition you should use Google Webmaster tools along with
rel="canonical"
http://www.seomoz.org/blog/how-to-advanced-relcanonical-http-headers
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=139394
Add a
rel="canonical"
link to the section of the non-canonical version of each HTML page.To specify a canonical link to the page http://www.example.com/product.php?item=swedish-fish, create a element as follows:
Copy this link into the section of all non-canonical versions of the page, such as http://www.example.com/product.php?item=swedish-fish&sort=price.
If you publish content on both http://www.example.com/product.php?item=swedish-fish and https://www.example.com/product.php?item=swedish-fish, you can specify the canonical version of the page. Create the element:
Add this link to the section of https://www.example.com/product.php?item=swedish-fish.
-
Hi,
If I was you I would go into Google Webmaster tools and verify both versions of the site, then go to Configuration >> Setting >> Preferred domain and tell Google your preference.
You will need to do this with both verified versions of the domain. You can then delete the listing for the one you do not want
You will only need one Google Verification code.
That will sort the issue with Google
Geoff
-
You have fixed the redirect loop, but the initial problem still exists. With the urgent issue now fixed (I can see the site now!) let's look at the problem:
You don't actually have a live problem at the moment, so any fix here will not improve your results. However you have a potential problem - one that could occur and cause you problems later. It's not hard to fix, so it is good to get it done.
Your server (like many) is set up to serve the same files to both klinehimalaya.com and www.klinehimalaya.com, which are technically different addresses. If google followed a link to klinehimalaya.com and a link to www.klinehimalaya.com if would find the content at two seperate places and could index both and consider them to be duplicate. Yes... all those PHDs and Google is still that stupid.
What you need to do (and I think you were trying to do this) is to pick one version of that address as the one that you are going to use. You can then redirect traffic from the other version to the chosen version.
In fact anythinghere.klinehimalaya.com is also serving up the same site, so you should fix the lot in one go.
If you are running apache (most sites are) then you can do this with an .htaccess redirect . There is a good example of how to do this at the top of this page: http://dense13.com/blog/2008/02/27/redirecting-non-www-to-www-with-htaccess/
In short it checks to see whether the address being loaded starts with www. . If it doesn't then it loads that version via a 301 redirect.
The issue Ross has pointed out is separate and not fixed by this. You should do both.
I hope that helps.
-
You are linking to your home page via your logo in the header.
However you are linking to this address http://www.klinehimalaya.com/index.php try linking to this address http://www.klinehimalaya.com/
This will remove your duplicate content issue.
-
Think I just fixed the redirect, but I still have an issue in SEOmoz where it is still pulling up both references to:
http://www.klinehimalaya.com/index.php
klinehimalaya.com/index.phpApparently this will effect my SEO efforts, hot to fix!!!
Thanks all!
-
I tried to look at this for you, but you have some issues on your site that are stopping me viewing it completely.
Both klinehimalaya.com and www.klinehimalaya.com create an infinite loop that prevents the page loading. klinehimalaya.com is correctly 301ing to www.klinehimalaya.com
However www.klinehimalaya.com is redirecting to itself.
You need to fix that URGENTLY. Start by looking at your htaccess file.
-
You need to do a 301 redirect
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Old domain still being crawled despite 301s to new domain
Hi there, We switched from the domain X.com to Y.com in late 2013 and for the most part, the transition was successful. We were able to 301 most of our content over without too much trouble. But when when I do a site:X.com in Google, I still see about 6240 URLs of X listed. But if you click on a link, you get 301d to Y. Maybe Google has not re-crawled those X pages to know of the 301 to Y, right? The home page of X.com is shown in the site:X.com results. But if I look at the cached version, the cached description will say :This is Google's cache of Y.com. It is a snapshot of the page as it appeared on July 31, 2014." So, Google has freshly crawled the page. It does know of the 301 to Y and is showing that page's content. But the X.com home page still shows up on site:X.com. How is the domain for X showing rather than Y when even Google's cache is showing the page content and URL for Y? There are some other similar examples. For instance, you would see a deep URL for X, but just looking at the <title>in the SERP, you can see it has crawled the Y equivalent. Clicking on the link gives you a 301 to the Y equivalent. The cached version of the deep URL to X also shows the content of Y.</p> <p>Any suggestions on how to fix this or if it's a problem. I'm concerned that some SEO equity is still being sequestered in the old domain.</p> <p>Thanks,</p> <p>Stephen</p></title>
Technical SEO | | fernandoRiveraZ1 -
Duplicate Content Problems
Hi I am new to the seomoz community I have been browsing for a while now. I put my new website into the seomoz dashboard and out of 250 crawls I have 120 errors! So the main problem is duplicate content. We are a website that finds free content sources for popular songs/artists. While seo is not our main focus for driving traffic I wanted to spend a little time to make sure our site is up to standards. With that said you can see when two songs by an artist are loaded. http://viromusic.com/song/125642 & http://viromusic.com/song/5433265 seomoz is saying that it is duplicate content even though they are two completely different songs. I am not exactly sure what to do about this situation. We will be adding more content to our site such as a blog, artist biographies and commenting maybe this will help? Although if someone was playing multiple bob marley songs the biography that is loaded will also be the same for both songs. Also when a playlist is loaded http://viromusic.com/playlist/sldvjg on the larger playlists im getting an error for to many links on the page. (some of the playlists have over 100 songs) any suggestions? Thanks in advance and any tips or suggestions for my new site would be greatly appreciated!
Technical SEO | | mikecrib10 -
Duplicates on the page
Hello SEOMOZ, I've one big question about one project. We have a page http://eb5info.com/eb5-attorneys and a lot of other similar pages. And we got a big list of errors, warnings saying that we have duplicate pages. But in real not all of them are same, they have small differences. For example - you select "State" in the left sidebar and you see a list on the right. List on the right panel is changing depending on the what you selecting on the left. But on report pages marked as duplicates. Maybe you can give some advices how to improve quality of the pages and make SEO better? Thanks Igor
Technical SEO | | usadvisors0 -
301s vs. rel=canonical for duplicate content across domains
Howdy mozzers, I just took on a telecommunications client who has spent the last few years acquiring smaller communications companies. When they took over these companies, they simply duplicated their site at all the old domains, resulting in a bunch of sites across the web with the exact same content. Obviously I'd like them all 301'd to their main site, but I'm getting push back. Am I OK to simply plug in rel=canonical tags across the duplicate sites? All the content is literally exactly the same. Thanks as always
Technical SEO | | jamesm5i0 -
302 multiple domains...
Hello, I have a few domain names with orthographic variations that I'd like to redirect to my main site. The problem is my registrar (OVH) does only 302 redirects, so what are my options ? Can I keep a dozen 302's ? Do I have to change all their DNS (it's a load on my server...) ? Thanks for any ideas Johann.
Technical SEO | | JohannCR0 -
Duplicate Page Title for Wordpress
Hello, We are using WP for our blog and keep getting Dup Page Title errors for our 12 author archives pages. The title of each page is the same, but I am wondering if this is WP issue with canonicalization working properly. The most recent four pages have a linking root domain and carry some Page Authority, but the older pages do not. Is this what Rand was talking about in his Google+ whiteboard Friday talk about blog post relevancy not lasting as long as articles? Here's what it looks like. Side question, is there a reason why the SEO Moz website doesn't have a Google+ button anywhere easy to find? Thank you, Michael
Technical SEO | | MKaloud1 -
Duplicate Content
Many of the pages on my site are similar in structure/content but not exactly the same. What amount of content should be unique for Google to not consider it duplicate? If it is something like 50% unique would it be preferable to choose one page as the canonical instead of keeping them both as separate pages?
Technical SEO | | theLotter0 -
External Links from own domain
Hi all, I have a very weird question about external links to our site from our own domain. According to GWMT we have 603,404,378 links from our own domain to our domain (see screen 1) We noticed when we drilled down that this is from disabled sub-domains like m.jump.co.za. In the past we used to redirect all traffic from sub-domains to our primary www domain. But it seems that for some time in the past that google had access to crawl some of our sub-domains, but in december 2010 we fixed this so that all sub-domain traffic redirects (301) to our primary domain. Example http://m.jump.co.za/search/ipod/ redirected to http://www.jump.co.za/search/ipod/ The weird part is that the number of external links kept on growing and is now sitting on a massive number. On 8 April 2011 we took a different approach and we created a landing page for m.jump.co.za and all other requests generated 404 errors. We added all the directories to the robots.txt and we also manually removed all the directories from GWMT. Now 3 weeks later, and the number of external links just keeps on growing: Here is some stats: 11-Apr-11 - 543 747 534 12-Apr-11 - 554 066 716 13-Apr-11 - 554 066 716 14-Apr-11 - 554 066 716 15-Apr-11 - 521 528 014 16-Apr-11 - 515 098 895 17-Apr-11 - 515 098 895 18-Apr-11 - 515 098 895 19-Apr-11 - 520 404 181 20-Apr-11 - 520 404 181 21-Apr-11 - 520 404 181 26-Apr-11 - 520 404 181 27-Apr-11 - 520 404 181 28-Apr-11 - 603 404 378 I am now thinking of cleaning the robots.txt and re-including all the excluded directories from GWMT and to see if google will be able to get rid of all these links. What do you think is the best solution to get rid of all these invalid pages. moz1.PNG moz2.PNG moz3.PNG
Technical SEO | | JacoRoux0