Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is it better to use XXX.com or XXX.com/index.html as canonical page
-
Is it better to use 301 redirects or canonical page? I suspect canonical is easier. The question is, which is the best canonical page, YYY.com or YYY.com/indexhtml? I assume YYY.com, since there will be many other pages such as YYY.com/info.html, YYY.com/services.html, etc.
-
Glad you got it sorted out. If you're 301-redirecting a lot of domains, I'd suggest doing it gradually or maybe holding off on the lowest-quality domains. Google can see a massive set of redirects as a bit of a red flag (too many people have bought up cheap domains and 301-redirected to consolidate the link equity). If the domains are really all closely related or if you're only talking about a handful (<5) then it's probably not a big issue.
-
I think things may be sorted out, but I am not sure. I actually put in 301-redirects from a bunch of domains that I own to this new domain, the content of which will eventually replace my main domain. But, I need to get the domain properly set up and optimized before I move it to my primary domain to replace the ancient web site. At that time, I will also redirect this site to the new, old site.
I used to have Google ad-words tied to some of the domains that I 301-redirected to the new web site that I am building. Those were just a waste of money, however, so I put them on hold. I also had a lot of problems with semel and buttons for web bouncing off those pages that I re-directed. I put in .htaccess commands to stop those spam sites and that seems to work.
-
Google seems to be indexing 30-ish pages, but when I look at the cached home-page, I'm actually seeing the home-page of http://rfprototype.com/. Did you recently change domains or 301-redirect the old site? The cache data is around Christmas (after the original question was posted), so I think we're missing part of the puzzle here.
-
So, I think I may have had things wrong. For one thing, it seems like moz and Google are only indexing 2 pages, while the site index shows something like 80 pages. (I suspect an image is a page, and there are a lot of images. But, there are about 10 or 12 distinct pages at the moment. Also, Google and moz do not seem to show the correct key words in any sense like they should, leading me to think that they were just spidering 2 pages. I don't know why. I added the following to my index.html header:
and
I assume I put them in the correct place. I also believe I don't need canonical pages anywhere else.
Should these changes to my index.html make the proper changes?
-
Yeah, I'd have to concur - all the evidence and case studies I've seen suggest that rel=canonical almost always passes authority (link equity). There are exceptions, but honestly, there are exceptions with 301s, too.
I think the biggest difference, practically, is the impact on human visitors. 301-redirects take people to a new page, whereas canonical tags don't.
-
In terms of rel=canonical that will pass value the same as a 301 redirect - for evidence have a look here:
http://moz.com/learn/seo/canonicalization
"Another option for dealing with duplicate content is to utilize the rel=canonical tag. The rel=canonical tag passes the same amount of link juice (ranking power) as a 301 redirect, and often takes much less development time to implement."
See DR Pete's response in this Moz Q&A:
http://moz.com/community/q/do-canonical-tags-pass-all-of-the-link-juice-onto-the-url-they-point-to
http://googlewebmastercentral.blogspot.co.uk/2009/02/specify-your-canonical.html
https://support.google.com/webmasters/answer/139066?rd=1
http://searchenginewatch.com/sew/how-to/2288690/how-and-when-to-use-301-redirects-vs-canonical
Matts Cutts stated there is not a whole lot of difference between the 301 and the canonical - they will both lose "just a tiny little amount bit, not very much at all" of credit from the referring page.
-
Ok, this is how I look at the situation.
So you have two URLs and the question is either to redirect301 or use canonical? In my opinion 301 is a better solution and this is because it will not only redirect people to the preferred version but the link value as well.
Whereas, with canonicals only search engines will know what is the preferred page but it will not transfer the link value which can help you with organic rankings.
Hope this helps!
-
You would put the canonical link in the index file and I would point that at the xxx.com version rather than the xxx.com/index.html version as people visiting your sites homepage are going to enter the domain and not the specific page so xxx.com rather than xxx.com/index.html.
There are some great articles on Moz explaining all this which I would suggest that you read -
http://moz.com/learn/seo/canonicalization
Dr Pete also did this post answering common questions on rel=canonical.
http://moz.com/blog/rel-confused-answers-to-your-rel-canonical-questions
In terms of 301 redirects and canonicalization both pass the same amount of authority gained by different pages. If you are trying to keep it as clean as possible you need to be careful you don't create an issue redirecting your index file to your domain - here is an old post explaining how moz solved this 301 redirect on an Apache server
http://moz.com/blog/apache-redirect-an-index-file-to-your-domain-without-looping
I personally find that if all your links on your site reference your preferred(canonical) URL for the homepage so in this case xxx.com and you redirect the www version to this or vice versa depending on your preference then you add a canonical in the index.html file pointing at xxx.com in this case unless you prefer to do it the other way round with www.xxx.com for both you will be fine.
Hope this helps
-
I forgot. Of course, there is no xxx.com page, per se. It is actually xxx.com/index.html so if you needed to put the canonical reference on xxx.com, how would you do it?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Keywords are indexed on the home page
Hello everyone, For one of our websites, we have optimized for many keywords. However, it seems that every keyword is indexed on the home page, and thus not ranked properly. This occurs only on one of our many websites. I am wondering if anyone knows the cause of this issue, and how to solve it. Thank you.
Technical SEO | | Ginovdw1 -
Why add .html to WordPress pages?
A site I may take over has a plugin that adds .html to the pages. I searched online but I’ve only found how to add it rather than why to add it. Is it needed? If I remove it, I’ll have to be careful with SEO / indexed pages and redirects. The site is running 3.x.x and 90% of the plugins have not been updated in over 5 years including this one. Before I update to 4.7.x, I am trying to understand the landscape (pros / cons) on why something could be used and if I need to find a suitable replacement for it.
Technical SEO | | acktivate2 -
Canonical homepage link uses trailing slash while default homepage uses no trailing slash, will this be an issue?
Hello, 1st off, let me explain my client in this case uses BigCommerce, and I don't have access to the backend like most other situations. So I have to rely on BG to handle certain issues. I'm curious if there is much of a difference using domain.com/ as the canonical url while BG currently is redirecting our domain to domain.com. I've been using domain.com/ consistently for the last 6 months, and since we switches stores on Friday, this issue has popped up and has me a bit worried that we'll loose somehow via link juice or overall indexing since this could confuse crawlers. Now some say that the domain url is fine using / or not, as per - https://moz.com/community/q/trailing-slash-and-rel-canonical But I also wanted to see what you all felt about this. What says you?
Technical SEO | | Deacyde0 -
How to Stop Google from Indexing Old Pages
We moved from a .php site to a java site on April 10th. It's almost 2 months later and Google continues to crawl old pages that no longer exist (225,430 Not Found Errors to be exact). These pages no longer exist on the site and there are no internal or external links pointing to these pages. Google has crawled the site since the go live, but continues to try and crawl these pages. What are my next steps?
Technical SEO | | rhoadesjohn0 -
Pages removed from Google index?
Hi All, I had around 2,300 pages in the google index until a week ago. The index removed a load and left me with 152 submitted, 152 indexed? I have just re-submitted my sitemap and will wait to see what happens. Any idea why it has done this? I have seen a drop in my rankings since. Thanks
Technical SEO | | TomLondon0 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0 -
OK to block /js/ folder using robots.txt?
I know Matt Cutts suggestions we allow bots to crawl css and javascript folders (http://www.youtube.com/watch?v=PNEipHjsEPU) But what if you have lots and lots of JS and you dont want to waste precious crawl resources? Also, as we update and improve the javascript on our site, we iterate the version number ?v=1.1... 1.2... 1.3... etc. And the legacy versions show up in Google Webmaster Tools as 404s. For example: http://www.discoverafrica.com/js/global_functions.js?v=1.1
Technical SEO | | AndreVanKets
http://www.discoverafrica.com/js/jquery.cookie.js?v=1.1
http://www.discoverafrica.com/js/global.js?v=1.2
http://www.discoverafrica.com/js/jquery.validate.min.js?v=1.1
http://www.discoverafrica.com/js/json2.js?v=1.1 Wouldn't it just be easier to prevent Googlebot from crawling the js folder altogether? Isn't that what robots.txt was made for? Just to be clear - we are NOT doing any sneaky redirects or other dodgy javascript hacks. We're just trying to power our content and UX elegantly with javascript. What do you guys say: Obey Matt? Or run the javascript gauntlet?0 -
/index.php in sitemap? take it out?
Hi Everyone, The following was automatically generated at xml-sitemaps.com Should I get rid of the index.php url from my sitemap? If so, how do I go about redirecting it in my htaccess ? <url><loc>http://www.mydomain.ca/</loc></url>
Technical SEO | | RogersSEO
<url><loc>http://www.mydomain.ca/index.php</loc></url> thank you in advance, Martin0