Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How to Remove Joomla Canonical and Duplicate Page Content
-
I've attempted to follow advice from the Q&A section.
Currently on the site www.cherrycreekspine.com, I've edited the .htaccess file to help with 301s - all pages redirect to www.cherrycreekspine.com.
Secondly, I'd added the canonical statement in the header of the web pages.
I have cut the Duplicate Page Content in half ... now I have a remaining 40 pages to fix up.
This is my practice site to try and understand what SEOmoz can do for me.
I've looked at some of your videos on Youtube ... I feel like I'm scrambling around to the Q&A and the internet to understand this product. I'm reading the beginners guide.... any other resources would be helpful.
-
Hi Todd,
Oh no, looks like all your canonicals are pointing towards the homepage....
<link rel="<a class="attribute-value">canonical</a>" href="[http://www.cherrycreekspine.com<](view-source:http://www.cherrycreekspine.com%3C/)"/>
Also looks like there's an extra carrot "<" at the end of the URL. Looks like it's coming from the wonky code.
Regardless, this isn't what you want. This basically tells search engines that every page is a canonical of the homepage - and that all these other pages aren't important on their own. It's likely search engines will start to drop these pages out of their index unless this tag is removed immediately. Reminds me of Dr. Pete's canonical nightmares.
In short... remove that canonical ASAP. It's probably better to have some duplicate content than a sitewide canonical that points to your homepage.
Have you tried the Joomla Canonicalisation Plugin? Haven't tried it myself, but it might be smidgen easier than trying to code the php yourself. You can find it here: http://extensions.joomla.org/extensions/site-management/seo-a-metadata/url-canonicalization-/5355
My guess is you can completely remove whatever manual canonical code you wrote, and the plugin will take care of the rest.
Remember, when the code is working properly, each page with point to itself (without extra parameters and so on) the way a proper canonical should.
Best of luck!
-
View page source shows HTML. If I can't see the PHP file i.e. which generates the HTML it's impossible for me to know how. I'm not very clear on this. You can PM the file to me if you want - but don't send me any passwords please.
-
Yikes... so, I checked the template html file and all including PHP info is there...
View page source did not show all the code.
So, still stuck.
-
Hold on .... not all the copy was properly entered by me... I'll fix and retest.
-
Mr., I replaced the line in the html but no change. I'm completely new to this, so please forgive me. All of these concepts are new to me.
Any other thoughts or direction? I looked at one of the Whiteboard videos... buy over my head.
Thanks in advance... Todd
-
You need to code it again:
In the HTML I can see on Line 15:
rel="canonical" href="http://www.cherrycreekspine.com<?php echo parse_url($canonical, PHP_URL_PATH); ?>"/>
You need to write everything after href="
and before "/>
in php; probably:
<link rel="canonical" href="<?php echo="" '<a="" class=" " href="http://www.cherrycreekspine.com%26lt%3B/?php%20echo%20parse_url($canonical,%20PHP_URL_PATH);%20?%3E" target="_blank">http://www.cherrycreekspine.com<'.parse_url($canonical, PHP_URL_PATH); ?>"/></link rel="canonical" href="<?php>
If it doesnt sort it, post again and I'll try to help you further.
Also, you either apply canonical values or remove duplicate content (the post title is slightly confused, I think)
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is it best practice to have a canonical tags on all pages
The website I'm working on has no canonical tags. There is duplicate content so rel=canonicals need adding to certain pages but is it best practice to have a tag on every page ?
Intermediate & Advanced SEO | | ColesNathan0 -
SEM Rush & Duplicate content
Hi SEMRush is flagging these pages as having duplicate content, but we have rel = next etc implemented: https://www.key.co.uk/en/key/brand/bott https://www.key.co.uk/en/key/brand/bott?page=2 Or is it being flagged as they're just really similar pages?
Intermediate & Advanced SEO | | BeckyKey0 -
Removing duplicate content
Due to URL changes and parameters on our ecommerce sites, we have a massive amount of duplicate pages indexed by google, sometimes up to 5 duplicate pages with different URLs. 1. We've instituted canonical tags site wide. 2. We are using the parameters function in Webmaster Tools. 3. We are using 301 redirects on all of the obsolete URLs 4. I have had many of the pages fetched so that Google can see and index the 301s and canonicals. 5. I created HTML sitemaps with the duplicate URLs, and had Google fetch and index the sitemap so that the dupes would get crawled and deindexed. None of these seems to be terribly effective. Google is indexing pages with parameters in spite of the parameter (clicksource) being called out in GWT. Pages with obsolete URLs are indexed in spite of them having 301 redirects. Google also appears to be ignoring many of our canonical tags as well, despite the pages being identical. Any ideas on how to clean up the mess?
Intermediate & Advanced SEO | | AMHC0 -
Adding a Canonical Tag to each page referencing itself?
Hey Mozers! I've noticed that on www.Zappos.com they have a Canonical tag on each page referencing it self. I have heard that this is a popular method but I dont see the point in canon tagging a page to its self. Any thoughts?
Intermediate & Advanced SEO | | rpaiva0 -
Partial duplicate content and canonical tags
Hi - I am rebuilding a consumer website, and each product page will contain a unique product image, and a sentence or two about the product (and we tend to use a lot of the same words in different ways across products). I'd like to have a tabbed area below the product info that talks about the overall product line, and this content would be duplicate across all the product pages (a "Why use our products" type of thing). I'd have this duplicate content also living on its own URL's so they can be found alone in the SERP's. Question is, do I need to add the canonical tag to this page, since there's partial duplicate content on the product pages? And if I did that, would my product pages go un-indexed?? I understand how to handle completely duplicated content, it's the partial duplicate that I'm having difficulty figuring out.
Intermediate & Advanced SEO | | Jenny10 -
Artist Bios on Multiple Pages: Duplicate Content or not?
I am currently working on an eComm site for a company that sells art prints. On each print's page, there is a bio about the artist followed by a couple of paragraphs about the print. My concern is that some artists have hundreds of prints on this site, and the bio is reprinted on every page,which makes sense from a usability standpoint, but I am concerned that it will trigger a duplicate content penalty from Google. Some people are trying to convince me that Google won't penalize for this content, since the intent is not to game the SERPs. However, I'm not confident that this isn't being penalized already, or that it won't be in the near future. Because it is just a section of text that is duplicated, but the rest of the text on each page is original, I can't use the rel=canonical tag. I've thought about putting each artist bio into a graphic, but that is a huge undertaking, and not the most elegant solution. Could I put the bio on a separate page with only the artist's info and then place that data on each print page using an <iframe>and then put a noindex,nofollow in the robots.txt file?</p> <p>Is there a better solution? Is this effort even necessary?</p> <p>Thoughts?</p></iframe>
Intermediate & Advanced SEO | | sbaylor0 -
Duplicate content on subdomains.
Hi Mozer's, I have a site www.xyz.com and also geo targeted sub domains www.uk.xyz.com, www.india.xyz.com and so on. All the sub domains have the content which is same as the content on the main domain that is www.xyz.com. So, I want to know how can i avoid content duplication. Many Thanks!
Intermediate & Advanced SEO | | HiteshBharucha0 -
Could you use a robots.txt file to disalow a duplicate content page from being crawled?
A website has duplicate content pages to make it easier for users to find the information from a couple spots in the site navigation. Site owner would like to keep it this way without hurting SEO. I've thought of using the robots.txt file to disallow search engines from crawling one of the pages. Would you think this is a workable/acceptable solution?
Intermediate & Advanced SEO | | gregelwell0