Magento Layered Navigation & Duplicate Content
-
Hello Dear SeoMoz,
I would like to ask your help with something that I am not sure off. Our ecommerce web site is built with Magento. I have found many problems so far and I know that there will be many more in the future. Currently, I am trying to find the best way to deal with the duplicate content that is produced from the layered navigation (size, gender etc). I have done a lot of research so far in order to understand which might be the best practice and I found the following practices:
- **Block layered navigation URLSs from the Google Webmaster Tools (**Apparently this works for Google Only).
- Block these URLs with the robots.txt file
- Make links no-follow
- **Make links JavaScript from Magento ***
- Avoid including these links in the xml site map.
- Avoid including these link in the A-Z Product Index.
- Canonical tag
- Meta Tags (noindex, nofollow)
Question
If I turn the layered navigation links into JavaScript links from the Magento Admin, the layered navigation links are still found by the crawlers but they look like that:
|
instead of:
http://www.mysite.com/girls-basics.html?gender_filte...
|
Can these new URLS (http://www.mysite.com/# ) solve the duplicate content problems with the layered navigation or do I need to implement other practices too to make sure that everything is done right.
Kind Regards
Stefanos Anastasiadis
-
I'm not sure if you guys found a solution to this but I've used Mageworx with my Magento sites and it seems to handle everything I need. I do have to do some Mod-rewrites but nothing too much for a developer to handle.
-
From what I can gather about Magento is the Layered Nav can create seemingly endless URL's. Even if you were to use one of the modules created to make them 'friendly', you would still technically have reems of duplicate pages...right? All nicely re-written but effectively with the same titles and meta...
You may be able to put a wildcard disallow in the robots file for the parameter 'dir=' , which is associated with all the filters. I dont know how well this will work or if Google may on occasision ignore this or find a way into the layered pages anyway? Does anyone know? What if the spider entered the site through a direct link to filtered page...would the robots.txt file go by the way side in this instance?
You could in theory also use WMT to dictate that Google does not index pages with the 'dir=' parameter. Again, I am not sure as to the success rate using this.
Its one of those areas that has many open and unaswered discussions but nothing definitive anywhere to address the issue. Yet Magento is very popular and as you look at people sites who use it you can see they have some how found a way to sort this out. Id love to be a fly on the wall in their office!
-
Stefanos:
Hi! Did you ever find an answer to this question? I have a Magento install as well and need some advanced technical SEO. Are you working with a Magento consultant at all?
Thanks!
Lynn
-
Thanks a lot for your reply. I already know this extension but it is not what I am looking for.
-
I don't know if you stumble upon this Extension,
but it may resolve your problems.
http://www.magentocommerce.com/magento-connect/EcommerceTeam/extension/4420/layered_navigation_seo
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Products description from third party vendor creating duplicate content issues?
Hi, I am running my client's e-store. The store sells different products from various vendors. Vendors provide us product descriptions. The problem is that these vendors also give these description to display their products on other similar sites and hence creating duplicate content issue. Thanks.
On-Page Optimization | | Kashif-Amin0 -
Is minor duplicate content on my website okay?
I know duplicate content across multiple websites is not a good thing, however I've always wondered about minor duplicate content on your own website. I know its good practice to have unique content on each page but what about the little stuff. For example on our website certain related pages share the same content in a right sidebar. Such as links to pdf leaflets, or "you can read our blog etc" . Is there a minimum number of repeated words required before its flagged as duplicate content? Another example is a customer gave two testimonials for two of our employees - the testimonials were identical other than the employee names - if these were posted on separate pages is it a problem for the site as a whole or for both those individual pages? Thanks
On-Page Optimization | | Brabian0 -
Content with changing URL and duplicate content
Hi everyone, I have a question regarding content (user reviews), that are changing URL all the time. We get a lot of reviews from users that have been dining at our partner restaurants, which get posted on our site under (new) “reviews”. My worry however is that the URL for these reviews is changing all the time. The reason for this is that they start on page 1, and then get pushed down to page 2, and so on when new reviews come in. http://www.r2n.dk/restaurant-anmeldelser I’m guessing that this could cause for serious indexing problems? I can see in google that some reviews are indexed multiple times with different URLs, and some are not indexed at all. We further more have the specific reviews under each restaurant profile. I’m not sure if this could be considered duplicate content? Maybe we should tell google not to index the “new reviews section” by using robots.txt. We don’t get much traffic on these URLs anyways, and all reviews are still under each restaurant-profile. Or maybe the canonical tag can be used? I look forward to your input. Cheers, Christian
On-Page Optimization | | Christian_T2 -
Home Page Content
Hello. i'm optimizing this website, > home page for one keyword phrase and i was wondering how many words article do i need with that keyword?and if i need it at all? as you can see if i add some content on my home page before the slider, it will ruin the look of the website, What is the right way to do it? Thank you!
On-Page Optimization | | KentR0 -
Creating Duplicate Content on Shopping Sites
I have a client with an eCommerce site that is interested in adding their products to shopping sites. If we use the same information that is on the site currently, will we run into duplicate content issues when those same products & descriptions are published on shopping sites? Is it best practice to rewrite the product title and descriptions for shopping sites to avoid duplicate content issues?
On-Page Optimization | | mj7750 -
What is the best way to manage industry required duplicate Important Safety Information (ISI) content on every page of a site?
Hello SEOmozzer! I have recently joined a large pharmaceutical marketing company as our head SEO guru, and I've encountered a duplicate content related issue here that I'd like some help on. Because there is so much red tape in the pharmaceutical industry, there are A LOT of limitations on website content, medication and drug claims, etc. Because of this, it is required to have Important Safety Information (ISI) clearly stated on every page of the client's website (including the homepage). The information is generally pretty lengthy, and in some cases is longer than the non-ISI content on each page. Here is an example: http://www.xifaxan.com/ All content under the ISI header is required on each page. My questions are: How will this duplicated content on each page affect our on-page optimization scores in the eyes of search engines? Is Google seeing this simply as duplicated content on every page, or are they "smart" enough to understand that because it is a drug website, this is industry standard (and required)? Aside from creating more meaty, non-ISI content for the site, are there any other suggestions you have for handling this potentially harmful SEO situation? And in case you were going to suggest it, we cannot simply have an image of the content, as it may not be visible by all internet users. We've already looked into that 😉 Thanks in advance! Dylan
On-Page Optimization | | MedThinkCommunications0 -
How woud you deal with Blog TAGS & CATEGORY listings that are marked a 'duplicate content' in SEOmoz campaign reports?
We're seeing "Duplicate Content" warnings / errors in some of our clients' sites for blog / event calendar tags and category listings. For example the link to http://www.aavawhistlerhotel.com/news/?category=1098 provides all event listings tagged to the category "Whistler Events". The Meta Title and Meta Description for the "Whistler Events" category is the same as another other category listing. We use Umbraco, a .NET CMS, and we're working on adding some custom programming within Umbraco to develop a unique Meta Title and Meta Description for each page using the tag and/or category and post date in each Meta field to make it more "unique". But my question is .... in the REAL WORLD will taking the time to create this programming really positively impact our overall site performance? I understand that while Google, BING, etc are constantly tweaking their algorithms as of now having duplicate content primarily means that this content won't get indexed and there won't be any really 'fatal' penalties for having this content on our site. If we don't find a way to generate unique Meta Titles and Meta Descriptions we could 'no-follow' these links (for tag and category pages) or just not use these within our blogs. I am confused about this. Any insight others have about this and recommendations on what action you would take is greatly appreciated.
On-Page Optimization | | RoyMcClean0 -
Percentage of duplicate content allowable
Can you have ANY duplicate content on a page or will the page get penalized by Google? For example if you used a paragraph of Wikipedia content for a definition/description of a medical term, but wrapped it in unique content is that OK or will that land you in the Google / Panda doghouse? If some level of duplicate content is allowable, is there a general rule of thumb ratio unique-to-duplicate content? thanks!
On-Page Optimization | | sportstvjobs0