What to do about similar product pages on major retail site
-
Hi all,
I have a dilemma and I'm hoping the community can guide me in the right direction. We're working with a major retailer on launching a local deals section of their website (what I'll call the "local site"). The company has 55 million products for one brand, and 37 million for another.
The main site (I'll call it the ".com version") is fairly well SEO'd with flat architecture, clean URLs, microdata, canonical tag, good product descriptions, etc.
If you were looking for a refrigerator, you would use the faceted navigation and go from department > category > sub-category > product detail page.
The local site's purpose is to "localize" all of the store inventory and have weekly offers and pricing specials. We will use a similar architecture as .com, except it will be under a /local/city-state/... sub-folder.
Ideally, if you're looking for a refrigerator in San Antonio, Texas, then the local page should prove to be more relevant than the .com generic refrigerator pages. (the local pages have the addresses of all local stores in the footer and use the location microdata as well - the difference will be the prices.)
MY QUESTION IS THIS:
If we pull the exact same product pages/descriptions from the .com database for use in the local site, are we creating a duplicate content problem that will hurt the rest of the site?
I don't think I can canonicalize to the .com generic product page - I actually want those local pages to show up at the top. Obviously, we don't want to copy product descriptions across root domains, but how is it handled across the SAME root domain?
Ideally, it would be great if we had a listing from both the .com and the /local pages in the SERPs.
What do you all think?
Ryan
-
Hi Ryan,
I guess the first point here is that Google doesn't treat this sort of filtering as "penalisation"; it's just filtering two or more versions of the same content because it believes (sometimes mistakenly) that users don't need to see two versions of the same thing. This gets REALLY tricky in fields like real estate when all the aggregators in the same town have access to pretty much the same feeds or properties.
If Google were perfect, you'd put up the two pieces of identical content for all 55 millions products, and Google would serve the right one given the appropriate query, like the example above ("fridge sale san antonio" brings up the local page; "refrigerator" has your main site rank). And this might happen, because Google is getting better at these sort of query-appropriate results. We still recommend not providing dupe content solely because we can't be sure that Google will get it right.
As an aside, it would be so great if they worked on a tool for localisation in the same way that they have given us the href lang tag for internationalisation. rel="city" or similar would be awesome, especially for big countries.
Your idea about serving the content from a shared source will certainly work (iframe, text hosted on separate URL, JS etc.). The pages serving this text clearly won't be credited with that text's content, which removes its SEO value of course.
-
Hi Jane, thanks for the response!
I can't understand why Google or any other search engine would penalize a brand for having the same product detail in more than one location on the same root domain. It's just not feasible to re-write all of the product descriptions for 55 million products. The only difference is going to be the price, and some localized content on the page in terms of store locations and addresses (perhaps multiple in one area).
What if - kind of like your M&S example - the local product pages pulled product descriptions from another location on the site, but displayed them in a modal window - so a JS event displayed the proper descriptions and details for the user experience, but the HTML is devoid of any "duplicate" product description content?
-
Hi Ryan,
It's going to be hard to do this without creating duplicates - if they aren't commissioning re-writes of descriptions but just pulling from the database, identical content like this is far from ideal.
One school of thought is that there really isn't any such thing as a "duplicate content penalty" unless you have some huge, gratuitous problem that results in a Panda issue. Google simply chooses the version of the content it favours and drops the other. The local site would still be much more relevant for a query like "fridge sale san antonio".
An example of a big retailer that has a similar(ish) site at the moment is Marks & Spencer Outlet here in the UK (outlet.marksandspencer.com). M&S is probably the most recognisable high street brand in the UK, to give you a perspective on size.
Looking at what they're doing, they're listing pages like this: http://outlet.marksandspencer.com/Limited-Edition-Jacquard-Textured-T69-1604J-S/dp/B00IIP7GY2?field_availability=-1&field_browse=1698309031&id=Limited+Edition+Jacquard+Textured+T69-1604J-S&ie=UTF8&refinementHistory=subjectbin%2Csize_name%2Ccolor_map%2Cbrandtextbin%2Cprice&searchNodeID=1698309031&searchPage=1&searchRank=-product_site_launch_date&searchSize=12
This is the same product as this: http://www.marksandspencer.com/jacquard-textured-coat-with-wool/p/p60056127. I love it that the "outlet" version is more expensive... anyway...
The product details, which are all included in the HTML of the main site, are not included in the Outlet page. The Outlet URL is indexed (what queries it ranks for / could potentially rank for are unknown) - but I would be keen to hypothesise / experiment with the idea that if that product was on a page about it only being available at M&S Moorgate, and looking for coats at M&S Moorgate was as popular a query as [fridge sale location], the Outlet page would rank.
You will never get an SEO to say that you should "copy and paste" descriptions across domains or within them, but essentially the pages have to provide a service / information that makes them worth ranking for relevant queries.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can noindexed pages accrue page authority?
My company's site has a large set of pages (tens of thousands) that have very thin or no content. They typically target a single low-competition keyword (and typically rank very well), but the pages have a very high bounce rate and are definitely hurting our domain's overall rankings via Panda (quality ranking). I'm planning on recommending we noindexed these pages temporarily, and reindex each page as resources are able to fill in content. My question is whether an individual page will be able to accrue any page authority for that target term while noindexed. We DO want to rank for all those terms, just not until we have the content to back it up. However, we're in a pretty competitive space up against domains that have been around a lot longer and have higher domain authorities. Like I said, these pages rank well right now, even with thin content. The worry is if we noindex them while we slowly build out content, will our competitors get the edge on those terms (with their subpar but continually available content)? Do you think Google will give us any credit for having had the page all along, just not always indexed?
Intermediate & Advanced SEO | | THandorf0 -
Product Page on Eccomerce Site ranking very poorly - Unique Product description but duplicate content on other tabs.
Hi All, I have a query regarding my Product pages on my eCommerce site. I have unique Product descriptions but some of the other page content on the other tabs i.e Hire Terms , Delivery , About the Hire Company - Are duplicated across ALL my products. Is that okay or how should I deal with them ? See example url of one of my products below below - http://goo.gl/aSFPqP My products currently rank very badly... 200 + so Any advice would be greatly appreciated thanks Peter
Intermediate & Advanced SEO | | PeteC120 -
Sort term product pages and fast indexing - XML sitemaps be updated daily, weekly, etc?
Hi everyone, I am currently working on a website that the XML sitemap is set to update weekly. Our client has requested that this be changed to daily. The real issue is that the website creates short term product pages (10-20 days) and then the product page URL's go 404. So the real problem is quick indexing not daily vs weekly sitemap. I suspect that daily vs weekly sitemaps may help solve the indexing time but does not completely solve the problem. So my question for you is how can I improve indexing time on this project? The real problem is how to get the product pages indexed and ranking before the 404 page shows u?. . Here are some of my initial thoughts and background on the project. Product pages are only available for 10 to 20 days (Auction site).Once the auction on the product ends the URL goes 404. If the pages only exist for 10 to 20 days (404 shows up when the auction is over), this sucks for SEO for several reasons (BTW I was called onto the project as the SEO specialist after the project and site were completed). Reason 1 - It is highly unlikely that the product pages will rank (positions 1 -5) since the site has a very low Domain Authority) and by the time Google indexes the link the auction is over therefore the user sees a 404. Possible solution 1 - all products have authorship from a "trustworthy" author therefore the indexing time improves. Possible solution 2 - Incorporate G+ posts for each product to improve indexing time. There is still a ranking issue here since the site has a low DA. The product might appear but at the bottom of page 2 or 1..etc. Any other ideas? From what I understand, even though sitemaps are fed to Google on a weekly or daily basis this does not mean that Google indexes them right away (please confirm). Best case scenario - Google indexes the links every day (totally unrealistic in my opinion), URL shows up on page 1 or 2 of Google and slowly start to move up. By the time the product ranks in the first 5 positions the auction is over and therefore the user sees a 404. I do think that a sitemap updated daily is better for this project than weekly but I would like to hear the communities opinion. Thanks
Intermediate & Advanced SEO | | Carla_Dawson0 -
"No index" page still shows in search results and paginated pages shows page 2 in results
I have "no index, follow" on some pages, which I set 2 weeks ago. Today I see one of these pages showing in Google Search Results. I am using rel=next prev on pages, yet Page 2 of a string of pages showed up in results before Page 1. What could be the issue?
Intermediate & Advanced SEO | | khi50 -
De-indexing product "quick view" pages
Hi there, The e-commerce website I am working on seems to index all of the "quick view" pages (which normally occur as iframes on the category page) as their own unique pages, creating thousands of duplicate pages / overly-dynamic URLs. Each indexed "quick view" page has the following URL structure: www.mydomain.com/catalog/includes/inc_productquickview.jsp?prodId=89514&catgId=cat140142&KeepThis=true&TB_iframe=true&height=475&width=700 where the only thing that changes is the product ID and category number. Would using "disallow" in Robots.txt be the best way to de-indexing all of these URLs? If so, could someone help me identify how to best structure this disallow statement? Would it be: Disallow: /catalog/includes/inc_productquickview.jsp?prodID=* Thanks for your help.
Intermediate & Advanced SEO | | FPD_NYC0 -
Ecommerce SEO - Indexed product pages are returning 404's due to product database removal. HELP!
Hi all, I recently took over an e-commerce start-up project from one of my co-workers (who left the job last week). This previous project manager had uploaded ~2000 products without setting up a robot.txt file, and as a result, all of the product pages were indexed by Google (verified via Google Webmaster Tool). The problem came about when he deleted the entire product database from our hosting service, godaddy and performed a fresh install of Prestashop on our hosting plan. All of the created product pages are now gone, and I'm left with ~2000 broken URL's returning 404's. Currently, the site does not have any products uploaded. From my knowledge, I have to either: canonicalize the broken URL's to the new corresponding product pages, or request Google to remove the broken URL's (I believe this is only a temporary solution, for Google honors URL removal request for 90 days) What is the best way to approach this situation? If I setup a canonicalization, would I have to recreate the deleted pages (to match the URL address) and have those pages redirect to the new product pages (canonicalization)? Alex
Intermediate & Advanced SEO | | byoung860 -
How to Build High Quality eCommerce Web Site during Low Quality Web Pages?
Today, I was reading Official Google Webmaster Central Blog: More guidance on building high-quality sites. I found one interesting statement over there. Low-quality content on some parts of a website can impact the whole site’s rankings. Why should I like to discuss on this topic? Because, I have made big change on my website via narrow by search. I want to give specific result to know more about it. This is my category page: http://www.vistastores.com/patio-umbrellas Left narrow by search section is creating accurate page for specific attribute products. California Umbrella:
Intermediate & Advanced SEO | | CommercePundit
http://www.vistastores.com/patio-umbrellas/shopby/manufacturer-california-umbrella From above page following page is accessible. http://www.vistastores.com/patio-umbrellas/shopby/canopy-shape-search-octagonal/manufacturer-california-umbrella Sunbrella Patio Umbrellas:
http://www.vistastores.com/patio-umbrellas/shopby/canopy-fabric-search-sunbrella Similar story for this page. Following page can accessible from above page. http://www.vistastores.com/patio-umbrellas/shopby/canopy-fabric-search-sunbrella/finish-search-wood My website have 100+ categories, 11,000 products. I have checked indexed pages in Google for my website. https://www.google.com/search?q=info%3Awww.vistastores.com&pws=0&gl=US#hl=en&safe=off&pws=0&gl=US&q=site:www.vistastores.com&bav=on.2,or.r_gc.r_pw.r_cp.r_qf.,cf.osb&fp=910893d99351c8f7&biw=1366&bih=547 It shows me 35,000+ crawled pages which are developed by left navigation section. So, Will it consider as low quality pages? I want to improve my website performance without delete these pages.0 -
On-Site Optimization Tips for Job site?
I am working on a job site that only ranks well for the homepage with very low ranking internal pages. My job pages do not rank what so ever and are database driven and often times turn to 404 pages after the job has been filled. The job pages have to no content either. Anybody have any technical on-site recommendations for a job site I am working on especially regarding my internal pages? (Cross Country Allied.com)
Intermediate & Advanced SEO | | Melia0