Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
ECommerce: Best Practice for expired product pages
-
I'm optimizing a pet supplies site (http://www.qualipet.ch/) and have a question about the best practice for expired product pages.
We have thousands of products and hundreds of our offers just exist for a few months. Currently, when a product is no longer available, the site just returns a 404. Now I'm wondering what a better solution could be:
1. When a product disappears, a 301 redirect is established to the category page it in (i.e. leash would redirect to dog accessories).
2. After a product disappers, a customized 404 page appears, listing similar products (but the server returns a 404)
I prefer solution 1, but am afraid that having hundreds of new redirects each month might look strange. But then again, returning lots of 404s to search engines is also not the best option.
Do you know the best practice for large ecommerce sites where they have hundreds or even thousands of products that appear/disappear on a frequent basis? What should be done with those obsolete URLs?
-
Unfortunately manually.
-
Yep, on two different sites we did thousands of redirects at a time with no issues. In one case it was annual and the other it was quarterly but I don't see any reason monthly would be any different.
Definitely post your findings after implementation or maybe even write a YouMoz post about what you find out!
-
Good luck
-
Thanks for your thoughts guys.
@Igal@Incapsula: I like your 302 idea! That might acutally make a lot of sense for some products that are short-lived.
@Matthew: Good to know that lots of 301s were not an issue on your sites. Are you talking about thousands of those, though?
Most importantly, I will have to find something that can be automated and doesn't require much extra-work. I will probably go for 301s and remove those after a few months
Remind me to post my learnings here after implementation:)
-
(+1) For redirect to main category page option. I did this several time, including for a very large tourism site which had a LOT of "inventory" changes (we are talking about dozens-hundreds/day) and had great results.
One thing I would like to suggest is to look into doing 302 and removing the redirects after 2-3 month.
The reason for this is purely practical. In our case, after just a few month, we were looking at many thousands of redirects and this is not something you want to "carry around".
My suggestion allows you to still make use of link juice for removed pages and, at the same time, have a manageable redirect profile.As a safe net you can have a generic: "404 >>> 301 >>> Homepage" rule underneath.
-
Hey,
In general, I would opt for option 1 as that would be the most scale-able solution. Whenever I've done this, I've not seen any issues with having lots of 301s appear. Given the shorter life span of those product pages you probably won't have lots of links going to those pages (or social, etc.) and I think that helps explain why I've not seen issues redirecting this many pages.
That being said, if you do have lots of links or social signals referencing a certain product page, that is when I'd opt for the custom page listing similar products. I've had success doing this for high-traffic product pages that have been removed as it can help maintain the sale. In terms of the signal, it really depends. If you are still offering unique content relevant to search queries and links referencing that page, I'd deliver a status 200 (it is still a good page worthy of attention). If the content isn't all that unique, and it is more for people (to maintain the sale) as opposed to search, I would have that page deliver a status 410 (saying it is gone).
I hope that helps!
Matthew
-
thanks Kevin, so you're also going with option 1.
Do you make those redirects manually, or does it run automated?
I should add that it's a Magento Webshop and we definitely need some automation since I am talking about hundreds of product pages.
-
We have a customize search page for each category. When a product has been discontinued, we do a 301 redirect those pages to the category search page.
We use to do a 301 redirect of list similar products (by doing a search and capturing the url with the search term), but it proved to be to time-consuming as these products did not traditionally sold that well and did not bring in much traffic.
Not saying it's the best way, but this is what we do.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Product Schema Markup for All Products
Hi Team, Google search console used to allow you to use their structured data markup helperhttps://www.google.com/webmasters/markup-helper/u/0/ to markup multiple product pages at once that were similar. I do not see this feature anymore with the new search console. Does anyone have a recommendation for marking up multiple product pages without having to have schema markup firing in GTM for each product page?
Technical SEO | | agrier0 -
Canonical for duplicate pages in ecommerce site and the product out of stock
I’m an SEO for an ecommerce site that sells shoes I have duplicate pages for different colors of the same product (unique URL for each color), Conventionally I have added canonical tags for each page, which direct to a specific product URL My question is what happens when a product which the googlbot is direct to, is out of stock but is still listed in the canonical tag ?
Technical SEO | | shoesonline0 -
Best practice for URL - Language/country
Hi, We are planning on having our website localized into more languages. We already have an English and German version. The German version is currently a sub-domain: www.example.com --> English version de.example.com --> German version Is this recommended? Or is it always better to have URLs with language prefixes such a: www.example.com/de www.example.com/es Which is a better practice in terms of SEO?
Technical SEO | | Kilgray1 -
Duplicate content on Product pages for different product variations.
I have multiple colors of the same product, but as a result I'm getting duplicate content warnings. I want to keep these all different products with their own pages, so that the color can be easily identified by browsing the category page. Any suggestions?
Technical SEO | | bobjohn10 -
Image Height/Width attributes, how important are they and should a best practice site include this as std
Hi How important are the image height/width attributes and would you expect a best practice site to have them included ? I hear not having them can slow down a page load time is that correct ? Any other issues from not having them ? I know some re social sharing (i know bufferapp prefers images with h/w attributes to draw into their selection of image options when you post) Most importantly though would you expect them to be intrinsic to sites that have been designed according to best practice guidelines ? Thanks
Technical SEO | | Dan-Lawrence0 -
Best way to handle pages with iframes that I don't want indexed? Noindex in the header?
I am doing a bit of SEO work for a friend, and the situation is the following: The site is a place to discuss articles on the web. When clicking on a link that has been posted, it sends the user to a URL on the main site that is URL.com/article/view. This page has a large iframe that contains the article itself, and a small bar at the top containing the article with various links to get back to the original site. I'd like to make sure that the comment pages (URL.com/article) are indexed instead of all of the URL.com/article/view pages, which won't really do much for SEO. However, all of these pages are indexed. What would be the best approach to make sure the iframe pages aren't indexed? My intuition is to just have a "noindex" in the header of those pages, and just make sure that the conversation pages themselves are properly linked throughout the site, so that they get indexed properly. Does this seem right? Thanks for the help...
Technical SEO | | jim_shook0 -
Cloaking? Best Practices Crawling Content Behind Login Box
Hi- I'm helping out a client, who publishes sale information (fashion sales etc.) In order for the client to view the sale details (date, percentage off etc.) they need to register for the site. If I allow google bot to crawl the content, (identify the user agent) but serve up a registration light box to anyone who isn't google would this be considered cloaking? Does anyone know what the best practice for this is? Any help would be greatly appreciated. Thank you, Nopadon
Technical SEO | | nopadon0 -
Merging several sites into one - best practice
I had 2 sites on the web (www.physicseditor.de, www.texutrepacker.com) and decided to move them all under one single domain (www.codeandweb.com) Both sites were ranking very good for several keywords. I not redirected the most important pages from the old domains with a 301 redirect to the new subpages (www.texturepacker.com => www.codeandweb.com/texturepacker) Google still delivers the old domains but the redirect take people directly to the new content. I've already submitted the new site map to google webmaster tools. Pages are already in the index but do not really show up in the search results. How long does it take until google accepts the new domain and delivers the new content in the search results? Was it ok what I did? Or is there some room for improvement? SeoMoz will of course not find any information about the new page since it is not yet directly linked in google. But I can't get ranking information for the "old" pages since SeoMoz tells me that it can't crawl the old domains....
Technical SEO | | gossi740