Thousands of 404-pages, duplicate content pages, temporary redirect
-
Hi,
i take over the SEO of a quite large e-commerce-site. After checking crawl issues, there seems to be +3000 4xx client errors, +3000 duplicate content issues and +35000 temporary redirects. I'm quite desperate regarding these results. What would be the most effective way to handle that. It's a magento shop.
I'm grateful for any kind of help!
Thx,
boris -
Hey guys,
thanks for your reactions. Appreciate it!
I guess it's time to roll up the sleaves...
cheers,
Boris
-
+1 to what Danny said. A couple other thoughts:
- +3000 4xx client errors: do any of these have many links or a significant amount of traffic? If not and they aren't a large portion of your overall site, it's not a big deal.
- **+3000 duplicate content issues: **as Danny said, there is likely a trend here, try to identify it and resolve it in mass rather than going page by page.
- **+35000 temporary redirects: **are any of these temporary redirects to important pages? If so, it's worth changing them to 301s. However, if they are all pointing to old, deep and weak pages then it's likely not a big concern again.
Daniel
-
This happens with all ecommerce platforms.
This is usually due to the categories on your site duplicating your pages.
You may have one product that is available in different colours so two links are being created. For example www.car.com/new-car
and
www.car.com/new-car=blue might be the exact same page.
Search engines are unsure which page to index from your website and they are very unlikely to show multiple or duplicate product pages within their index.
So, you need to inform search engines which page you wish for them to index.
The best way to solve these issues is to simply go through each error and solve it. You will start to notice a pattern in the duplicate content URLs and redirects. This should speed up the process. You could either add in canonical tags or simply use your robots file to block google crawling particular URL extensions. I have stopped bots from crawling my ecommerce pages that end with the parameter "route=product/search&tag" and "product-id" as these are non SEO friendly URLS that are duplicated versions of my pages.
Make sure that you also remove dead links and remove links that are going to versions of the page you don't want them too.
It's a lengthy process but it needs to be done.
Danny
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Getting rid of pagination - redirect all paginated pages or leave them to 404?
Hi all, We're currently in the process of updating our website and we've agreed that one of the things we want to do is get rid of all our pagination (currently used on the blog and product review areas) and instead implement load more on scroll. The question I have is... should we redirect all of the paginated pages and if so, where to? (My initial thoughts were either to the blog homepage or to the archive page) OR do we leave them to just 404? Bear in mind we have thousands of paginated pages 😕 Here's our blog area btw - https://www.ihasco.co.uk/blog Any help would be appreciated, thanks!
Technical SEO | | iHasco0 -
Database driven content producing false duplicate content errors
How do I stop the Moz crawler from creating false duplicate content errors. I have yet to submit my website to google crawler because I am waiting to fix all my site optimization issues. Example: contactus.aspx?propid=200, contactus.aspx?propid=201.... these are the same pages but with some old url parameters stuck on them. How do I get Moz and Google not to consider these duplicates. I have looked at http://moz.com/learn/seo/duplicate-content with respect to Rel="canonical" and I think I am just confused. Nick
Technical SEO | | nickcargill0 -
Client error 404 pages!
I have a number of 404 pages coming up which are left over in Google from the clients previous site. How do I get them out of Google please?
Technical SEO | | PeterC-B0 -
Duplicate content by php id,page=... problem
Hi dear friends! How can i resolve this duplicate problem with edit the php code file? My trouble is google find that : http://vietnamfoodtour.com/?mod=booking&act=send_booking&ID=38 and http://vietnamfoodtour.com/.....booking.html are different page, but they are one but google indexed both of them. And the Duplcate content is raised 😞 how can i notice to google that they are one?
Technical SEO | | magician0 -
Duplicate content due to csref
Hi, When i go trough my page, i can see that alot of my csref codes result in duplicate content, when SeoMoz run their analysis of my pages. Off course i get important knowledge through my csref codes, but im quite uncertain of how much it effects my SEO-results. Does anyone have any insights in this? Should i be more cautios to use csref-codes or dosent it create problems that are big enough for me to worry about them.
Technical SEO | | Petersen110 -
API for testing duplicate content
Does anyone know a service or API or php lib to compare two (or more) pages and to return their similiarity (Level-3-Shingles). API would be greatly prefered.
Technical SEO | | Sebes0 -
Sharing the same content on every page
As an ecommerce site, one of the tabs on the product description is filled with delivery information. This tab is populated the same way on every product page. I think this is contributing to an increased score on my pages similarity to each other. Is there a way to obscure this info for se's and is it worthwhile doing so?
Technical SEO | | LadyApollo0 -
Duplicate Content issue
I have been asked to review an old website to an identify opportunities for increasing search engine traffic. Whilst reviewing the site I came across a strange loop. On each page there is a link to printer friendly version: http://www.websitename.co.uk/index.php?pageid=7&printfriendly=yes That page also has a link to a printer friendly version http://www.websitename.co.uk/index.php?pageid=7&printfriendly=yes&printfriendly=yes and so on and so on....... Some of these pages are being included in Google's index. I appreciate that this can't be a good thing, however, I am not 100% sure as to the extent to which it is a bad thing and the priority that should be given to getting it sorted. Just wandering what views people have on the issues this may cause?
Technical SEO | | CPLDistribution0