Duplicate Content/Missing Meta Description | Pages DO NOT EXISIT!
-
Hello all,
For the last few months, Moz has been showing us that our site has roughly 2,000 duplicate content errors. Pages that were actually duplicate content, I took care of accordingly using best practice (301 redirects, canonicalization,etc.). Still remaining after these fixes were errors showing for pages that we have never created.
Our homepage is www.primepay.com. An example of pages that are being shown as duplicate content is http://primepay.com/blog/%5BLink%20to%20-%20http:/www.primepay.com/en/payrollservices/payroll/payroll/payroll/online-payroll with a referring page of http://primepay.com/blog/%5BLink%20to%20-%20http:/www.primepay.com/en/payrollservices/payroll/payroll/online-payroll. Some of these are even now showing up as 403 and 404 errors.
The only real page on our site within that URL strand is primepay.com/payroll or primepay.com/payroll/online-payroll. Therefore, I am not sure where Moz is getting these pages from.
Another issue we are having in relation to duplicate content is that moz is showing old campaign url’s tacked on to our blog page i.e. http://primepay.com/blog?title=&page=2&utm_source=blog&utm_medium=blogCTA&utm_campaign=IRSblogpost&qt-blog_tabs=1.
As of this morning, our duplicate content went from 2,000 to 18,000. I exported all of our crawl diagnostics data and looked to see what the referring pages were, and even they are not pages that we have created. When you click on these links, they take you to a random point in time from the homepage of our blog; some dating back to 2010.
I checked our crawl stats in both Google and Bing’s Webmaster tool, and there are no duplicate content or 400 level errors being reporting from their crawl. My team is truly at a loss with trying to resolve this issue and any help with this matter would be greatly appreciated.
-
Thanks Dirk. Very insightful tip about not using campaign tracking to check internal links. There was an old blog post that had anchor text with campaign tracking that was causing many SEO issues. As for the latter part, it is unknown why a string of gibberish can be placed after /blog/ and also for our locations page. Our team's web developer is looking further into this issue. If anyone has any more advice on the matter it would be greatly appreciated.
-
Hey there
Dirk pretty much hit upon the issue, which I'll reiterate with a visual. If you enter any gibberish /blog URL (like this: http://primepay.com/blog/jglkjglkjg) in the browser it returns a 200 OK which, but it should return a 404 code --> http://screencast.com/t/cStpPB5zE
Otherwise pages that are really broken will look to crawlers like they are supposed to exist.
-
You shouldn't use campaign tracking to check internal links - you have to use event tracking. Check http://cutroni.com/blog/2010/03/30/tracking-internal-campaigns-with-google-analytics/ . Apart from the reporting issue - it's also generating a huge number of url's that need to be crawled by Google bot and is just wasting it's time (most of these tagged url have a correct canonical version). You mention these tags are old - but they are still present on a lost of pages.
For cases like this it's better to check with a local tool like Screaming Frog which gives you a much better view which pages are generating these links.The other issue you have is probably related to a few pages that have a bad formatted (relative) url in a link - the way your site is configured it's just rendering a page on your site - so the bots are then crawling your site over and over again, each time encountering the same bad relative link - and each time adding the bad formatting to the url. It's an endless loop - best way to avoid this is to use absolute internal links rather than relative links. Not sure if it's the only one - but one of the pages with this error is :http://primepay.com/blog/7-ways-find-right-payroll-service-your-company - it contains a link to
[Your payroll service is no different.]([Link to - http://www.primepay.com/en/payrollservices/] "Your payroll service is no different.")
This page should generate a 404 but is generating a 200 and the loop starts here.
Again - with screaming frog you can for each of these bad url's you can generate a crawl path report which shows you exactly on which page the error is generated.
Hope this helps,
Dirk
-
Example:
http://primepay.com/blog/hgehergreg
Status:
My site as an example:
https://caseo.ca/blog/hgehergreg
If I put in random gibberish in this URL, it should be displaying a 404 page and not the blog page.
-
Getting you some help for direct advice on your problem, but wanted to leave a comment about the tool itself. When you are looking at the Moz crawl tool, it only updates once a week, so if there hasn't been that long between the last crawl and when you did the work, it won't be updated. Here's more info.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Blog article split into a few pages
Hi, I came across multiple websites where blog articles are split into multiple pages and once you reach the end of the page, you need to click 'Next' or '2/3' to continue reading. Is this a good practice? I understand you keep clicking through multiple pages instead of reading an article only on one page, but you also split the strength of the targeted keyword between multiple pages. Can anyone advise? Thanks Katarina
Moz Pro | | Katarina-Borovska0 -
Missing Description Tag
I pulled the latest moz report and my category pages are being flagged as missing description tag, but not only is the description visible on each page, it is also in the code for all to see. Here is the code, what am I missing? | |
Moz Pro | | moon-boots
| | |
| | |
| | |
| | dir="ltr" class="ltr" lang="en"> | | |
| | |
| | |
| | |
| | |
| | <title>Astronaut Costumes</title> |
| | <base href="<a href=" http:="" www.interstellarstore.com="" "="">http://www.interstellarstore.com/" /> |
| | |
| | |
| | http://www.interstellarstore.com/image/catalog/Earth.png" rel="icon" /> |
| | catalog/view/theme/pav_WindOnline_Store/stylesheet/bootstrap.css" rel="stylesheet" /> |
| | catalog/view/theme/pav_WindOnline_Store/stylesheet/stylesheet.css" rel="stylesheet" /> |
| | catalog/view/theme/pav_WindOnline_Store/stylesheet/customize/1455569423.css" rel="stylesheet" /> |
| | catalog/view/javascript/font-awesome/css/font-awesome.min.css" rel="stylesheet" /> |
| | catalog/view/theme/pav_WindOnline_Store/stylesheet/animate.css" rel="stylesheet" /> |
| | catalog/view/javascript/jquery/magnific/magnific-popup.css" rel="stylesheet" /> |
| | catalog/view/javascript/jquery/owl-carousel/owl.carousel.css" rel="stylesheet" /> |
| | catalog/view/theme/pav_WindOnline_Store/stylesheet/fonts.css" rel="stylesheet" /> |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |0 -
Magento Dynamic Pages Being Indexed
Hi there, I have about 50k Moz medium priority errors in my Crawl Diagnostic report. The bulk of them are classified as "Temporary Redirect" problems. Then if you drill into those further, I can see that the problem urls all kinda are center around: mysite.com/catalogsearch/result.. mysite.com/wishlist.. mysite.com/catalog.. Is this something I should disallow in my Robstxt file? And if so how specific do I get with it.. Disallow /catalogsearch/result/?q= Will listing the /catalogsearch be enough to cover anything after it? thanks
Moz Pro | | Shop-Sq0 -
Aren't domain.com/page and domain.com/page/ the same thing?
Hi All, A recent Moz scan has turned up quite a few duplicate content notifications, all of which have the same issue. For instance: domain.com/page and domain.com/page/ are listed as duplicates, but I was under the impression that these pages would, in fact, be the same page. Is this even something to bother fixing or a fluke scan? If I should fix it does anyone know of an .htaccess modification that might be used? Thanks!
Moz Pro | | G2W0 -
Does a url with no trailing slash (/)need A special redirect to the same url with a trailing slash (/)
I recently moved a website to wordpress which the wordpress default includes the trailing slash (/) after ALL urls. My url structure used to look like: www.example.com/blue-widgets Now it looks like: www.example.com/blue-widgets/ Today I checked the urls using Open Site Explorer and below is what I discovered: www.example.com/blue-widgets returned all my links, authority, etc HOWEVER there is a note that says........."Oh Hey! it looks like that URL redirects to www.example.com/blue-widgets/. Would you like to see data for that URL instead?" When I click on the link to THAT URL I get a note that says_.....NO DATA AVAILABLE FOR THIS URL._ Does this mean that www.example.com/blue-widgets/ really has NO DATA? How do I fix this?
Moz Pro | | webestate0 -
Too many on-page links
one of my SEOmoz pro campaigns has given me the warning: Too many on-page links and the page in question is my html sitemap. How do i resolve this because I obviously need my sitemap. How do i get around this?
Moz Pro | | CompleteOffice1 -
Duplicate page title
I own a store www.mzube.co.uk and the scam always says that I have duplicate page titles or duplicate page. What happens is thn I may have for example www.mzube.co.uk/allproducts/page1. And if I hve 20 pages all what will change from each page is the number at the end and all the rest of the page name will be the same but really the pages are if different products. So the scans think I have 20 pages the same but I havent Is this a concern as I don't think I can avoid this Hope you can answer
Moz Pro | | mzube0 -
On Page Report Card... with or w/o local modifiers?
Hey all! So I am curious how you recommend using the "on page report card" (which is really helpful) along with the concept of local modifiers. IE, here is a term I am going after: business forums but really I care about a specific location: business forums | Greensboro NC So the word I hear is typically to do your keyword research & page optimization FOR the primary term, but then tack on your local modifiers after. So which do you run reports on? Probably both is the best answer, eh? Obviously my local sites won't have a shot at ranking nationally/internationally for such a broad term as "business forums", especially with some monster sites out there with some serious clout. This is more of a best practices question. Thanks dudes.
Moz Pro | | nsmcseo20