Huge in crease in on page errors
-
Hi guys I’ve just checked my online campaign and I see errors in my crawl diagnostics have almost doubled from the 21<sup>st</sup> of October to the 25<sup>th</sup> of October going from 6708 errors to 11 599. Can anyone tell me what may have caused this?
Also I notice we have a lot of issues with duplicate page titles which seems strange as no new pages have been added, can anyone explain why this might be?
I look forward to hearing from you
-
Just noting that this discussion continues here:
-
Hi Matt,
I hope things are going well. I'm just following up on the duplicate page issue .I have spoken to our web company and they have correct the issue last week and I noted the amount of duplicates dropped significantly. Ive just checked today and i see it's back up (I'm following this up with our web company). Are you able to offer any insights as to why this problem seems to reoccur
I was of the understanding this is a permanent fix so once the change has been made, I cant understand why it then seems to reoccur?. Any insights would be much appreciated.
Regards
Pete
-
Well, it stands to reason that something must've changed is order to cause such a huge increase. Looking through the list of duplicate URLs, I'm seeing a lot that could be fixed by rel="canonical". There's enough of them that adding a canonical link to each would be a huge undertaking or require some careful coding. I'm wondering if this increase could've been partially caused by someone removing rel="canonical" from a lot of pages.
For example, I'm seeing a lot of this:
http://www.health2000.co.nz/shop/aromatherapy/lemongrass-essential-oil/P4494/C56
vs.
http://www.health2000.co.nz/shop/aromatherapy/lemongrass-essential-oil/p4494/c56
The only difference between those URLs is capitalization. The first, capitalized version is the one that appears on your XML sitemap. I'm not 100% sure why both versions would be appearing to Roger—it may be an issue with the CMS—but a rel="canonical" on the former pointing to the latter would solve that problem.
Now, that doesn't look to be the only issue, but it _is _a large one.
Let me know what you find out!
-
That's ok Matt. I've put those two questions to our web company, although I don't think any changes happened then, although I do know that they did work on the 27th of September. I m fairly sure in was rel =canonical in nature. I have asked them to confirm and will let you know in due course. As an aside why do you think the changes you mentioned would be of effect on our web site?
-
Hi Pete,
Sure thing. Sorry it's taken so long!
May I ask what, if any, changes were made on the site between the 21st and 25th of October? In particular, were there any changes made involving rel="canonical"?
-
Hi Matt,
Ive just had a look an I can now see the amount of pages, crawled on our site, is comparable to our competitors, which is great!
However the amount of on page errors is significantly higher. In particularly the amount of duplicate errors is about 10 000 which is the same amount we had, before our web company fixed this issue. Are you able to give me any feed back as to what's happened there?. Thanks again for your help with this!
Pete
-
Hi Matt,
No probs, I look forward to hearing from you
Regards
Pete
-
That's awesome Keri, thanks for following this up : )
-
Hi Pete,
Sorry for the delay! I just wanted to let you know I'm looking into it, and should get back to you shortly.
Matt
-
Hi Pete,
This is going to need a bit more digging than I can do from where I sit. I'm going to ask a colleague of mine to come in and lend you a hand. Thanks for your patience!
Keri
-
Thanks Keri here is our site http://www.health2000.co.nz we have recently asked our IT company to make amendments to offset the duplicates page issue.
The attached graphic shows the problem was in recline but now it seems to have come back. Any idea why that might be? I would have thought 301 redirecting would be an all or nothing solution. Also Ive asked our IT company and they have said this may take a while for google to indexes our page. If that is correct , how long do you think will take?
I’ve set up campaigns for our organisation and four of our competitors and note that on average we have had 6500 pages crawled where as our competitors have over 11000 pages crawled is there any reason why that might be? Thanks again for your help!
Pete
-
If you give him a 301 redirect it should help him, and the search engines, which is the most important part.
If you can touch base with your IT team and see if they changed something and ask them to change it back, that'd be a good place to start. If you can share your URL here, we can look at it and help direct you to the easiest way to fix things (if it is the www and non-www problem), or help identify the source of the problem.
-
Hi Keri,
Thanks for drooping me a line...... How do we make that cheeky little robot unfind one of them? : )
Cheers
Pete
-
Hmmm...my first thought is that if it's sudden duplicate content and doubling of errors is that perhaps Roger found both the www and non-www versions of the site?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page Authority 2.0 is coming soon!
UPDATE: We’ve made the tough decision to delay the launch of Page Authority 2.0 as our team investigates an unforeseen issue. **To learn more about the rigorous testing process and updates to the timeline, **check out this blog post. Hey there Moz family, We’re stoked to let you know about an upcoming change to a beloved metric — similar to our refresh of the Domain Authority algorithm last year, we’ve been working on developing an improvement to its sibling metric, Page Authority (also known as “PA”). Page Authority (PA) identifies the strength of a particular page (on a 1-100 scale) and its ability to rank in search results in comparison to other pages. PA is a Moz metric, and while it can be used as a good gauge of page strength and ranking potential, it is not used by search engines to determine ranking outcome. On September 30, 2020, we will release the new and improved Page Authority algorithm that will be updated in a similar fashion to last year’s update to DA. The updated algorithm will take into account more modern ranking factors to make the score even more accurate and predictive. We recognize that the update to the DA algorithm took time to communicate to clients and stakeholders, and we wanted to be sure to give you advance notice again this time around. We’ve created a number of resources to help you understand the what, the why, and the how of this update to PA. Let’s start with a few FAQs that you might be curious about! Why didn’t PA update when DA updated? Although many folks associate DA and PA with one another, the two metrics are calculated independently. We chose to update the two metrics separately in order to take the care that each metric deserved, and to provide the highest quality algorithm updates for the SEO community. Why is Moz changing the PA algorithm? As with our update to the DA algorithm, we want to ensure that you have confidence in our metrics and the predictions that they provide. Data integrity is an integral part of our tools and something that we hold in the highest regard. To be sure that PA can best reflect the potential for a page to rank on the SERP, we’re making the necessary improvements. What can I expect to see from the PA algorithm update? Many pages will see changes to their PA scores as a result of this algorithm update. While the changes to scores may be somewhat minimal, there is a possibility that some pages will see material change to their scores. The new PA algorithm takes into consideration Spam Score and link patterns, in addition to dozens of other factors, so your PA scores may see noticeable change if your pages have spammy or unnatural link patterns. How can I prepare for the update? As with any change to a metric that you know and love, we recommend getting in touch with your stakeholders to let them know of the upcoming update. For those who are used to seeing this metric in your SEO reports, giving them a heads-up will help them to prepare for any fluctuations they might see to PA scores once the new PA algorithm rolls out. We also recommend using this update as an opportunity to educate them on the use of Page Authority and how you might use this refreshed metric for future link building projects. Our resource center has a few helpful pieces of collateral that can support these conversations with stakeholders and clients. Is Page Authority an absolute score or a relative one? Page Authority should always be used as a relative metric, to compare the score of your pages to the scores of other sites’ pages. Link Explorer looks at over 7 trillion pages and 40 trillion links to inform the Page Authority metric that you see. As such, it is always a wise idea to use PA as a comparative score to understand where your page stacks up in comparison to the other pages that are present on the SERPs you care about. Will Domain Authority (DA) be impacted by this update? No, DA will not be affected by this update. This particular algorithm update is specific to Page Authority only. Will API users be affected at the same time? Yes, API users will see the update to Page Authority at the same time as users of Moz Pro. We’d love for you to check out our resource page for links to a slide deck, a whitepaper, and other helpful information. The full announcement blog post from Russ Jones can be found here. Happy to chat with you here in the Q&A thread, or feel free to send an email to help@moz.com with any questions. Best, Igor
API | | IgorJesovnik8 -
Navigation pages with a PA 1
Ok, you guys probably think this is a new website and i should just wait, but this is not the case.
API | | Forresult
We have 2 websites (old) websites with a DA of 34 and a DA of 19 and high PA values on the mainpage. Our problem: All the other pages stay at a page authority of 1. One website is build in Magento and one in Wordpress. Both websites have deeplinks, footerlinks en in-contentlinks. The other pages don't get any linkjuice according to Moz. We don't use any robot noindex,nofollow or nofollow links and the menu structure isn't the problem. Is anyone familiare with the problem? I know is shouldn't be concerned about PA/DA, but i just can't explain what's going on.0 -
Error Code 803
Hello all, With every new Moz report that comes in, an error code 803 appears. I check each link that comes up as an 803 but they work perfectly. What could be causing this to happen? Error Code 803: Incomplete HTTP Response Received Your page did not send a complete response to the Moz Crawler's request. Our crawler is pretty standard, so it's likely that other browsers and crawlers may also encounter this error. If you have this error on your homepage, it prevents the Moz crawler (and some search engines) from crawling the rest of your site.
API | | moon-boots0 -
Moz Crawl: Can't check page optimization error https
Help needed, when I try to do a page optimization check i get the following error : The URL you entered does not appear to be returning a page successfully. Please make sure that you've entered the URL of valid, working page. But i can do a site crawl, what should be the problem? Checked with frog seo spider and add no problem, robots.txt its also clean. Anyone knows what can be wrong? Thanks
API | | Luis-Pereira0 -
803 Crawl attempt error
Hello I'd be very grateful for any advice with this: My insights show I have an 803 error. Now, under "pages with crawl attempt error" the page in question is just an uploaded image to wordpress. However, above the graph it says: "We were unable to access your homepage, which prevented us from crawling the rest of your site. It is likely that other browsers as well as search engines may encounter this problem and abort their sessions." Does this really mean my homepage? or is the only issue with the image? I have noticed for the past 8 weeks I'm getting 1 crawl attempt error every 2 weeks (so when viewed weekly I have 1 error one week, 0 error the next week etc) Is this normal? Since receiving this 803 error, I have significantly dropped in SERPS for 3 key terms I was on page 1 for (now dropped to pages 3-4). Could this be related? I realise this is a bit specific, but thanks in advance. Cheers 🙂
API | | wearehappymedia0 -
3 result limit to Top Pages API call
I am using the MOZ API to make calls for the top pages for a particular URL. However, when I pass in any limit value greater than 3 the API only returns 3 results. I have even tried to put in URLs like 'www.moz.com' and still only 3 results. Sample call to the API below: http://lsapi.seomoz.com/linkscape/top-pages/www.moz.com?AccessID=member-xxxxxxxxx&Expires=1419020831&Signature=xxxxxxxxx&Cols=2052&Offset=0&Limit=50
API | | solodev0 -
API - Internal Links to page and related metrics
Hi dear moz Team! Currently I´m building a Java application accessing your API. But there are some metrics I urgently need which I can´t get out of the API until now: The total number of internal links to a page The total number of internal links to a page with partial anchor text match MozRank passed by all internal links w. part. match anchor text (would be nice) For example, if I try this by your links endpoint, my idea was: http://lsapi.seomoz.com/linkscape/links/http%3A%2F%2Fwww.jetztspielen.de%2F?AccessID=..
API | | pollierer
&Expires=..
&Signature=..
&Scope=domain_to_page
&Filter=internal
&Sort=domain_authority
&SourceCols=4 (or any other value)
&SourceDomain=www.jetztspielen.de
&Offset=0
&Limit=50 If I try this, the API says: {"status": "400", "error_message": "Cannot set a source domain when filtering for internal links."} Is there any way to get the data I need by your API endpoints? I´m currently writing my master thesis and it is very important to me to solve this somehow. Thank you very much in advance! Best, Andreas Pollierer1 -
Top Pages metrics in OSE
Not sure if this is an API question or a feature request, but wondering if other folks had a way to do this: In OSE there is the dashboard for a specific URL that is entered into the search bar, giving you metrics at a glance on it. But I often find myself going to the Top Pages tab to get a sense of the domain as a whole. First off, wouldn't it be nice/is there a way to build my own "dashboard" based on info from that section? Specifically, I'd love to see at a glance the number of "top pages" that exist (many websites are well under the 10,000 page limit for this section, but there's no quick glance metric showing that) One thing that would be very handy for me would be a breakdown of HTTP Status info across the whole domain, as being able to see the raw total of different statuses (and the percentages of each based on the total number of pages) would be really helpful, giving me a sense if I should dig into any issues before exporting the list to CSV. I've found myself needing this type of info for multiple domains at once, so what would be REALLY cool would be a Google Doc where I could paste in different domains in one column, and this info being returned in other columns. I've searched through the Q&A and didn't find anything like this, and I didn't know how easy/hard any of this would be to do, but I was wondering if anyone else had a sense of how to solve this problem and how feasible it would be to tackle it. Thanks!
API | | John-E-Turner0