Sitemaps and Indexed Pages
-
Hi guys,
I created an XML sitemap and submitted it for my client last month.
Now the developer of the site has also been messing around with a few things.
I've noticed on my Moz site crawl that indexed pages have dropped significantly.
Before I put my foot in it, I need to figure out if submitting the sitemap has caused this.. can a sitemap reduce the pages indexed?
Thanks
David.
-
Thanks Eli!
I guess I was wondering if the MOZ Bot only followed pages that were in the sitemap. It was generated by Screaming Frog I have trusted it to include all relevant pages!
I have put in a more detailed description in the response below. Overall I need to investigate further but i'm satisfied that the sitemap has not caused the drop!
-
Thanks Martijn!
I guess I was wondering if the MOZ Bot only followed pages that were in the sitemap. It was generated by Screaming Frog I have trusted it to include all relevant pages!
To elaborate.
There were about 80,000 pages and I used canonical, no index, and redirects to clean up a rather large mess of filter URL's and dup content.
That dropped the pages to about 14k. Then I submitted the sitemap last month and now the crawl only found 4k pages.
Further investigation is needed on my behalf but I wanted to double check that this sudden drop was not because of a sitemap! Thanks for clarifying that!
-
Hi David,
Messing up, Changing or Updating, Deleting a Sitemap is not necessarily something that will decrease the number of ranked or crawled pages. It usually is used a signal to find new pages and figure out if old ones are deleted. But the chances that your sitemap have had a significant impact in what kind of pages went down is something I would find unlikely. It could happen though that you'd see the opposite, an increase in pages indexed/submitted/crawled after you submit a sitemap.
Martijn.
-
Hey David!
Thanks for reaching out to us!
Unfortunately I am not an SEO consultant / Web Developer so I cannot offer specific advice, but I'm sure there are loads of members here who would love to help and have a lot more knowledge than I do! A few things I have picked up which may help are the following:
Try to determine when the drop started, did it drop when you submitted the XML sitemap or when the developer changed certain things? This could help point to the reason for the drop in indexing. There are a variety of reasons as to why Google may not choose to index pages, however some of the common ones are:
-
Check your robots.txt to ensure those pages are still crawlable
-
Check for duplicate content / was there any canonical changes?
-
One of the tools you could use to help keep track of ranking fluctuations is mozcast (http://mozcast.com/). Was there turbulence in the Google algorithm when the indexed pages dropped significantly?
If you want us to have a look at your specific campaign to investigate further could you please pop an email over to help@moz.com.
Thanks!
Eli
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page Authority 2.0 is coming soon!
UPDATE: We’ve made the tough decision to delay the launch of Page Authority 2.0 as our team investigates an unforeseen issue. **To learn more about the rigorous testing process and updates to the timeline, **check out this blog post. Hey there Moz family, We’re stoked to let you know about an upcoming change to a beloved metric — similar to our refresh of the Domain Authority algorithm last year, we’ve been working on developing an improvement to its sibling metric, Page Authority (also known as “PA”). Page Authority (PA) identifies the strength of a particular page (on a 1-100 scale) and its ability to rank in search results in comparison to other pages. PA is a Moz metric, and while it can be used as a good gauge of page strength and ranking potential, it is not used by search engines to determine ranking outcome. On September 30, 2020, we will release the new and improved Page Authority algorithm that will be updated in a similar fashion to last year’s update to DA. The updated algorithm will take into account more modern ranking factors to make the score even more accurate and predictive. We recognize that the update to the DA algorithm took time to communicate to clients and stakeholders, and we wanted to be sure to give you advance notice again this time around. We’ve created a number of resources to help you understand the what, the why, and the how of this update to PA. Let’s start with a few FAQs that you might be curious about! Why didn’t PA update when DA updated? Although many folks associate DA and PA with one another, the two metrics are calculated independently. We chose to update the two metrics separately in order to take the care that each metric deserved, and to provide the highest quality algorithm updates for the SEO community. Why is Moz changing the PA algorithm? As with our update to the DA algorithm, we want to ensure that you have confidence in our metrics and the predictions that they provide. Data integrity is an integral part of our tools and something that we hold in the highest regard. To be sure that PA can best reflect the potential for a page to rank on the SERP, we’re making the necessary improvements. What can I expect to see from the PA algorithm update? Many pages will see changes to their PA scores as a result of this algorithm update. While the changes to scores may be somewhat minimal, there is a possibility that some pages will see material change to their scores. The new PA algorithm takes into consideration Spam Score and link patterns, in addition to dozens of other factors, so your PA scores may see noticeable change if your pages have spammy or unnatural link patterns. How can I prepare for the update? As with any change to a metric that you know and love, we recommend getting in touch with your stakeholders to let them know of the upcoming update. For those who are used to seeing this metric in your SEO reports, giving them a heads-up will help them to prepare for any fluctuations they might see to PA scores once the new PA algorithm rolls out. We also recommend using this update as an opportunity to educate them on the use of Page Authority and how you might use this refreshed metric for future link building projects. Our resource center has a few helpful pieces of collateral that can support these conversations with stakeholders and clients. Is Page Authority an absolute score or a relative one? Page Authority should always be used as a relative metric, to compare the score of your pages to the scores of other sites’ pages. Link Explorer looks at over 7 trillion pages and 40 trillion links to inform the Page Authority metric that you see. As such, it is always a wise idea to use PA as a comparative score to understand where your page stacks up in comparison to the other pages that are present on the SERPs you care about. Will Domain Authority (DA) be impacted by this update? No, DA will not be affected by this update. This particular algorithm update is specific to Page Authority only. Will API users be affected at the same time? Yes, API users will see the update to Page Authority at the same time as users of Moz Pro. We’d love for you to check out our resource page for links to a slide deck, a whitepaper, and other helpful information. The full announcement blog post from Russ Jones can be found here. Happy to chat with you here in the Q&A thread, or feel free to send an email to help@moz.com with any questions. Best, Igor
API | | IgorJesovnik8 -
Navigation pages with a PA 1
Ok, you guys probably think this is a new website and i should just wait, but this is not the case.
API | | Forresult
We have 2 websites (old) websites with a DA of 34 and a DA of 19 and high PA values on the mainpage. Our problem: All the other pages stay at a page authority of 1. One website is build in Magento and one in Wordpress. Both websites have deeplinks, footerlinks en in-contentlinks. The other pages don't get any linkjuice according to Moz. We don't use any robot noindex,nofollow or nofollow links and the menu structure isn't the problem. Is anyone familiare with the problem? I know is shouldn't be concerned about PA/DA, but i just can't explain what's going on.0 -
Mozscape Index update frequency problems?
I'm new to Moz, only a member for a couple months now. But I already rely heavily on the mozscape index data for link building, as I'm sure many people do. I've been waiting for the latest update (due today after delay), but am not seeing any mention of the data yet - does it normally get added later in the day? I'm not that impatient that I can't wait until later today or tomorrow for this index update, but what I am curious about is whether Moz is struggling to keep up, and if updates will continue to get more and more rare? For example, in 2013 I count 28 index updates. In 2014 that number dropped to 14 updates (50% drop). In 2015, there was only 8 (another 43% drop), and so far this year (until the March 2nd update is posted) there has only been 1. This isn't just a complaint about updates, I'm hoping to get input from some of the more experienced Moz customers to better understand (with the exception of the catastrophic drive failure) the challenges that Moz is facing and what the future may hold for update frequency.
API | | kevin.kembel1 -
10/14 Mozscape Index Update Details
Howdy gang, As you might have seen, we've finally been able to update the Mozscape index after many challenging technical problems in the last 40 days. However, this index has some unique qualities (most of them not ideal) that I should describe. First, this index still contains data crawled up to 100 days ago. We try to make sure that what we've crawled recently is stuff that we believe has been updated/changed, but there may be sites and pages that have changed significantly in that period that we didn't update (due to issues I've described here previously with our crawlers & schedulers). Second, many PA/DA and other metric scores will look very similar to the last index because we lost and had problems with some metrics in processing (and believe that much of what we calculated may have been erroneous). We're using metrics from the prior index (which had good correlations with Google, etc) until we can feel confident that the new ones we're calculating are correct. That should be finished by the next index, which, also, should be out much faster than this one (more on that below). Long story short on this one - if your link counts went up and you're seeing much better/new links pointing to you, but DA/PA remain unchanged, don't panic - that's due to problems on our end with calculations and will be remedied in the next index. Third - the good news is that we've found and fixed a vast array of issues (many of them hiding behind false problems we thought we had), and we now believe we'll be able to ship the next index with greater quality, greater speed, and better coverage. One thing we're now doing is taking every URL we've ever seen in Google's SERPs (via all our rank tracking, SERPscape, the corpus for the upcoming KW Explorer product, etc) and prioritizing them in Mozscape's crawl, so we expect to be matching what Google sees a bit more closely in future indices. My apologies for the delay in getting this post up - I was on a plane to London for Searchlove - should have got it up before I left.
API | | randfish4 -
Lost many links and keyword ranks since moz index update
Hi All, I came back from work today from a week off to find my site has gone from 681 external inbound links to 202. With this my domain authority, moz trust and moz rank have all also taken a slip. Compounding this, I am seeing a slip most of my keywords rankings. If i try to use the open site explorer to explore my links and see what going on i get the message It looks like we haven't discovered link data for this site or URL. If i check the just discovered links like it suggests I get It looks like there's no Just-Discovered Links data for this URL yet. I know these features worked before the index as i used them. Is this all attributable to the moz index issues that have been noted or could something have happened to my site? Since i started 2 months ago I have made many changes including... Updating the site map that was 4 years out of date and included 400 broken urls Removed blank pages and other useless webpages on the site that contained no content (from the previous administrator) Edited a few pages content from keyword spammy stuff to nicely written and relevant content Fixed url rewrites that made loops and un-accessible product pages All these changes should be for the better but the latest readings have me a little worried. Thanks.
API | | ATP0 -
Does Moz's crawlers use _escaped_fragment_ to inspect pages on a single-page application?
I just got started, but got a 902 error code on some pages, with a message saying there might be an outage on my site. That's certainly not the case, so I'm wondering if the crawlers actually respect and use the escaped_fragment query parameter. Thanks, David.
API | | CareerDean0 -
On page reports
Hi everyone I have just been going through the online page I see that I have quite a few words that have an F rating I was wondering if I have all the keywords with an A ranking would that improve our Moz rating? Also of the below elements can anyone tell which, if any are more important than the others? Title URL Meta Desc H1 H2-4 Body B / Strong IMG ALT
API | | Hardley1110 -
Top Pages metrics in OSE
Not sure if this is an API question or a feature request, but wondering if other folks had a way to do this: In OSE there is the dashboard for a specific URL that is entered into the search bar, giving you metrics at a glance on it. But I often find myself going to the Top Pages tab to get a sense of the domain as a whole. First off, wouldn't it be nice/is there a way to build my own "dashboard" based on info from that section? Specifically, I'd love to see at a glance the number of "top pages" that exist (many websites are well under the 10,000 page limit for this section, but there's no quick glance metric showing that) One thing that would be very handy for me would be a breakdown of HTTP Status info across the whole domain, as being able to see the raw total of different statuses (and the percentages of each based on the total number of pages) would be really helpful, giving me a sense if I should dig into any issues before exporting the list to CSV. I've found myself needing this type of info for multiple domains at once, so what would be REALLY cool would be a Google Doc where I could paste in different domains in one column, and this info being returned in other columns. I've searched through the Q&A and didn't find anything like this, and I didn't know how easy/hard any of this would be to do, but I was wondering if anyone else had a sense of how to solve this problem and how feasible it would be to tackle it. Thanks!
API | | John-E-Turner0