Sitemap Best Practices
-
My question is regarding the URL structure best practices of a sitemap. My website allows search any number of ways, i.e.
1. http://www.website.com/category/subcategory/product
2. http://www.website.com/subcategory/product
3. http://www.website.com/product
However, I am not sure which structure to use in the sitemap (which is being written manually). I know that for SEO purposes the 3rd option is best as the link is more relevant to that individual product, but the Moz tool states that the home page should have less than 100 links (although Google doesn't penalise for having more) and by writing my entire site in the 3rd way it would result in a lot more links adjoining to the home page.
It is either the 2nd or 3rd option, I think, as the 1st category is not keyword specific (rather a generic term, i.e. novelties).
Does anyone have experience with this?
-
Happy to help!
-
Thanks Logan!
-
Google is less concerned about the actual structure of your URLs, and more concerned that you pick a horse and ride it, which you've done by canonicalizing 2 variations to the third. In your example, the third URL is perfectly fine, since it will always remain constant. The other 2 can change depending on how someone navigates to that product. I'd keep it the way you have it.
-
Hello Logan, thanks for responding, although you've not responded to my actual question as such.
Yes, currently I am canonicalising links 1 and 2 toward link 3, but my question wasn't regarding which URL to use in the sitemap, but rather what Googles preferred URL structure was.
Does Google dislike the link 3 structure because it makes it links every product and category directly to the home page? It would appear that the Moz tool seems to think so (although they state that you're not penalised for it).
In your experience, what is Googles preferred URL structure, link 1, 2 or 3? I can easily change the Canonical tag to either of the three, that isn't an issue.
-
Hi,
In your example, you have 3 URLs that render the same content it sounds like. If this is the case, I would assume you're canonicalizing 2 versions to the third. In this situation, you'd want to use the canonical version in your XML sitemaps. You don't want to point search engines to URLs in a XML sitemap then have them go elsewhere when they find the canonical tag.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Large site with content silo's - best practice for deep indexing silo content
Thanks in advance for any advice/links/discussion. This honestly might be a scenario where we need to do some A/B testing. We have a massive (5 Million) content silo that is the basis for our long tail search strategy. Organic search traffic hits our individual "product" pages and we've divided our silo with a parent category & then secondarily with a field (so we can cross link to other content silo's using the same parent/field categorizations). We don't anticipate, nor expect to have top level category pages receive organic traffic - most people are searching for the individual/specific product (long tail). We're not trying to rank or get traffic for searches of all products in "category X" and others are competing and spending a lot in that area (head). The intent/purpose of the site structure/taxonomy is to more easily enable bots/crawlers to get deeper into our content silos. We've built the page for humans, but included link structure/taxonomy to assist crawlers. So here's my question on best practices. How to handle categories with 1,000+ pages/pagination. With our most popular product categories, there might be 100,000's products in one category. My top level hub page for a category looks like www.mysite/categoryA and the page build is showing 50 products and then pagination from 1-1000+. Currently we're using rel=next for pagination and for pages like www.mysite/categoryA?page=6 we make it reference itself as canonical (not the first/top page www.mysite/categoryA). Our goal is deep crawl/indexation of our silo. I use ScreamingFrog and SEOMoz campaign crawl to sample (site takes a week+ to fully crawl) and with each of these tools it "looks" like crawlers have gotten a bit "bogged down" with large categories with tons of pagination. For example rather than crawl multiple categories or fields to get to multiple product pages, some bots will hit all 1,000 (rel=next) pages of a single category. I don't want to waste crawl budget going through 1,000 pages of a single category, versus discovering/crawling more categories. I can't seem to find a consensus as to how to approach the issue. I can't have a page that lists "all" - there's just too much, so we're going to need pagination. I'm not worried about category pagination pages cannibalizing traffic as I don't expect any (should I make pages 2-1,000) noindex and canonically reference the main/first page in the category?). Should I worry about crawlers going deep in pagination among 1 category versus getting to more top level categories? Thanks!
Moz Pro | | DrewProZ1 -
What's the best way to eliminate "429 : Received HTTP status 429" errors?
My company website is built on WordPress. It receives very few crawl errors, but it do regularly receive a few (typically 1-2 per crawl) "429 : Received HTTP status 429" errors through Moz. Based on my research, my understand is that my server is essentially telling Moz to cool it with the requests. That means it could be doing the same for search engines' bots and even visitors, right? This creates two questions for me, which I would greatly appreciate your help with: Are "429 : Received HTTP status 429" errors harmful for my SEO? I imagine the answer is "yes" because Moz flags them as high priority issues in my crawl report. What can I do to eliminate "429 : Received HTTP status 429" errors? Any insight you can offer is greatly appreciated! Thanks,
Moz Pro | | ryanjcormier
Ryan0 -
How to make a best SEO strategy
Hi I want to make a seo strategy for website. what things I need and how to make is there any tool?
Moz Pro | | Kabooter0 -
What is the best method to solve duplicate page content?
The issue I am having is an overwhelmingly large number of pages on cafecartel.com show that they have duplicate page content. But when I check the errors on SEOmoz it shows that the duplicate content is from www.cafecartel.com not cafecartel.com. So first of all, does this mean that there are two sites? and is this a problem I can fix easily? (i.e. redirecting the URL and deleting the extra pages) Is this going to make all other SEO useless due to the fact that it shows that nearly every page has duplicate page content? Or am I just completely reading the data wrong?
Moz Pro | | MarkP_0 -
SEO Web Crawler - Referrer Lists XML Sitemap URL
Hello!, I recently ran the crawl tool on a client site. Opening up the file, I noticed that the referring URLs listed are my XML sitemaps and not (X)HTML pages. Any reason or thoughts behind why this is happening? Thanks!
Moz Pro | | MorpheusMedia0 -
What's been the best SERP rank tracker for you?
I do a lot of my SERP ranking tracking via Aaron Wall's Rank Checker, but I'm realizing it's taking up too much time. I was looking at Raven's tools and a few others. What's been your favorite thus far? If you do reply, can you talk about: price capabilities time savings b/c of it graphical UI other thoughts I like SEOMoz's UI, but I can only manage 5 domains and I don't have a huge budget yet. Also, if there's a rank checker out there that incorporates google places results, that would be great too! Maybe too good, eh?
Moz Pro | | SeattleOrganicSEO0 -
What's the best practice for tracking broad match traffic?
We use the Pro Web App for our keyword/ranking reporting. First, am I correct in assuming that this uses exact match to report the traffic? That being the case, is there a way in Moz (beyond entering endless permutations of our targeted keywords) to track traffic we're getting out of broad match searches? For example, say we optimized for and are tracking the keyword "lambada dancing llamas" for a company called Llarry's Llamas. Let's say that Moz reports little to no traffic for that keyword, but GoogleAnalytics indicates that we've gotten traffic from "larrys dancing llamas", "dancing llamas", "llarry llamas" and so on. So, obviously we're getting broad match traffic out of the lambada llamas, just not a lot of exact match, which is all that the Moz app shows. How do other people track this kind of traffic?
Moz Pro | | MackenzieFogelson0 -
Where is the best SEO Glossary of Terms?
Where do you feel the best existing Glossary of SEO / SEM Terms explained currently resides. (and is there a better way to ask this question?) 🙂
Moz Pro | | iansears0