Overly-Dynamic Urls how to fix in SEOMOZ?
-
Hello.
I have about 300 warnings of overly-dynamic urls.
In urls like this:
http://www.theprinterdepo.com/clearance?dir=asc&order=price&p=10
As you can see all parameters are needed, and my ecommerce solution generates them automatically.
How can I get rid of these warnings? I suppose that by using robots.txt, but I have no idea about it.
In my google webmaster tools I have already configured that these parameteres the crawler should not index them.
Check the image here:
-
Hi Kami,
You might want to ask this in a new post. New posts on old threads don't bump a post in the Q&A forums, so not too many people will see this post (I only knew about it as I was notified from having subscribed last year).
-
Hi, We sort of have the same issue, We have over 5000 pages with the same issues actually. Our ecommerce site uses several different filter (Using Ajax) and we have many different urls like,
http://www.dellamoda.com/Designer-Pumps.html?sort=price&sort_direction=1&use_selected_filter=Y
http://www.dellamoda.com/Designer-Accessories.html?sort=title&use_selected_filter=Y&view=all
http://www.dellamoda.com/Designer-Accessories.html?sort=title&sort_direction=1&use_selected_filter=Y
Could we use the robots.txt file to disallow these as well? and do we need to put the whole url in there?
like:
Disallow: /*?sort=price&sort_direction=1&use_selected_filter=Y
if not how far into the url should be disallowed?
Any help would be greatly appreciated.
Thank you, Tony
-
John, having a ton of URLs indexed for the same page will actually dilute things, not help your rankings. Dr. Pete wrote a great post at http://www.seomoz.org/blog/duplicate-content-in-a-post-panda-world explaining duplicate content that should help give you a better understanding of things.
-
Hello there,
I have encounter a similar problem in a similar scenario. If we do not allow this pages to be crawled, wouldn't it reduce the number of pages indexed resulting in lower Google ranking?
-
Hi! Did Rasmus answer your question, or are you looking for some more help?
-
Disallow: /*?dir=desc
Disallow: /*&order=
I think yout should try with these lines, and test in Google Webmaster tools. This should leave only this page
https://www.theprinterdepo.com/clearance
That is what you want right?
-
Can I know the syntax for robots.txt to ignore those?
-
Sure thing!
For example:
<a <span="">href</a><a <span="">="</a>https://www.theprinterdepo.com/clearance?dir=desc&order=price" title="Set Descending Direction">src="https://www.theprinterdepo.com/skin/frontend/default/MAG060062/images/i_asc_arrow.gif" alt="Set Descending Direction" class="v-middle" />
This code is the code for reversing the result. So you can check to see if the page has a url query with "dir=asc" (the standard). If so the code above should instead be:
<a rel="nofollow" <span="">href</a><a rel="nofollow" <span="">="</a>https://www.theprinterdepo.com/clearance?dir=desc&order=price" title="Set Descending Direction">src="https://www.theprinterdepo.com/skin/frontend/default/MAG060062/images/i_asc_arrow.gif" alt="Set Descending Direction" class="v-middle" />However, I believe the best approach will be to change the meta tag for robots for the page.
if the url query is dir=asc, order=price then robots="index, follow". If dir is not asc, and order is not price then robots="noindex, follow".
-
I am a really newbie to SEO, can you please explain me how to do it?
-
Can you either set rel="nofollow" on the links on the page that changes the sorting, such that moz and google do not check these pages? Or you can set the robots="noindex, follow" on pages which are not the "standard" sort?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
URL has caps, but canonical does not. Now what?
Hi, Just started working with a site that has the occasional url with a capital, but then the url in the canonical as lower case. Neither, when entered in a browser, resolves to the other. It's a Shopify site. What do you think I should do?
Technical SEO | | 945010 -
Language Specific Characters in URLs for
Hi People, would really appreciate your advice as we are debating best practice and advice seems very subjective depending if we are talking to our dev or SEO team. We are developing a website aimed at the South American market with content entirely in Spanish. This is our first international site so our experience is limited. Should we be using Spanish characters (such as www.xyz.com/contáctanos) in URLs or should we use ASCII character replacements? What are the pros and cons for SEO and usability? Would really be great to get advice from the Moz community and make me look good at the same time as it was my suggestion 🙂 Nick
Technical SEO | | nickspiteri0 -
Special characters in URL
Will registered trademark symbol within a URL be bad? I know some special characters are unsafe (#, >, etc.) but can not find anything that mentions registered trademark. Thanks!
Technical SEO | | bonnierSEO0 -
How important is keyword usage in the URL?
Hi,
Technical SEO | | Whebb
We have a client who has engaged us recently for some SEO work and most of their website looks pretty good seo-wise already. Many of their site pages rank at the top or middle of page two for their targeted keywords. In many cases they are not using the targeted keyword in the URL and most pages could use some additional on-page clean up. My question is, is it worth it to re-write the URLs to include the targeted keyword and then do 301 redirects to send the old pages to the new ones in order to improve the ranking? Or should we just do the minor on page work in hopes that this will be enough to improve the rankings and push them on to the first page. Thanks.0 -
Changing all urls
A client of mine has a wordpress website that is installed in a directory, called "site". So when you go to www.domain.com you are redirected to www.domain.com/site. We all know how bad it is to have a redirect fron your subdomain to another page. In this case I measured a loss of 5 points of page authority. The question is: what is the best practice to remove the "site" from the address and changing all the urls? Should I use the webmaster tool to tell to Google that the site is moving? It's not 100% true, cause the site is just moving one level up. Should I install a copy of the website under www.domain.com and just redirect 301 every old page to its new url? This way I think the site would be deindexet for 2/3 months. Any suggestions or tips welcome! Thanks DoMiSol
Technical SEO | | DoMiSoL0 -
Best Practices for adding Dynamic URL's to XML Sitemap
Hi Guys, I'm working on an ecommerce website with all the product pages using dynamic URL's (we also have a few static pages but there is no issue with them). The products are updated on the site every couple of hours (because we sell out or the special offer expires) and as a result I keep seeing heaps of 404 errors in Google Webmaster tools and am trying to avoid this (if possible). I have already created an XML sitemap for the static pages and am now looking at incorporating the dynamic product pages but am not sure what is the best approach. The URL structure for the products are as follows: http://www.xyz.com/products/product1-is-really-cool
Technical SEO | | seekjobs
http://www.xyz.com/products/product2-is-even-cooler
http://www.xyz.com/products/product3-is-the-coolest Here are 2 approaches I was considering: 1. To just include the dynamic product URLS within the same sitemap as the static URLs using just the following http://www.xyz.com/products/ - This is so spiders have access to the folder the products are in and I don't have to create an automated sitemap for all product OR 2. Create a separate automated sitemap that updates when ever a product is updated and include the change frequency to be hourly - This is so spiders always have as close to be up to date sitemap when they crawl the sitemap I look forward to hearing your thoughts, opinions, suggestions and/or previous experiences with this. Thanks heaps, LW0 -
While SEOMoz currently can tell us the number of linking c-blocks, can SEOMoz tell us what the specific c-blocks are?
I know it is important to have a diverse set of c-blocks, but I don't know how it is possible to have a diverse set if I can't find out what the c-blocks are in the first place. Also, is there a standard for domain linking c-blocks? For instance, I'm not sure if a certain amount is considered "average" or "above-average."
Technical SEO | | Todd_Kendrick0 -
Do keywords in url parameter count?
I have a client who is on an older ecommerce platform that does not allow url rewrites in anyway. It would cost a ton of money to custom dev a solution. Anyways right now they have set up a parameter on their product urls to at least get the keyword in there. My question is, will this keyword actually be counted since it is in a parameter? An example url is http://domain.com/Catalog.aspx?Level1=01&Level2=02&C=Product-name-here Does this 'product-name-here' count as having the keyword in the url according to google?
Technical SEO | | webfeatseo0