Skip to content
    Moz logo Menu open Menu close
    • Products
      • Moz Pro
      • Moz Pro Home
      • Moz Local
      • Moz Local Home
      • STAT
      • Moz API
      • Moz API Home
      • Compare SEO Products
      • Moz Data
    • Free SEO Tools
      • Domain Analysis
      • Keyword Explorer
      • Link Explorer
      • Competitive Research
      • MozBar
      • More Free SEO Tools
    • Learn SEO
      • Beginner's Guide to SEO
      • SEO Learning Center
      • Moz Academy
      • MozCon
      • Webinars, Whitepapers, & Guides
    • Blog
    • Why Moz
      • Digital Marketers
      • Agency Solutions
      • Enterprise Solutions
      • Small Business Solutions
      • The Moz Story
      • New Releases
    • Log in
    • Log out
    • Products
      • Moz Pro

        Your all-in-one suite of SEO essentials.

      • Moz Local

        Raise your local SEO visibility with complete local SEO management.

      • STAT

        SERP tracking and analytics for enterprise SEO experts.

      • Moz API

        Power your SEO with our index of over 44 trillion links.

      • Compare SEO Products

        See which Moz SEO solution best meets your business needs.

      • Moz Data

        Power your SEO strategy & AI models with custom data solutions.

      Let your business shine with Listings AI
      Moz Local

      Let your business shine with Listings AI

      Learn more
    • Free SEO Tools
      • Domain Analysis

        Get top competitive SEO metrics like DA, top pages and more.

      • Keyword Explorer

        Find traffic-driving keywords with our 1.25 billion+ keyword index.

      • Link Explorer

        Explore over 40 trillion links for powerful backlink data.

      • Competitive Research

        Uncover valuable insights on your organic search competitors.

      • MozBar

        See top SEO metrics for free as you browse the web.

      • More Free SEO Tools

        Explore all the free SEO tools Moz has to offer.

      NEW Keyword Suggestions by Topic
      Moz Pro

      NEW Keyword Suggestions by Topic

      Learn more
    • Learn SEO
      • Beginner's Guide to SEO

        The #1 most popular introduction to SEO, trusted by millions.

      • SEO Learning Center

        Broaden your knowledge with SEO resources for all skill levels.

      • On-Demand Webinars

        Learn modern SEO best practices from industry experts.

      • How-To Guides

        Step-by-step guides to search success from the authority on SEO.

      • Moz Academy

        Upskill and get certified with on-demand courses & certifications.

      • MozCon

        Save on Early Bird tickets and join us in London or New York City

      Unlock flexible pricing & new endpoints
      Moz API

      Unlock flexible pricing & new endpoints

      Find your plan
    • Blog
    • Why Moz
      • Digital Marketers

        Simplify SEO tasks to save time and grow your traffic.

      • Small Business Solutions

        Uncover insights to make smarter marketing decisions in less time.

      • Agency Solutions

        Earn & keep valuable clients with unparalleled data & insights.

      • Enterprise Solutions

        Gain a competitive edge in the ever-changing world of search.

      • The Moz Story

        Moz was the first & remains the most trusted SEO company.

      • New Releases

        Get the scoop on the latest and greatest from Moz.

      Surface actionable competitive intel
      New Feature

      Surface actionable competitive intel

      Learn More
    • Log in
      • Moz Pro
      • Moz Local
      • Moz Local Dashboard
      • Moz API
      • Moz API Dashboard
      • Moz Academy
    • Avatar
      • Moz Home
      • Notifications
      • Account & Billing
      • Manage Users
      • Community Profile
      • My Q&A
      • My Videos
      • Log Out

    The Moz Q&A Forum

    • Forum
    • Questions
    • Users
    • Ask the Community

    Welcome to the Q&A Forum

    Browse the forum for helpful insights and fresh discussions about all things SEO.

    1. Home
    2. SEO Tactics
    3. Intermediate & Advanced SEO
    4. Pagination parameters and canonical

    Moz Q&A is closed.

    After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.

    Pagination parameters and canonical

    Intermediate & Advanced SEO
    5
    6
    4950
    Loading More Posts
    • Oldest to Newest
    • Newest to Oldest
    • Most Votes
    Reply
    • Reply as question
    Log in to reply
    This topic has been deleted. Only users with question management privileges can see it.
    • teconsite
      teconsite last edited by

      Hello,

      We have a site that manages pagination through parameters in urls, this way:

      friendly-url.html
      friendly-url.html?p=2
      friendly-url.html?p=3
      ...

      We've rencently added the canonical tag pointing to friendly-url.html for all paginated results.

      In search console, we have the "p" parameter identified by google. 
      Now that the canonical has been added, should we still configure the parameter in search console, and tell google that it is being use for pagination?

      Thank you!

      1 Reply Last reply Reply Quote 0
      • dohertyjf
        dohertyjf last edited by

        Hi Teconsite, this is a great question.

        I would not recommend marketing the "p" parameter in Search Console. Instead, I'd leave it as "Let Google Decide" and use  your pagination SEO implementation to guide the search engines.

        There is still a lot of debate around pagination as it relates to SEO. The way I have always implemented is is:

        • Every paginated page canonicals to itself, because you do not want the search engines to start ignoring your paginated pages which are there somewhat for users, but also for SEO.
        • Use rel next/prev to help Google understand that they are in pagination, which will also help them rank the beginning of pagination for the terms you are trying to rank for.
        • Use noindex/follow on pages 2-N to be sure they stay out of Google's index.
        • Use the numbers showing how long pagination is to drive the search engines deep into your pagination to get all of your products/whatever indexed.  This is often done through linking to page 1, the last page, and the 3-5 pages on either side of the page you are currently on. So page 7 of 20 would like to page 1, pages 5-9, and page 20.

        The reason most people say to canonical pages 2-N to the base page is to preserve any link equity pointing to these pages and help the first page rank. However, I have almost never seen a deep paginated page with links, and if you have architected pagination correctly then the equity going into pages 2-N will also flow to page 1, just like product pages linking to category pages.

        Hope this helps!

        1 Reply Last reply Reply Quote 0
        • Gabriele_Layoutweb
          Gabriele_Layoutweb last edited by

          In this Moz guide regarding Google webmaster recommendations, it says you should still set the paginated page parameter in Google's Webmaster Tools:

          https://moz.com/ugc/seo-guide-to-google-webmaster-recommendations-for-pagination (search for the part "Coding Instruction for the View-All Option")

          Hope this helps!

          1 Reply Last reply Reply Quote 0
          • anthonydnelson
            anthonydnelson @teconsite last edited by

            You are sort of in an odd situation. You could tell Google that the "p" parameter is for pagination and they would better understand that. However, the canonical tag usage sort of tells Google that all of your paginated pages are actually duplicates of the first page.

            1 Reply Last reply Reply Quote 2
            • teconsite
              teconsite @anthonydnelson last edited by

              Hello Anthony!

              Thank you for your answer. I have been reading about the rel/prev and the canonical, and I found two different points of view about this. I know the recommendation of Google is the one that you have mentioned above, but as the CMS (Prestashop) is managing the paginated results the way I have shown, that is the one I am using.

              The question is, imagine that I have implemented the canonical the way you say before (or the way I did, I doesn't really matter for my question), should I still tell google that "p" parameter is a pagination parameter in Google Webmaster Tools or it's not necessary?

              Thank you!

              anthonydnelson 1 Reply Last reply Reply Quote 0
              • anthonydnelson
                anthonydnelson last edited by

                Typically, if you want to use the Canonical Tag for pagination, you would have it point to a View All style page, such as friendly-url.html&view=all.

                If you have too many products/pages in the pagination series, you might want to consider removing the canonical tag and implementing rel=prev/next. You can get more info here: http://googlewebmastercentral.blogspot.com/2011/09/pagination-with-relnext-and-relprev.html

                teconsite 1 Reply Last reply Reply Quote 3
                • 1 / 1
                • First post
                  Last post

                Got a burning SEO question?

                Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.


                Start my free trial


                Browse Questions

                Explore more categories

                • Moz Tools

                  Chat with the community about the Moz tools.

                • SEO Tactics

                  Discuss the SEO process with fellow marketers

                • Community

                  Discuss industry events, jobs, and news!

                • Digital Marketing

                  Chat about tactics outside of SEO

                • Research & Trends

                  Dive into research and trends in the search industry.

                • Support

                  Connect on product support and feature requests.

                • See all categories

                Related Questions

                • moon-boots

                  Google Pagination Changes

                  What with Google recently coming out and saying they're basically ignoring paginated pages, I'm considering the link structure of our new, sooner to launch ecommerce site (moving from an old site to a new one with identical URL structure less a few 404s). Currently our new site shows 20 products per page but with this change by Google it means that any products on pages 2, 3 and so on will suffer because google treats it like an entirely separate page as opposed to an extension of the first. The way I see it I have one option: Show every product in each category on page 1. I have Lazy Load installed on our new website so it will only load the screen a user can see and as they scroll down it loads more products, but how will google interpret this? Will Google simply see all 50-300 products per category and give the site a bad page load score because it doesn't know the Lazy Load is in place? Or will it know and account for it? Is there anything I'm missing?

                  Intermediate & Advanced SEO | | moon-boots
                  0
                • ostesmorbrod

                  Landing pages for paid traffic and the use of noindex vs canonical

                  A client of mine has a lot of differentiated landing pages with only a few changes on each, but with the same intent and goal as the generic version. The generic version of the landing page  is included in navigation, sitemap and is indexed on Google. The purpose of the differentiated landing pages is to include the city and some minor changes in the text/imagery to best fit the Adwords text. Other than that, the intent and purpose of the pages are the same as the main / generic page. They are not to be indexed, nor am I trying to have hidden pages linking to the generic and indexed one (I'm not going the blackhat way). So – I want to avoid that the duplicate landing pages are being indexed (obviously), but I'm not sure if I should use noindex (nofollow as well?) or rel=canonical, since these landing pages are localized campaign versions of the generic page with more or less only paid traffic to them. I don't want to be accidentally penalized, but I still need the generic / main page to rank as high as possible... What would be your recommendation on this issue?

                  Intermediate & Advanced SEO | | ostesmorbrod
                  0
                • YairSpolter

                  Block in robots.txt instead of using canonical?

                  When I use a canonical tag for pages that are variations of the same page, it basically means that I don't want Google to index this page. But at the same time, spiders will go ahead and crawl the page. Isn't this a waste of my crawl budget? Wouldn't it be better to just disallow the page in robots.txt and let Google focus on crawling the pages that I do want indexed? In other words, why should I ever use rel=canonical as opposed to simply disallowing in robots.txt?

                  Intermediate & Advanced SEO | | YairSpolter
                  0
                • aactive

                  Yoast & rel canonical for paginated Wordpress URLs

                  Hello, our Wordpress blog at http://www.jobs.ca/career-resources has a rel canonical issue since we added pagination to the front page and category-pages. We're using Yoast and it's incorrectly applying a rel-canonical meta tag referencing page 1 on page 2, 3, etc. This is a known misuse of the rel-canonical tag (per Google's Webmaster Blog - http://googlewebmastercentral.blogspot.ca/2013/04/5-common-mistakes-with-relcanonical.html, which says rel-canonical should be replaced with rel-prev and rel-next for page 2, 3, etc.). We don't see a way to specify anywhere in Yoast's options to correct this behaviour for page 2, 3, etc. Yoast allows you to override a page's canonical URL, otherwise it automatically uses the Wordpress permalink. My question is, does anyone know how to configure Yoast to properly replace rel-canonical tags with rel-prev and rel-next for paginated URLs, or do I need to look at another plugin or customize the behavior directly in my child theme code? This issue was brought up here as well: http://moz.com/community/q/canonical-help, but the only response did not relate to Yoast. (We're using Wordpress 3.6.1 and Yoast "Wordpress SEO" 1.4.18)

                  Intermediate & Advanced SEO | | aactive
                  0
                • esiow2013

                  May know what's the meaning of these parameters in .htaccess?

                  Begin HackRepair.com Blacklist RewriteEngine on Abuse Agent Blocking RewriteCond %{HTTP_USER_AGENT} ^BlackWidow [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Bolt\ 0 [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Bot\ mailto:craftbot@yahoo.com [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} CazoodleBot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^ChinaClaw [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Custo [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Default\ Browser\ 0 [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^DIIbot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^DISCo [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} discobot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Download\ Demon [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^eCatch [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ecxi [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^EirGrabber [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^EmailCollector [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^EmailSiphon [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^EmailWolf [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Express\ WebPictures [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^ExtractorPro [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^EyeNetIE [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^FlashGet [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^GetRight [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^GetWeb! [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Go!Zilla [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Go-Ahead-Got-It [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^GrabNet [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Grafula [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} GT::WWW [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} heritrix [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^HMView [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} HTTP::Lite [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} HTTrack [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ia_archiver [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} IDBot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} id-search [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} id-search.org [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Image\ Stripper [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Image\ Sucker [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Indy\ Library [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^InterGET [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Internet\ Ninja [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^InternetSeer.com [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} IRLbot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ISC\ Systems\ iRc\ Search\ 2.1 [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Java [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^JetCar [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^JOC\ Web\ Spider [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^larbin [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^LeechFTP [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} libwww [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} libwww-perl [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Link [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} LinksManager.com_bot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} linkwalker [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} lwp-trivial [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Mass\ Downloader [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Maxthon$ [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} MFC_Tear_Sample [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^microsoft.url [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Microsoft\ URL\ Control [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^MIDown\ tool [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Mister\ PiX [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Missigua\ Locator [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Mozilla.*Indy [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Mozilla.NEWT [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^MSFrontPage [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Navroad [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^NearSite [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^NetAnts [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^NetSpider [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Net\ Vampire [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^NetZIP [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Nutch [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Octopus [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Offline\ Explorer [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Offline\ Navigator [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^PageGrabber [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} panscient.com [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Papa\ Foto [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^pavuk [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} PECL::HTTP [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^PeoplePal [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^pcBrowser [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} PHPCrawl [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} PleaseCrawl [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^psbot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^RealDownload [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^ReGet [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Rippers\ 0 [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} SBIder [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^SeaMonkey$ [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^sitecheck.internetseer.com [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^SiteSnagger [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^SmartDownload [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Snoopy [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Steeler [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^SuperBot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^SuperHTTP [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Surfbot [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^tAkeOut [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Teleport\ Pro [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Toata\ dragostea\ mea\ pentru\ diavola [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} URI::Fetch [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} urllib [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} User-Agent [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^VoidEYE [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Web\ Image\ Collector [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Web\ Sucker [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Web\ Sucker [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} webalta [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebAuto [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^[Ww]eb[Bb]andit [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} WebCollage [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebCopier [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebFetch [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebGo\ IS [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebLeacher [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebReaper [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebSauger [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Website\ eXtractor [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Website\ Quester [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebStripper [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebWhacker [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WebZIP [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} Wells\ Search\ II [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} WEP\ Search [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Wget [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Widow [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WWW-Mechanize [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^WWWOFFLE [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Xaldon\ WebSpider [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} zermelo [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^Zeus [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ^(.)Zeus.Webster [NC,OR]
                  RewriteCond %{HTTP_USER_AGENT} ZyBorg [NC]
                  RewriteRule ^. - [F,L] Abuse bot blocking rule end End HackRepair.com Blacklist

                  Intermediate & Advanced SEO | | esiow2013
                  1
                • mark_baird

                  Should I use rel=canonical on similar product pages.

                  I'm thinking of using rel=canonical for similar products on my site. Say I'm selling pens and they are al very similar. I.e. a big pen in blue, a pack of 5 blue bic pens, a pack of 10, 50, 100 etc. should I rel=canonical them all to the best seller as its almost impossible to make the pages unique. (I realise the best I realise these should be attributes and not products but I'm sure you get my point) It seems sensible to have one master canonical page for bic pens on a site that has a great description video content and good images plus linked articles etc rather than loads of duplicate looking pages. love to hear thoughts from the Moz community.

                  Intermediate & Advanced SEO | | mark_baird
                  0
                • JosephFrost

                  Canonical link vs root domain

                  I have a wordpress website installed on http://domain.com/home/ instead of http://domain.com  - Does it matter whether I leave it that way with a canonical link from the domain.com to the domain.com/home/ or should I move the wordpress files and database to the root domain?

                  Intermediate & Advanced SEO | | JosephFrost
                  0
                • CommercePundit

                  How to fix issues regarding URL parameters?

                  Today, I was reading help article for URL parameters by Google. http://www.google.com/support/webmasters/bin/answer.py?answer=1235687 I come to know that, Google is giving value to URLs which ave parameters that change or determine the content of a page. There are too many pages in my website with similar value for Name, Price and Number of product. But, I have restricted all pages by Robots.txt with following syntax. URLs:
                  http://www.vistastores.com/table-lamps?dir=asc&order=name
                  http://www.vistastores.com/table-lamps?dir=asc&order=price
                  http://www.vistastores.com/table-lamps?limit=100 Syntax in Robots.txt
                  Disallow: /?dir=
                  Disallow: /?p=
                  Disallow: /*?limit= Now, I am confuse. Which is best solution to get maximum benefits in SEO?

                  Intermediate & Advanced SEO | | CommercePundit
                  0

                Get started with Moz Pro!

                Unlock the power of advanced SEO tools and data-driven insights.

                Start my free trial
                Products
                • Moz Pro
                • Moz Local
                • Moz API
                • Moz Data
                • STAT
                • Product Updates
                Moz Solutions
                • SMB Solutions
                • Agency Solutions
                • Enterprise Solutions
                • Digital Marketers
                Free SEO Tools
                • Domain Authority Checker
                • Link Explorer
                • Keyword Explorer
                • Competitive Research
                • Brand Authority Checker
                • Local Citation Checker
                • MozBar Extension
                • MozCast
                Resources
                • Blog
                • SEO Learning Center
                • Help Hub
                • Beginner's Guide to SEO
                • How-to Guides
                • Moz Academy
                • API Docs
                About Moz
                • About
                • Team
                • Careers
                • Contact
                Why Moz
                • Case Studies
                • Testimonials
                Get Involved
                • Become an Affiliate
                • MozCon
                • Webinars
                • Practical Marketer Series
                • MozPod
                Connect with us

                Contact the Help team

                Join our newsletter
                Moz logo
                © 2021 - 2025 SEOMoz, Inc., a Ziff Davis company. All rights reserved. Moz is a registered trademark of SEOMoz, Inc.
                • Accessibility
                • Terms of Use
                • Privacy

                Looks like your connection to Moz was lost, please wait while we try to reconnect.