Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Paginated Pages Which Shouldnt' Exist..
-
Hi
I have paginated pages on a crawl which shouldn't be paginated:
https://www.key.co.uk/en/key/chairs
My crawl shows:
<colgroup><col width="377"></colgroup>
| https://www.key.co.uk/en/key/chairs?page=2 |
| https://www.key.co.uk/en/key/chairs?page=3 |
| https://www.key.co.uk/en/key/chairs?page=4 |
| https://www.key.co.uk/en/key/chairs?page=5 |
| https://www.key.co.uk/en/key/chairs?page=6 |
| https://www.key.co.uk/en/key/chairs?page=7 |
| https://www.key.co.uk/en/key/chairs?page=8 |
| https://www.key.co.uk/en/key/chairs?page=9 |
| https://www.key.co.uk/en/key/chairs?page=10 |
| https://www.key.co.uk/en/key/chairs?page=11 |
| https://www.key.co.uk/en/key/chairs?page=12 |
| https://www.key.co.uk/en/key/chairs?page=13 |
| https://www.key.co.uk/en/key/chairs?page=14 |
| https://www.key.co.uk/en/key/chairs?page=15 |
| https://www.key.co.uk/en/key/chairs?page=16 |
| https://www.key.co.uk/en/key/chairs?page=17 |Where is this coming from?
Thank you
-
You will also have to get those URLs out of the index once you fix the rel next/prev issue. In order to do that effectively, they should return a 404 or 410 status code in the HTTP header so Google knows that they no longer exist (even though they never really did in the first place). Otherwise, it's what is known as a "soft 404" in which the page doesn't really exist, but returns a 200 (OK) status code, which is confusing to Google if you don't want them indexed.
-
Hi Becky
I can see chairs:
https://www.key.co.uk/en/key/chairs
But the paginated versions above are not in there. (can you see them?)
All you need to do is remove this directive for pages without a page 2: rel="next" href="https://www.key.co.uk/en/key/chairs?page=2" > as there is no page 2 for chairs.
Regards
Nigel
-
Hi Nigel
Thanks for jumping in. I'm confused as I have found the pages on my screaming frog crawl?
This page https://www.key.co.uk/en/key/chairs shouldn't have any pagination as there are no additional pages, but there is rel=next in the source code...
Now I'm a bit confused!
Becky
-
Yes I've just gone through every top level page too & pagination is awful, so I'm compiling a list and a case to push it.
It's pretty bad across the site, so I'll push for this to be updated. I find new issues with it all the time..
Thanks for your help!
-
Yes exactly. Even though the pages don't exist to the user, they still technically exist. If I were you, I'd take a very deep look at pagination on your site. If this is happening at scale, then fixing it could be a major improvement to your site. I took a look and it seems to be happening on all your top-level category pages like Chairs, Office Furniture, Shelving & Racking, etc.
These paginated pages are essentially a bunch of duplicate pages of your main category pages, each with a self-referencing canonical (which is the proper way to set up pagination). So Google could be extremely confused about which one to rank. In most cases, Google will rank page 1 because the use of rel="next"/rel="prev" is essentially telling Google that page 1 is the canonical version. However, you're still opening yourself up to the possibility of Google crawling all of these duplicate pages which is a huge waste on your crawl budget.
Hope that helps!
-
Hi
Thank you both.
We do have issues with our pagination which I've raised with developers, but it's taking forever to sort out. I'll flag this as well.
So even though the content on the paginated pages for Chairs doesn't exist we still need to remove the tags on these - https://www.key.co.uk/en/key/chairs?page=10
-
If you view your source code, you'll notice you are actually using rel="next" and rel="prev" on the main category page (https://www.key.co.uk/en/key/chairs). This is why you (and most likely Googlebot as well) are crawling these paginated pages. Even though you don't have links to the paginated pages on the main category page, they still exist and you're giving crawlers the directive (rel next / rel prev) to crawl them.
If you remove rel="next" on the category home page, that should help but you should really remove rel="next" and rel="prev" on the paginated pages as well. Unless you do that, Google will still find them and crawl them because they're aware these pages exist and they're likely indexed.
Here's a great resource on understanding pagination as well as the correct use of rel="next" and rel="prev" from Maile Ohye at Google: https://www.youtube.com/watch?v=njn8uXTWiGg
Hope this helps!
Cheers!
-Tyler -
Nice website by the way. It looks very professional. And your 49 DA is very impressive.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I redirect a domain we control but which has been labeled 'toxic' or just shut it down?
Hi Mozzers: We recently launched a site for a client which involved bringing in and redirecting content which formerly had been hosted on different domains. One of these domains still existed and we have yet to bring over the content from it. It has also been flagged as a suspicious/toxic backlink source to our new domain. Would I be wise to redirect this old domain or should I just shut it down? None of the pages seem to have particular equity as link sources. Part of me is asking myself 'Why would we redirect a domain deemed toxic, why not just shut it down.' Thanks in advance, dave
Intermediate & Advanced SEO | | Daaveey0 -
Fresh page versus old page climbing up the rankings.
Hello, I have noticed that if publishe a webpage that google has never seen it ranks right away and usually in a descend position to start with (not great but descend). Usually top 30 to 50 and then over the months it slowly climbs up the rankings. However, if my page has been existing for let's say 3 years and I make changes to it, it takes much longer to climb up the rankings Has someone noticed that too ? and why is that ?
Intermediate & Advanced SEO | | seoanalytics0 -
How will changing my website's page content affect SEO?
Our company is looking to update the content on our existing web pages and I am curious what the best way to roll out these changes are in order to maintain good SEO rankings for certain pages. The infrastructure of the site will not be modified except for maybe adding a couple new pages, but existing domains will stay the same. If the domains are staying the same does it really matter if I just updated 1 page every week or so, versus updating them all at once? Just looking for some insight into how freshening up the content on the back end pages could potentially hurt SEO rankings initially. Thanks!
Intermediate & Advanced SEO | | Bankable1 -
Do 404s really 'lose' link juice?
It doesn't make sense to me that a 404 causes a loss in link juice, although that is what I've read. What if you have a page that is legitimate -- think of a merchant oriented page where you sell an item for a given merchant --, and then the merchant closes his doors. It makes little sense 5 years later to still have their merchant page so why would removing them from your site in any way hurt your site? I could redirect forever but that makes little sense. What makes sense to me is keeping the page for a while with an explanation and options for 'similar' products, and then eventually putting in a 404. I would think the eventual dropping out of the index actually REDUCES the overall link juice (ie less pages), so there is no harm in using a 404 in this way. It also is a way to avoid the site just getting bigger and bigger and having more and more 'bad' user experiences over time. Am I looking at it wrong? ps I've included this in 'link building' because it is related in a sense -- link 'paring'.
Intermediate & Advanced SEO | | friendoffood0 -
My landing pages don't show up in the SERPs, only my frontpage does.
I am having some trouble with getting the landing pages for a clients website to show up in the SERPs.
Intermediate & Advanced SEO | | InmediaDK
As far as I can see, the pages are optimized well, and they also get indexed by Google. The website is a danish webshop that sells wine, www.vindanmark.com Take for an instance this landing page, http://www.vindanmark.com/vinhandel/
It is optimzied for the keywords "Vinhandel Århus". Vinhandel means "Winestore" and "Århus" is a danish city. As you can see, I manage to get them at page 1 (#10), but it's the frontpage that ranks for the keyword. And this goes for alle the other landing pages as well. But I can't figure out, why the frontpage keep outranking the landingpages on every keyword.
What am I doing wrong here?1 -
Is 301 redirect suggested on pagination pages
Hi - Due to pagination the default page of site is coming in 2 url with - ?page=1/ sub-url and /sub-url is 301 a recommended solution due to this pagination urls Also - is it required to create separate title and meta description of every pagination page We are taking specifically in context of our discounts and offer section http://www.mycarhelpline.com/index.php?option=com_offers&view=list&Itemid=9
Intermediate & Advanced SEO | | Modi0 -
Blocking Pages Via Robots, Can Images On Those Pages Be Included In Image Search
Hi! I have pages within my forum where visitors can upload photos. When they upload photos they provide a simple statement about the photo but no real information about the image,definitely not enough for the page to be deemed worthy of being indexed. The industry however is one that really leans on images and having the images in Google Image search is important to us. The url structure is like such: domain.com/community/photos/~username~/picture111111.aspx I wish to block the whole folder from Googlebot to prevent these low quality pages from being added to Google's main SERP results. This would be something like this: User-agent: googlebot Disallow: /community/photos/ Can I disallow Googlebot specifically rather than just using User-agent: * which would then allow googlebot-image to pick up the photos? I plan on configuring a way to add meaningful alt attributes and image names to assist in visibility, but the actual act of blocking the pages and getting the images picked up... Is this possible? Thanks! Leona
Intermediate & Advanced SEO | | HD_Leona0 -
Should the sitemap include just menu pages or all pages site wide?
I have a Drupal site that utilizes Solr, with 10 menu pages and about 4,000 pages of content. Redoing a few things and we'll need to revamp the sitemap. Typically I'd jam all pages into a single sitemap and that's it, but post-Panda, should I do anything different?
Intermediate & Advanced SEO | | EricPacifico0