Should I nofollow search results pages
-
I have a customer site where you can search for products they sell
url format is:
domainname/search/keywords/
keywords being what the user has searched for.
This means the number of pages can be limitless as the client has over 7500 products.
or should I simply rel canonical the search page or simply no follow it?
-
cheers
-
Hi there,
You've got the right idea, but let me suggest another tactic.
It's true that search functions can generate 1000's of urls that all tend to look like one another. Google suggests that you keep search result pages non-indexed, as these pages offer very little value and create tons of duplicate content.
http://www.seomoz.org/learn-seo/duplicate-content
Here's one way to handle your situation:
1. Put a meta "noindex,follow" tag in your search pages header, like this:
This tells search engines not to index the page, but allows them to follow the links on the page and flow link juice.
2. Hopefully you have a good site architecture and ways for search engines to discover your content. After step one, you can put a directive in your robots.txt file to block that directory from being crawled.
Something like:
User-agent: * Disallow: /search/
Which blocks anything in the search directory.
3. Find out if search engines have already indexed a lot of your search pages by performing a site: search in Google, like so:
site:yourdomain.com/search
If you find pages in Google's index that shouldn't be there, you can use Google Webmasters URL removal tool to take these out of the index. You can remove the entire search directory with a single request.
http://support.google.com/webmasters/bin/answer.py?hl=en&answer=1663427
This is a powerful and sometimes dangerous tool, so be careful!
4. Finally, if you'd like to add "nofollow" to your search results pages, this should be fine, but only after you've completed the steps above.
Keep in mind, this is only one possible solution. If you have significant link juice flowing through your search results, this strategy may not be the best. But in general, you want to keep search results out of Google's index, so I'm comfortable recommending this strategy for 90% of all cases.
Hope this helps! Best of luck with your SEO.
-
yes i would leave them.
-
so even though the search could generate lots of extra pages you think I should leave the pages as is?
-
Don't use the "no follow" attribute. The only time i'd recommend using "no follow" is on pages where you have external links . Blog comment pages, resources page etc.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Pages with Duplicate Content
When I crawl my site through moz, it shows lots of Pages with Duplicate Content. The thing is all that pages are pagination pages. How should I solve this issue?
Technical SEO | | 100offdeal0 -
Search Console Indexed Page Count vs Site:Search Operator page count
We launched a new site and Google Search Console is showing 39 pages have been indexed. When I perform a Site:myurl.com search I see over 100 pages that appear to be indexed. Which is correct and why is there a discrepancy? Also, Search Console Page Index count started at 39 pages on 5/21 and has not increased even though we have hundreds of pages to index. But I do see more results each week from Site:psglearning.com My site is https://wwww.psglearning.com
Technical SEO | | pdowling0 -
What should I do with all these 404 pages?
I have a website that Im currently working on that has been fairly dormant for a while and has just been given a face lift and brought back to life. I have some questions below about dealing with 404 pages. In Google WMT/search console there are reports of thousands of 404 pages going back some years. It says there are over 5k in total but I am only able to download 1k or so from WMT it seems. I ran a crawl test with Moz and the report it sent back only had a few hundred 404s in, why is that? Im not sure what to do with all the 404 pages also, I know that both Google and Moz recommend a mixture of leaving some as 404s and redirect others and Id like to know what the community here suggests. The 404s are a mix of the following: Blog posts and articles that have disappeared (some of these have good back-links too) Urls that look like they used to belong to users (the site used to have a forum) which where deleted when the forum was removed, some of them look like they were removed for spam reasons too eg /user/buy-cheap-meds-online and others like that Other urls like this /node/4455 (or some other random number) Im thinking I should permanently redirect the blog posts to the homepage or the blog but Im not sure what to do about all the others? Surely having so many 404s like this is hurting my crawl rate?
Technical SEO | | linklander0 -
Advice on improving ecommerce product detail pages to rank better in google search results.
Hi all, I run an ecommerce website, not a great ranked site, however i want to try and improve the product detail pages. To do this, i am first going to focus on 1 page (this one: http://goo.gl/eS62SU) If i type the product code directly into google.co.uk search i am on the 8th page (see https://www.google.co.uk/#q=hac-hfw2220r-z&start=70) which is a bit poor to say the least. I see this kind of thing for a lot of my products. Hence, i am going to see if over the next month or two i can get this one page moving up the rankings purely with on page optimisation. I would like to ask a couple of things: 1. Is there anything that jumps out at you as to why that product detail page could NOT ever rank well, i.e some code / set up of page etc that prevents google ranking it 2. Any advice you could give that might improve that page in rankings for its product code. FYI - I can not change the dynamic URL, I only have control over such things as product name / summary / features / spec etc any advice welcome
Technical SEO | | isntworkdull0 -
Why are only a few of our pages being indexed
Recently rebuilt a site for an auctioneers, however it has a problem in that none of the lots and auctions are being indexed by Google on the new site, only the pages like About, FAQ, home, contact. Checking WMT shows that Google has crawled all the pages, and I've done a "Fetch as Google" on them and it loads up fine, so there's no crawling issues that is standing out. I've set the "URL Parameters" to no effect too. Also built a sitemap with all the lots in, pushed to Google which then crawled them all (massive spike in Crawl rate for a couple days), and still just indexing a handful of pages. Any clues to look into would be greatly appreciated. https://www.wilkinsons-auctioneers.co.uk/auctions/
Technical SEO | | Blue-shark0 -
Duplicate page content
Hello, My site is being checked for errors by the PRO dashboard thing you get here and some odd duplicate content errors have appeared. Every page has a duplicate because you can see the page and the page/~username so... www.short-hairstyles.com is the same as www.short-hairstyles.com/~wwwshor I don't know if this is a problem or how the crawler found this (i'm sure I have never linked to it). But I'd like to know how to prevent it in case it is a problem if anyone knows please? Ian
Technical SEO | | jwdl0 -
"nofollow pages" or "duplicate content"?
We have a huge site with lots of geographical-pages in this structure: domain.com/country/resort/hotel domain.com/country/resort/hotel/facts domain.com/country/resort/hotel/images domain.com/country/resort/hotel/excursions domain.com/country/resort/hotel/maps domain.com/country/resort/hotel/car-rental Problem is that the text on ie. /excursions is often exactly the same on .../alcudia/hotel-sea-club/excursion and .../alcudia/hotel-beach-club/excursion The two hotels offer the same excursions, and the intro text on the pages are the exact same throughout the entire site. This is also a problem on the /images and /car-rental pages. I think in most cases the only difference on these pages is the Title, description and H1. These pages do not attract a lot of visits through search-engines. But to avoid them being flagged as duplicate content (we have more than 4000 of these pages - /excursions, /maps, /car-rental, /images), do i add a nofollow-tag to these, do i block them in robots.txt or should i just leave them and live with them being flagged as duplicate content? Im waiting for our web-team to add a function to insert a geographical-name in the text, so i could add ie #HOTELNAME# in the text and thereby avoiding the duplicate text. Right now we have intros like: When you visit the hotel ... instead of: When you visit Alcudia Sea Club But untill the web-team has fixed these GEO-tags, what should i do? What would you do and why?
Technical SEO | | alsvik0 -
SEOMoz Crawl Diagnostic indicates duplicate page content for home page?
My first SEOMoz Crawl Diagnostic report for my website indicates duplicate page content for my home page. It lists the home page URL Page Title and URL twice. How do I go about diagnosing this? Is the problem related to the following code that is in my .htaccess file? (The purpose of the code was to redirect any non "www" backlink referrals to the "www" version of the domain.) RewriteCond %{HTTP_HOST} ^whatever.com [NC]
Technical SEO | | Linesides
RewriteRule ^(.*)$ http://www.whatever.com/$1 [L,R=301] Should I get rid of the "http" reference in the second line? Related to this is a notice in the "Crawl Notices Found" -- "301 Permanent redirect" which shows my home page title as "http://whatever.com" and shows the redirect address as http://http://www.whatever.com/ I'm guessing this problem is again related to the redirect code I'm using. Also... The report indicates duplicate content for those links that have different parameters added to the URL i.e. http://www.whatever.com?marker=Blah Blah&markerzoom=13 If I set up a canonical reference for the page, will this fix this? Thank you.0