Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Membership/subscriber (/customer) only content and SEO best practice
-
Hello Mozzers, I was wondering whether there's any best practice guidance out there re: how to deal with membership/subscriber (existing customer) only content on a website, from an SEO perspective - what is best practice?
A few SEOs have told me to make some of the content visible to Google, for SEO purposes, yet I'm really not sure whether this is acceptable / manipulative, and I don't want to upset Google (or users for that matter!)
Thanks in advance, Luke
-
I'd say it's mostly transferable as plenty of content is found in both news and the main index. News is more of a service overlay that attempts to better handle user expectations for frequency and speed of response when it comes to news items. Still, old news gets into the index and treated like content from most any site so if you have a subscription based model that aligns with what they're recommending for more news orientated sites, at least you're fitting into a form of what they outline.
-
Everything I could find was related to Google News, but not the main index? Is it directly transferrable? Especially given it's the _oldest _content that's going to end up being paid for in my example.
-
As an example, the New York Times does this via tracking of how many full articles a user reads while allowing Googlebot full access to its articles. Sites that use this method employ "no cache" on Google so articles can't be read there and then various forms of tracking to ensure users are being counted correctly. Here are some thoughts on this and more from Google's side that might help you out: https://support.google.com/news/publisher/answer/40543. Cheers!
-
Don't want to hijack this thread at all, but I was looking for something very similar and wonder if we're thinking of the same thing?
A blog wants to make it's older content only available to premium members - but still retain a snippet of that content (perhaps the first few paragraphs (the posts are quite long) as visible to search engines. Thus allowing traffic to arrive on the site from the content, but not necessarily view it.
I saw that as being against the spirit of what Google wants to do, but was hoping for a little clarity on that. I wonder if the OP was thinking of something similar?
-
As Leonie states, the search engines are for public facing content. If your site is completely private then you'd be more interested in making sure it's not found anywhere other than by members, however it sounds like you have some aspects of the site that could be public or created to attract new members. Typically in these cases you pull small topical samples from the site that are shown to benefit the members and help articulate why membership is valuable. It may be a matter of having what is practically like two sites: the public facing, membership recruitment site, and the private, non-indexed membership site. Cheers!
-
Hi, if your whole website is for members and behind a login and password, Searchengines can't index the website and thus not visisble for others than your members.
if you want other people to find your website, you'll need a public part, which you can optimize for your users and searchengines.
the question is: do you want other people than your members find the website, if yes, than you'll need content that searchengines can find. If the answer is no you can hide the whole website behind a login and password.
i manage a website which a part of that is only for members. that part is not optimized and behind a login and password. The rest of the site is public and need to be found in the searchengines. This part is optimized for on - and off page seo.
Grtz, Leonie
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
SEO Best Practices regarding Robots.txt disallow
I cannot find hard and fast direction about the following issue: It looks like the Robots.txt file on my server has been set up to disallow "account" and "search" pages within my site, so I am receiving warnings from the Google Search console that URLs are being blocked by Robots.txt. (Disallow: /Account/ and Disallow: /?search=). Do you recommend unblocking these URLs? I'm getting a warning that over 18,000 Urls are blocked by robots.txt. ("Sitemap contains urls which are blocked by robots.txt"). Seems that I wouldn't want that many urls blocked. ? Thank you!!
Intermediate & Advanced SEO | Sep 4, 2017, 6:02 AM | jamiegriz0 -
Best way to "Prune" bad content from large sites?
I am in process of pruning my sites for low quality/thin content. The issue is that I have multiple sites with 40k + pages and need a more efficient way of finding the low quality content than looking at each page individually. Is there an ideal way to find the pages that are worth no indexing that will speed up the process but not potentially harm any valuable pages? Current plan of action is to pull data from analytics and if the url hasn't brought any traffic in the last 12 months then it is safe to assume it is a page that is not beneficial to the site. My concern is that some of these pages might have links pointing to them and I want to make sure we don't lose that link juice. But, assuming we just no index the pages we should still have the authority pass along...and in theory, the pages that haven't brought any traffic to the site in a year probably don't have much authority to begin with. Recommendations on best way to prune content on sites with hundreds of thousands of pages efficiently? Also, is there a benefit to no indexing the pages vs deleting them? What is the preferred method, and why?
Intermediate & Advanced SEO | Dec 4, 2016, 3:14 PM | atomiconline0 -
Does blocking foreign country IP traffic to site, hurt my SEO / US Google rankings?
I have a website is is only of interest to US visitors. 99% (at least) of Adsense income is from the US. But I'm getting constant attempts by hackers to login to my admin account. I have countermeasures fo combat that and am initiating others. But here's my question: I am considering not allowing any non US, or at least any non-North American, traffic to the site via a Wordpress plugin that does this. I know it will not affect my business negatively, directly. However, are there any ramifications of the Google bots of these blocked countries not being able to access my site? Does it affect the rankings of my site in the US Google searches. At the very least I could block China, Russia and some eastern European countries.
Intermediate & Advanced SEO | Sep 15, 2014, 4:18 PM | bizzer0 -
Best Practice For Company/Client Logo Endorsement
Article: http://searchengineland.com/homepage-sliders-are-bad-for-seo-usability-163496 I came across the following article and somewhat agree with the authors summary.
Intermediate & Advanced SEO | Oct 11, 2013, 12:25 AM | Mark_Ch
I find sliders a distraction to B2B users and overall offers no SEO benefits. Scenario
As a service provider, over time I have worked with many high profile blue chip comnpanies. As part of my site redesign, I'm looking to show users my client achievements. My initial thoughts are to carry out the following: On the home page I'm looking to incorporate some high profile company logos (similar to http://www.semrush.com) with a hyperlink "more customers" to the right of logo caption. The link will take the user to a dedicated page (www.mydomain.co.uk/customer) showing a comprehensive list of company logos. Questions
#1 Is the above practice good or bad.
#2 Is there a better way to achieve the above Any other practical advise on user experience, social engagement, website speed, etc would be much appreciated. Thanks Mark0 -
How to perform Local SEO for sites like Angies List/Task Rabbit or Craigslist
I have a new SEO client that has a business model similar to Criagslist and Angies List or Task Rabbit, Where they offer local based services nationwide. My first thought was Local link building and citation building etc. But the issue is they are a purely online service company and they don't have a phyiscal address in every city/state they will be offering their services in. What is the best course of action for providing SEO services for this type of business model. I am pretty much at a stand still on how to rank them locally for the areas they provide services in. it's a business model that involves local businesses and customers looking for services from those local businesses.
Intermediate & Advanced SEO | Apr 24, 2013, 6:22 PM | VITALBGS0 -
Best practice for removing indexed internal search pages from Google?
Hi Mozzers I know that it’s best practice to block Google from indexing internal search pages, but what’s best practice when “the damage is done”? I have a project where a substantial part of our visitors and income lands on an internal search page, because Google has indexed them (about 3 %). I would like to block Google from indexing the search pages via the meta noindex,follow tag because: Google Guidelines: “Use robots.txt to prevent crawling of search results pages or other auto-generated pages that don't add much value for users coming from search engines.” http://support.google.com/webmasters/bin/answer.py?hl=en&answer=35769 Bad user experience The search pages are (probably) stealing rankings from our real landing pages Webmaster Notification: “Googlebot found an extremely high number of URLs on your site” with links to our internal search results I want to use the meta tag to keep the link juice flowing. Do you recommend using the robots.txt instead? If yes, why? Should we just go dark on the internal search pages, or how shall we proceed with blocking them? I’m looking forward to your answer! Edit: Google have currently indexed several million of our internal search pages.
Intermediate & Advanced SEO | Feb 10, 2016, 6:49 AM | HrThomsen0 -
Domain Name Change - Best Practices?
Good day guys, We got a restaurant that is changing its name and domain. However they are keeping the same server location, same content and same pages (we are just changing the logo on the website). It just has to go a new domain. We don't want to lose the value of the current site, and we want to avoid any duplicate penalties. Could you please advise of the best practices of doing a domain name change? Thank you.
Intermediate & Advanced SEO | Jul 30, 2012, 2:57 PM | Michael-Goode0 -
Duplicate page titles Wordpress SEO/Yoast
Hi I have a Wordpress site using the Wordpress SEO plugin by Yoast. Everything appears to be fine except that on 1 Feb SEOMoz crawl suddenly picked up a bunch of errors. The errors are duplicate page titles, and these exist only for the mysite.com/page/X pages. I can't find any setting in Yoast that looks wrong or tells me how to fix this. The pages are also dynamically canonicalizing to themselves - not sure if this makes any difference although I don't know how this is happening. Does anyone know how to fix this duplicate title error? Alex
Intermediate & Advanced SEO | Mar 26, 2012, 7:51 AM | alextanner0