Is CloudFlare bad for SEO?
-
I have been hit by DDoS attacks lately...not on a huge scale, but probably done by some "script kiddies" or competitors of mine.
Still, I need to take some action in order to protect my server and my site against all of this spam traffic that is being sent to it. In the process of researching the tools available for defending a website from a DDoS attack, I came across the service offered by CloudFlare.com.
According to the CloudFlare website, they protect your site against a DDoS attack by showing users/visitors they find suspicious an interstitial that asks them if they are a real user or a bot...this interstitial contains a Captcha that suspicious users are asked to enter in order to visit the site.
I'm just wondering what kind of an effect such an interstitial could have on my Google rankings...I can imagine that such a thing could add to increased click-backs to the SERPs and, if Google detects this, to lower rankings.
Has anyone had experience with the DDoS protection services offered by CloudFlare, who can say a word or two regarding any effects this may have on SEO?
Thanks
-
i know this is a old thread and i see where people says cloudflare dont have a bad effect but i need someone to tell me more about it because am scare loosing all my hardwork on my Music niche
-
Wanna know if cloudflare depends on some country and unavailable in some countries? thinking of using this as well on my subscene page
-
it should not negatively impact your ranking in any way shape or form. If you are seeing the speed increase from it it will most likely be a positive effect on your rankings. Google no longer uses IP's so it's fine. to have a reverse proxy in front like CloudFlare. you like it and you think it's fine to help move site, by all means, try it out.
It will only have a negative impact on your rankings if you do something that's almost impossible in block Google Bot with your firewall which is not going to happen unless you really start getting technical with the WAF, in other words, don't worry about it
sincerely,
Tom
-
Seriously I wanted to go for cloudflare too on my music blog because they make site super fast but I'm scared to loose my ranking
Should I or not
-
I agree that they are much better than they where I’m still using Incapsula & Fastly.
I use CloudFlare Enterprise for China.
Glad to hear you’re doing well!
Tom
-
We have hundreds of sites managed through Cloudflare. So far the good far outweighs the bad, although I am beginning to suspect bad neighbors could certainly negatively influence ranking. We have a few comparisons ongoing. We will know for sure in the coming months. I only wish Cloudflare would take a higher stance about which sites it is willing to get in bed with.
Here https://sarkariresultsite.com/ is the our new website which is managed by Cloudflare.
-
With CloudFlare the challenge page you refer to is dependent on your security questions. My knowledge of DDos mitigation is not great, but what I do is have the CloudFlare security settings on Low. If I was to come under attack, I would react and put the security settings on high for the duration.
If the DDos is bad, than you may need the expensive plan that has specialist DDos protection.
today i transfer my site https://sarkariresultsite.in on cloud flare with cloudflare flexible SSL
-
Generally speaking, CAPTCHA based DDoS protection is not advised as it will drive away clients and DDoS bots alike.
Same goes for delay pages.Also, DDoS attacks are swift and can bring site/server down in minutes so the manual approach is risky and probably less effective than it sounds on paper. Even if you are always near your PC, like many of us here are, you`ll still be exposed on weekends, nighttime, launch breaks, etc... Hackers know that and will exploit it. (favorite attack time - VERY EARLY Sunday mornings... )
You should also consider that DDoS can serve as a means to an end. It's often used to "soften" the protective layer to hack the database, inject the site with backdoors, deface or delete it and so on and so forth.
What you really need is a combination of auto-triggering and transparent mitigation, which shies away from intrusive mitigation methods in favor of more subtle progressive challenges (i.e. JS challenge combined with reputation and behavior monitoring).
This may sound fancy but it really this is the emerging industry standard and growing competition between vendors actually drives costs down. -
With CloudFlare the challenge page you refer to is dependent on your security questions. My knowledge of DDos mitigation is not great, but what I do is have the CloudFlare security settings on Low. If I was to come under attack, I would react and put the security settings on high for the duration.
If the DDos is bad, then you may need the expensive plan that has specialist DDos protection.
-
Hi
SEO wise, there are many things to consider before on-boarding a CDN service and bot filtering is one of them.
I've covered the topic in 2 blogs posts about a year ago:1. CDN's SEO related advantages 2. CDN's SEO related myths (and possible issues to consider)
PS: I work for CDN-based security and acceleration service provider - Incapsula.
-
Hi
http://www.neustar.biz/enterprise/ddos-protection
https://www.verisigninc.com/en_US/forms/ddosattackservice.xhtml
http://www.incapsula.com/ddos/anti-ddos-protection
What can happen
If you want a true built for enterprise host that actually has a perfect record and will care more than any other that I know of about security, up time and would you benefit yes you would fire host is the best they offer built-in DoS/DDoS mitigation
2 answer your question about cloud flair I do not feel adding them will give you the same advantages as a enterprise level hosting company.
For instance I believe fire host to be the best let's say you think liquid Web managed dedicated with layer 7 DDoS protection is better.
Well that's up to you and me to make our decisions. Based on past history and our own opinions of the companies.
If you where to add CloudFlare
you would receive CloudFlare layer 7 DDoS protection only once you started paying $200 a month of free cloud for a service is not a system that will protect you against an attack
http://www.cloudflare.com/features-security
There content delivery network as they call it is simply a reverse proxy it's not what other people would call a true CDN once again the issues would be a large with the base plan http://www.cloudflare.com/features-cdn
Your so-called anti-cast DNS CloudFlare says is standard changes your IP address by definition any cast and DNS does not change your IP. There are some methods two get around this but it's not simple.
A $200 a month "CloudFlare Railgun origin network optimizer"
http://www.cloudflare.com/railgun
Will do the same thing as any high end content delivery network as far as fixing the front end code edge cast will do that as will NetDNA, Akamai, numerous others so may I can't list them all.
Information on Akamai stopping attacks
The unique thing is edge cast is very fast so is Akamai you can purchase a Akamai CDN from liquid Web for $120 a month with a terabyte of bandwidth that's a lot.
It will do more than the CloudFlare Railgun including help stop DDoS attacks it will do it faster and if were to spend the money on a liquid Web VPS with DDoS you would be at about $175 a month or you could start off on the base package with fire host and that would cost you $200 a month however for a month $250 you can you can add edge casts outstanding content delivery network
I have accounts with liquid Web and fire host I use fire host primarily and really do not do much with liquid Web however both of them are better than CloudFlare in terms of protection and in terms of speed.
http://www.thewhir.com/web-hosting-news/cn-domain-service-restored-following-massive-ddos-attack
I of course have tried to use CloudFlare and found it worthless to me.
My main point is for $200 a month you can get a much faster and safer setup.
The other thing to remember when CloudFlare said it blocked a 300 GB attack that is considered fictitious and debatable by many in the
industry myself included. I don't think switches that At 10 Gb can do that when there's only two pointing at it
Here is a link from cloud flair talking about how good of a job they did saving the Internet and stopping the biggest attack ever. They did not succeed in keeping Spamhaus online so they did not stop the attack in my opinion.
http://blog.cloudflare.com/the-ddos-that-almost-broke-the-internet
"On Monday, March 18, 2013 Spamhaus contacted CloudFlare regarding an attack they were seeing against their website spamhaus.org. They signed up for CloudFlare and we quickly mitigated the attack." & said there was "more then than 300Gbps of attack traffic"
The other big difference between CF and everyone else is they utilize there any cast network to accept the attack this slows down any one on that network instantly as it did when "the attack that almost broke the Internet happened"
Instead of moving the traffic to a private cloud and dealing with it so it would not affect others using their service they basically did not have any regard for their users at that time.
"The attackers were quiet for a day. Then, on March 22 at 18:00 UTC, the attack resumed, peaking at 120Gbps of traffic hitting our network. As we discussed in the previous blog post, CloudFlare uses Anycast technology which spreads the load of a distributed attack across all our data centers. This allowed us to mitigate the attack without it affecting Spamhaus or any of our other customers. The attackers ceased their attack against the Spamhaus website four hours after it started."
CloudFlare actually spread the attack on to their own customers this is something that harmed everyone using their service that day for I believe five days total.
http://www.theregister.co.uk/Print/2013/03/28/spamhaus_mega_ddos_little_collateral_damage/
http://www.webhostingtalk.com/showthread.php?t=1065748
Nearly half of DDoS attacks in the first half of 2013 are now larger than 1Gbps, up from 13.5 percent in 2012,
http://www.thewhir.com/web-hosting-news/prolexic-stops-massive-dns-reflection-attack
largest DNS reflection attack ever recorded,” which peaked at 167Gbps.
If this is now the largest what was the 300gbps?
It is not bad for SEO in itself however using a CAPTCHA page if it thinks your visitors is a attack or bot is very bad & happens allot I know changing your IP address can cause some issues. What I'm getting at is if you want the type of protection you were talking about there are much better methods of going about getting it.
Try them out see what you think you can really only learn more about it.
I hope this helps you,
Thomas
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page Speed Factors For SEO
Hey Guys, I have developed a page and optimised it. I have got a dilemma, I have 2 variants of the optimised page I could use. The page is responsive and uses bootstrap from an external CDN. The 2 variants: External CDN - This is adding an an extra request and is delivering the entire framework (not ideal for mobile) I've looked in the node/grunt.js route (+unCSS) to remove redundant CSS, which led me to my next variant. Inline CSS. After doing some grunt.js work I shaved out the redundant code from the framework then added it inline. I will also point out that all assets are optimised, all CSS/JS/HTML is minifed. In terms for score the 1st variant is less than the second, but I believe that most users of the internet already have bootstrap cached due to it being so common. The ultimate question comes down to ranking, I'm not entirely sure where I draw the line between development and SEO (I will also ask in Stack Overflow). Which one would rank better? all other factors being equal.
Intermediate & Advanced SEO | | AkashMakwana0 -
Odesk SEO Cover letter
How write and what to include in an SEO cover letter to apply a job on Odesk? Any example??
Intermediate & Advanced SEO | | ross254sidney0 -
SEO Priorities for Ecommerce Sites
Hello All! What is the best way to rank SEO tasks by PRIORITY for Ecommerce sites to improve?? It can be quite overwhelming with all the types of projects/tasks needed to improve organic rankings... How would you rank the most CRITICAL tasks to spend the MOST TIME on to the tasks you spend less on. Appreciate your input in advance 🙂 Thank you! Mark
Intermediate & Advanced SEO | | wickerparadise0 -
SEO Feedback Please
Hi, I was hoping to get some feedback on this site's SEO: http://www.chantre.com/ I also have different URLs which can be accessed from the left side of the site under "Quick Search Links" that go to the different offices within this company. Thoughts on how to improve would be great. Thanks in advance.
Intermediate & Advanced SEO | | gXe0 -
How do I make my URLs SEO friendly?
Hi all, I am aware that overly-dynamic URLs hurt a website's SEO potential and I want to fix mine. At present they look like this: http://www.societyboardshop.co.uk/products.php?brand=Girl+Skateboards&BrandID=153 What do I need to do to fix them please... do I add some code to the htaccess file? Many thanks, much apreciated. Paul.
Intermediate & Advanced SEO | | Paul530 -
Corporate pages and SEO help
We own and operate more than two dozen educational related sites. The business team is attempting to standardize some parts of our site hierarchy so that our sitemap.php, about.php, privacy.php and contact.php are all at the root directory. Our sitemap.php is generated by our sitemap.xml files, which are generated from our URLlist.txt files. I need to provide some feedback on this initiative. I'm worried about adding more stand-alone pages to our root directory and as part of a separate optimization in the future I was planning to suggest we group the "privacy", "about" and "contact" pages in a separate folder. We generally try to put our most important pages/directories for SEO in the root as our homepages pass a lot of link juice and have high authority. We do not invest SEO time into optimizing these pages as they're not pages we're trying to rank for, and I've already been looking into even no-following all links to them from our footer, sitemap, etc. I know that adding these "corporate" pages to a site are usually a standard part of the design process but is there any SEO benefit to having them at the root? And along the same lines, is there any SEO harm to having unimportant pages at the root? What do you guys think out there in Moz land?
Intermediate & Advanced SEO | | Eric_edvisors0 -
Disabled/Accessibilty vs SEO?
Can anyone point me to resources that helps website owners balance these two issues? Or how to SEO a site meant for disabled users? or how to make an SEO'd site more accessible? Thanks!
Intermediate & Advanced SEO | | mjcarrjr0 -
The SEO process...step by step
In what order do you carry out your SEO? do you start with keyword research or competitor analysis or is it on-page and then on to link building, do you concentrate on article submissions and then move on to social networking? List your step by step process here.
Intermediate & Advanced SEO | | IPIM0