Is the TTFB for different locations and browsers irrelevant if you are self-hosting?
-
Please forgive my ignorance on this subject. I have little to no experience with the technical aspects of setting up and running a server.
Here is the scenario:
We are self-hosted on an Apache server. I have been on the warpath to improve page load speed since the beginning of the year. I have been on this warpath not so much for SEO, but for conversion rate optimization. I recently read the Moz Post "How Website Speed Actually Impacts Search Rankings" and was fascinated by the research regarding TTFB. I forwarded the post to my CEO, who promptly sent me back a contradictory post from Cloudflare on the same topic. Ily Grigorik published a post in Google+ that called Cloudflare's experiment "silly" and said that "TTFB absolutely does matter."
I proceeded to begin gathering information on our site's TTFB using data provided by http://webpagetest.org. I documented TTFB for every location and browser in an effort to show that we needed to improve. When I presented this info to my CEO (I am in-house) and IT Director, that both shook their heads and completely dismissed the data and said it was irrelevant because it was measuring something we couldn't control.
Ignorant as I am, it seems that Ilya Grigorik, Google's own Web Dev Advocate says it absolutely is something that can be controlled, or at least optimized if you know what you are doing.
Can any of you super smart Mozzers help me put the words together to express that TTFB from different locations and for different browsers is something worth paying attention to? Or, perhaps they are right, and it's information I should ignore?
Thanks in advance for any and all suggestions!
Dana
-
Yes, very helpful guys. I appreciate it!
-
Thanks Igal and hopefully you have some info to work with Dana!
-
Many thanks to both Vadim and Igal for such great information and also a really great thread on the subject. I really, really appreciate your answers.!
-
Honestly, I don't know. I don't think TTFB was ever comparatively tested - at least no to the best of my knowledge.
For security, these are some of the resources I can point to.
I understand that this is not the main issue
Still, I wanted to provide some factual context to my previous statements.
http://zeroscience.mk/files/wafreport2013.pdf http://ddos-protection-services-review.toptenreviews.com/ http://tonyonsecurity.com/2012/11/13/protecting-your-website-cloudflare-or-incapsula/
(This last one is interesting since Tony is a COO of Sucuri. Some would call his our competitor. I prefer 'colleague' )
-
security wise it seems both of you guys have stellar options. for me the issue is performance, caching for dynamic sites, CDN performance, and in this case TTFB response. I was not sure with your response do you have faster TTFB to CF?
Thanks
-
Hi Vadim
Thanks.
Yep, I work for Incapsula but no, we are not the said "Mod".As for CF comparison... Generally speaking, we are more business oriented and security focused. I know that our security offering is more comprehensive, especially because both WAFs were comparatively pen-tested on several occasions and we always came out as consistently (and significantly) better option.We also have addition security features - like 2FA support and backdoor shell protection - which CF simply doesn't offer and we do more in way of ddos mitigation, especially against smart application layer attacks which require security capabilities, besides network muscle.
Still, speed wise, I always considered us to be pretty much on the same level. However, until few days ago I never considered TTFB to be such core SEO factor, so maybe we have better performance there...
But again, to be fair, I`m only speculating - mostly based on the CF blog you've shared.
(if TTFB is considered un-important, it might also be under developed...)Might be an interesting thing to test and document.
-
Hi Igal,
Do you work for incapsula, you are mentioned as a Mod on the blog?
I have heard great things about incapsula from others, but in terms of TTFB is it better than cloudflare? If so, how so?
Also any other ways that it excels Cloudflare? any ways its inferior to Cloudflare in your opinion?
Thanks I am really looking for more info, as I had great results with Cloudflares features and offering, wondering if I should give Incapsula a run
Thanks
-
I absolutely agree with Vadim. (+1)
Google is the best source for Google facts. Everything else is just speculation.
And yes, generally speaking, the best answer is to use a CDN....
The reason is simple. CNDs proxy technology, which was designed to minimize "physical" distances between the site's content and browsers, directly influences TTFB.Being an in-house SEO for a CDN company I get a lot of questions about this from our support and clients. I have to admit, until recent Moz post, I wasn't aware of full implications of TTFB and considered it to be one of few page load speed related metrics. (http://moz.com/blog/how-website-speed-actually-impacts-search-ranking)
This post really helped me get a better grasp on things. Interestingly enough, few month ago one of our clients Guest Posted in our blog about speed improvement gained by our free plan. Among other things, he mentioned 70% improvement in TTFB (grade going from F to A)
(http://www.incapsula.com/the-incapsula-blog/item/718-what-incapsula-free-did-for-my-site)At the time I didn't give it much attention. Because, like many others, I was focusing on overall load speeds....
Now I can't help but feel that this was a missed opportunity.
This post could be even better with the added SEO angle...
If anyone here is interested in giving this a try and guest posting about it, I`ll be happy to provide all resources needed on our end. -
Yea this makes sense as others have said that Cloudflare is trying to say that TTFB is not the most important metric, and so they published this study, as it aids their business model.
I would do just that listen to Google dev vs Cloudflare. Also the way I think about it even if their studies are true, where for the overall benefit TTFB would have to increase if you are using some compression, you still need to work and decrease your TTFB either way, that is just intuition. I apologize if I made it seem that TTFB is to be ignored, because Cloudflare state's that quite boldly,
Again some things that affect TTFB:
- Move your website to a faster/better server (If an option)
- Use a CDN or something similar to reduce the load on the server (repeated requests to a server will increase the TTFB)
- Reduce the time the server spends processing the request for information (sent above) and more here: http://stackoverflow.com/questions/10938682/how-to-reduce-server-wait-time
-
Thanks Vadim. Yes, this Cloudflare post is exactly the one I was referencing in my question. As I mentioned, Ilya Gregorik posted a rebuttal to their experiment here: post in Google+
It seems to me that if a Google developer says TTFB absolutely does matter that this would take precedence over anything Cloudflare might say.
What do you think?
-
Databases? Optimize any database queries that are slow This should help: http://www.techfounder.net/2011/03/25/database-profiling-and-optimizing-your-database-the-generic-version/
Now before you pass anything over to the IT this issue is a heated one in some cases where you have people saying that TTFB is not might not be the key metric to go after, here is more food for thought:
http://blog.cloudflare.com/ttfb-time-to-first-byte-considered-meaningles
"At CloudFlare we make extensive use of nginx and while investigating TTFB came across a significant difference in TTFB from nginx when compression is or is not used. Gzip compression of web pages greatly reduces the time it takes a web page to download, but the compression itself has a cost. That cost causes TTFB to be greater even though the complete download is quicker."
-
Thanks Vadim. This is helpful. In the first article the author writes:
"The only thing that is controllable is the server you are on." He suggests optimizing the database. What specific & measurable directive might I give to our IT manager that would accomplish this goal?
The second post looks very helpful indeed. I am downloading Microsoft's VRTA right now. It's a bit technically over my head, but I get the concepts. This should be something I can pass on to IT...however, it seems the info could be a bit dated (it repeatedly references IE 7)...Is there anything additional that might be more current?
Thanks again!
-
Hi Dana,
Yes TTFB is something you can control with the type of server you use. And where that server is in relation to your visitors. You cannot control the browsers they use, but hear are some thoughts on possible optimizations:
Server side: http://createdevelop.blog.com/2010/10/12/how-to-reduce-time-to-first-byte/
Location (plus other suggestions): http://msdn.microsoft.com/en-us/magazine/dd188562.aspx
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate content across different domains
Hi Guys, Looking for some advice regarding duplicate content across different domains. I have reviewed some previous Q&A on this topic e.g. https://moz.com/community/q/two-different-domains-exact-same-content but just want to confirm if I'm missing anything. Basically, we have a client which has 1 site (call this site A) which has solids rankings. They have decided to build a new site (site B), which contains 50% duplicate pages and content from site A. Our recommendation to them was to make the content on site B as unique as possible but they want to launch asap, so not enough time. They will eventually transfer over to unique content on the website but in the short-term, it will be duplicate content. John Mueller from Google has said several times that there is no duplicate content penalty. So assuming this is correct site A should be fine, no ranking losses. Any disagree with this? Assuming we don't want to leave this to chance or assume John Mueller is correct would the next best thing to do is setup rel canonical tags between site A and site B on the pages with duplicate content? Then once we have unique content ready, execute that content on the site and remove the canonical tags. Any suggestions or advice would be very much appreciated! Cheers, Chris
Intermediate & Advanced SEO | | jayoliverwright0 -
We used to speak of too many links from same C block as bad, have CDN's like CloudFlare made that concept irrelevant?
Over lunch with our head of development, we were discussing the way CloudFlare and other CDN's help prevent DDOS attacks, etc. and I began to wonder about the IP address vs. the reverse proxy IP address. Before we would look to see commonalities in the IP as a way that search engines would modify the value to given links and most link software showed this. For ahrefs, I know they still show common IPs using the C block as the reference point. I began to get curious about what was the real IP when our head of dev said, that is the IP from CloudFlare... So, I ran a site in ahrefs and we got an older site we had developed years ago that showed up as follows: Actos-lawsuit.org 104.28.13.57 and again as 104.28.12.57 (duplicate C block is first three sets of numbers are the same and obviously, this has a .12 and a .13 so not duplicate.) Then we looked at our host to see what was the IP shown there: 104.239.226.120. So, this really begs a question of is C Block data or even IP address data still relevant with regard to links? What do the search engines see when they look for IP address now? Yes, I have an opinion, but would love to hear yours first!
Intermediate & Advanced SEO | | RobertFisher0 -
Changing Domain Host and Potential SEO Effects
Hi Community, We are currently thinking of changing our domain hosting package and I am wondering if that could have any effects on rankings, SEO, etc. The thing here is that we aren't changing the hosting provider, we are only changing the package from the same provider. Most of the info I have found on the internet has been mostly about changing actual hosting providers versus packages from the same provider (if the makes sense). If anyone knows if this might affect our SEO and what steps we can take to safe guard against any potential pitfalls, please let us know. Thanks
Intermediate & Advanced SEO | | Brian_Dowd0 -
Leverage Browser Caching: Do I need Last-Modified?
Per Page Speed recommendations I specified the Expires header in my .htaccess file. Do I need to add code for Last-Modified too? I thought I read somewhere that it will put the date next to the meta description in the SERPS, which might cause the result to seem outdated after a while. Are there any problems that could crop up if these aren't implemented correctly.
Intermediate & Advanced SEO | | kimmiedawn0 -
IP Address: Ownership Location Versus IP Resolve
We are a US based ecommerce company that recently switched hosting to a Canadian owned company. I was told we would have a US based IP address but noticed yesterday that the MOZ bar is listing my website, 1800doorbell.com as a Canadian company. I've researched this online and what's typically stated is that your IP location needs to be in the Geo area you serve. When I brought his up to my host they stated: "The location being reported by many of these tools will be the one from the WHOIS. Since our corporation is registered in Canada, it will return a matching result. You can verify the location of the address by issuing a traceroute and examining the location codes at the end of the traceroute. For example, on: 96.125.180.207" So now I am really confused. What matters to me is how the search engines see my IP address. Will/do they see it as a US IP address? Below is the output from DNSstuff and thanks for any help: This is what I received back from DNSstuff: | ASN | 12179 |
Intermediate & Advanced SEO | | jake372
| Name | INTERNAP-2BLK |
| Description | - Internap Network Services Corporation |
| # Peers | 11 |
| # IPv4 Origin Ranges | 32 |
| # IPv6 Origin Ranges | 2 |
| Registrar | ARIN |
| Allocation date | Apr 13, 1999 |
| Country Code | US | | |
| Reverse | unknown.static.dal01.cologlobal.com. |
| Reverse-verified | No |
| Origin AS | - Internap Network S... |
| Country Code | CA |
| Country | Canada |
| Region | North America |
| Population | 31592805 |
| Top-level Domain | CA |
| IPv4 Ranges | 5944 |
| IPv6 Ranges | 336 |
| Currency | Canadian Dollar |
| Currency Code | CAD |
| IP Range - Start | 96.125.176.0 |
| IP Range - End | 96.125.191.255 |
| Registrar | ARIN |
| Allocation date | May 10, 2011 |0 -
Domain vs Subdomain for Multi-Location Practice
I have a client who has 2 locations (Orlando & Tampa) and would like to keep the current domain for both locations (DA 29). We want to target additional cities within each service area (Orlando & Tampa). Each service area would target 2 cities on the main pages and 4-5 cities with "SEO" pages which contains unique content specific to the given city. Would I be better off creating sub domains (www.orlando.domain.com & www.tampa.domain.com), creating subfolders (www.domain.com/orlando, etc) or keeping the domain as is and create SEO pages specific to each city? We want to spread the domain authority to both locations.
Intermediate & Advanced SEO | | Red_Spot_Interactive0 -
Looking for an open source way to host product reviews...
We currently use power reviews for our item reviews but want to move towards something that can be hosted internally and thus forgo the cost or power reviews and actually see user generated content play a role in our site. Any suggestions?
Intermediate & Advanced SEO | | MichealGooden0 -
Serving different content based on IP location
I have city centric website. For sake of simplicity, say I only have 2 cities -- City A and City B. Depending on a user's IP address, they will either get City A or City B. Users can change their location through javascript on pages. But there is no cross-linking between cities. By this, I mean that unless you can read or execute javascript, there is no way for you to get from city A to City B. My concern is this: googlebot comes to my site, and we serve them up City A. How does City B get discovered if Googlebot doesn't read javascript? We have an xml sitemap plus plenty of backlinks to City B. Is this sufficient? Should I provide a static link to City B (and vice versa) on the homepage for crawling purposes?
Intermediate & Advanced SEO | | ChatterBlock0