Is the TTFB for different locations and browsers irrelevant if you are self-hosting?
-
Please forgive my ignorance on this subject. I have little to no experience with the technical aspects of setting up and running a server.
Here is the scenario:
We are self-hosted on an Apache server. I have been on the warpath to improve page load speed since the beginning of the year. I have been on this warpath not so much for SEO, but for conversion rate optimization. I recently read the Moz Post "How Website Speed Actually Impacts Search Rankings" and was fascinated by the research regarding TTFB. I forwarded the post to my CEO, who promptly sent me back a contradictory post from Cloudflare on the same topic. Ily Grigorik published a post in Google+ that called Cloudflare's experiment "silly" and said that "TTFB absolutely does matter."
I proceeded to begin gathering information on our site's TTFB using data provided by http://webpagetest.org. I documented TTFB for every location and browser in an effort to show that we needed to improve. When I presented this info to my CEO (I am in-house) and IT Director, that both shook their heads and completely dismissed the data and said it was irrelevant because it was measuring something we couldn't control.
Ignorant as I am, it seems that Ilya Grigorik, Google's own Web Dev Advocate says it absolutely is something that can be controlled, or at least optimized if you know what you are doing.
Can any of you super smart Mozzers help me put the words together to express that TTFB from different locations and for different browsers is something worth paying attention to? Or, perhaps they are right, and it's information I should ignore?
Thanks in advance for any and all suggestions!
Dana
-
Yes, very helpful guys. I appreciate it!
-
Thanks Igal and hopefully you have some info to work with Dana!
-
Many thanks to both Vadim and Igal for such great information and also a really great thread on the subject. I really, really appreciate your answers.!
-
Honestly, I don't know. I don't think TTFB was ever comparatively tested - at least no to the best of my knowledge.
For security, these are some of the resources I can point to.
I understand that this is not the main issue
Still, I wanted to provide some factual context to my previous statements.
http://zeroscience.mk/files/wafreport2013.pdf http://ddos-protection-services-review.toptenreviews.com/ http://tonyonsecurity.com/2012/11/13/protecting-your-website-cloudflare-or-incapsula/
(This last one is interesting since Tony is a COO of Sucuri. Some would call his our competitor. I prefer 'colleague' )
-
security wise it seems both of you guys have stellar options. for me the issue is performance, caching for dynamic sites, CDN performance, and in this case TTFB response. I was not sure with your response do you have faster TTFB to CF?
Thanks
-
Hi Vadim
Thanks.
Yep, I work for Incapsula but no, we are not the said "Mod".As for CF comparison... Generally speaking, we are more business oriented and security focused. I know that our security offering is more comprehensive, especially because both WAFs were comparatively pen-tested on several occasions and we always came out as consistently (and significantly) better option.We also have addition security features - like 2FA support and backdoor shell protection - which CF simply doesn't offer and we do more in way of ddos mitigation, especially against smart application layer attacks which require security capabilities, besides network muscle.
Still, speed wise, I always considered us to be pretty much on the same level. However, until few days ago I never considered TTFB to be such core SEO factor, so maybe we have better performance there...
But again, to be fair, I`m only speculating - mostly based on the CF blog you've shared.
(if TTFB is considered un-important, it might also be under developed...)Might be an interesting thing to test and document.
-
Hi Igal,
Do you work for incapsula, you are mentioned as a Mod on the blog?
I have heard great things about incapsula from others, but in terms of TTFB is it better than cloudflare? If so, how so?
Also any other ways that it excels Cloudflare? any ways its inferior to Cloudflare in your opinion?
Thanks I am really looking for more info, as I had great results with Cloudflares features and offering, wondering if I should give Incapsula a run
Thanks
-
I absolutely agree with Vadim. (+1)
Google is the best source for Google facts. Everything else is just speculation.
And yes, generally speaking, the best answer is to use a CDN....
The reason is simple. CNDs proxy technology, which was designed to minimize "physical" distances between the site's content and browsers, directly influences TTFB.Being an in-house SEO for a CDN company I get a lot of questions about this from our support and clients. I have to admit, until recent Moz post, I wasn't aware of full implications of TTFB and considered it to be one of few page load speed related metrics. (http://moz.com/blog/how-website-speed-actually-impacts-search-ranking)
This post really helped me get a better grasp on things. Interestingly enough, few month ago one of our clients Guest Posted in our blog about speed improvement gained by our free plan. Among other things, he mentioned 70% improvement in TTFB (grade going from F to A)
(http://www.incapsula.com/the-incapsula-blog/item/718-what-incapsula-free-did-for-my-site)At the time I didn't give it much attention. Because, like many others, I was focusing on overall load speeds....
Now I can't help but feel that this was a missed opportunity.
This post could be even better with the added SEO angle...
If anyone here is interested in giving this a try and guest posting about it, I`ll be happy to provide all resources needed on our end. -
Yea this makes sense as others have said that Cloudflare is trying to say that TTFB is not the most important metric, and so they published this study, as it aids their business model.
I would do just that listen to Google dev vs Cloudflare. Also the way I think about it even if their studies are true, where for the overall benefit TTFB would have to increase if you are using some compression, you still need to work and decrease your TTFB either way, that is just intuition. I apologize if I made it seem that TTFB is to be ignored, because Cloudflare state's that quite boldly,
Again some things that affect TTFB:
- Move your website to a faster/better server (If an option)
- Use a CDN or something similar to reduce the load on the server (repeated requests to a server will increase the TTFB)
- Reduce the time the server spends processing the request for information (sent above) and more here: http://stackoverflow.com/questions/10938682/how-to-reduce-server-wait-time
-
Thanks Vadim. Yes, this Cloudflare post is exactly the one I was referencing in my question. As I mentioned, Ilya Gregorik posted a rebuttal to their experiment here: post in Google+
It seems to me that if a Google developer says TTFB absolutely does matter that this would take precedence over anything Cloudflare might say.
What do you think?
-
Databases? Optimize any database queries that are slow This should help: http://www.techfounder.net/2011/03/25/database-profiling-and-optimizing-your-database-the-generic-version/
Now before you pass anything over to the IT this issue is a heated one in some cases where you have people saying that TTFB is not might not be the key metric to go after, here is more food for thought:
http://blog.cloudflare.com/ttfb-time-to-first-byte-considered-meaningles
"At CloudFlare we make extensive use of nginx and while investigating TTFB came across a significant difference in TTFB from nginx when compression is or is not used. Gzip compression of web pages greatly reduces the time it takes a web page to download, but the compression itself has a cost. That cost causes TTFB to be greater even though the complete download is quicker."
-
Thanks Vadim. This is helpful. In the first article the author writes:
"The only thing that is controllable is the server you are on." He suggests optimizing the database. What specific & measurable directive might I give to our IT manager that would accomplish this goal?
The second post looks very helpful indeed. I am downloading Microsoft's VRTA right now. It's a bit technically over my head, but I get the concepts. This should be something I can pass on to IT...however, it seems the info could be a bit dated (it repeatedly references IE 7)...Is there anything additional that might be more current?
Thanks again!
-
Hi Dana,
Yes TTFB is something you can control with the type of server you use. And where that server is in relation to your visitors. You cannot control the browsers they use, but hear are some thoughts on possible optimizations:
Server side: http://createdevelop.blog.com/2010/10/12/how-to-reduce-time-to-first-byte/
Location (plus other suggestions): http://msdn.microsoft.com/en-us/magazine/dd188562.aspx
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Location Links in Footer
Our business is in 10 cities. We offer identical services in each city, there's absolutely nothing different about the services we offer based on location. We have a contact page for each city with a bit of unique content (phone, address, photo of city, list of counties we service). It really would be a grey area to create subsites for each city and try to rewrite the service description content 10 times. However, we want to improve organic results. We of course have Google Places listings for each city. From an on-page SEO perspective, wouldn't it only have the possibility of benefiting, not hurting local SEO but add the city name linked to that city's contact page in the footer? I've seen arguments against it, and could see maybe if you were in like 50 cities instead of 10, but is there really any observed downside to doing that in the footer for every page? We can't title the difference service pages with the city name in the headings or page title, so at least we'd have anchor text in the footer.
Intermediate & Advanced SEO | | Wizkids9640 -
Self referencing canonicals and paginated content - advice needed
Hi, I help manage a large site that uses a lot of params for tracking, testing and to help deal with paginated content e.g. abc.com/productreview?page=2. The paginated review content correctly uses rel next and rel prev tags to ensure we get the value of all of the paginated review content that we have. The volume of param exclusions I need to maintain in Google & Bing Webmaster tools is getting clunky and frustrating. I would like to use self referencing canonicals, which would make life a lot easier. Here's my issue: If I use canonicals on the review pages the paginated content urls would also use the same canonical e.g. /productreview?page=2 pointing to /productreview I believe I am going to lose the value of those reviews, even though they use the rel next rel prev tags. BTW airbnb do this - do they know something I don't, don't care about the paginated reviews, or are they doing it incorrectly, see http://d.pr/i/14mPU Is my assertion above correct about losing the value of the paginated reviews if I use self referencing canonicals? Any thoughts on a solution to clearing up the param problem or do I have to live with it? Thanks in advance, Andy
Intermediate & Advanced SEO | | AndyMacLean0 -
What's the best way to A/B test new version of your website having different URL structure?
Hi Mozzers, Hope you're doing good. Well, we have a website, up and running for a decent tenure with millions of pages indexed in search engines. We're planning to go live with a new version of it i.e a new experience for our users, some changes in site architecture which includes change in URL structure for existing URLs and introduction of some new URLs as well. Now, my question is, what's the best way to do a A/B test with the new version? We can't launch it for a part of users (say, we'll make it live for 50% of the users, an remaining 50% of the users will see old/existing site only) because the URL structure is changed now and bots will get confused if they start landing on different versions. Will this work if I reduce crawl rate to ZERO during this A/B tenure? How will this impact us from SEO perspective? How will those old to new 301 URL redirects will affect our users? Have you ever faced/handled this kind of scenario? If yes, please share how you handled this along with the impact. If this is something new to you, would love to know your recommendations before taking the final call on this. Note: We're taking care of all existing URLs, properly 301 redirecting them to their newer versions but there are some new URLs which are supported only on newer version (architectural changes I mentioned above), and these URLs aren't backward compatible, can't redirect them to a valid URL on old version.
Intermediate & Advanced SEO | | _nitman0 -
Can you use multiple rel alternate tags for different device subdomains?
When redirecting from desktop to mobile with a separate URL structure, you need to have a rel alternate - rel canonical handshake to define the relationship between the pages. But if you have a different subdomain for different mobile devices, can you add more than one rel alternate tag on the desktop page? EG if site.com is redirecting to iphone.site.com, m.site.com, android.site.com
Intermediate & Advanced SEO | | AdiRste0 -
Profit from domain authority, while shop hosted on other server
Hi, What if you have a blog and you want to add a shop to your site (like http://blog.com/shop) to profit from the domain autority, but... the shop uses a specific technique that needs to be hosted on an other server? I know a solution when I locate the shop at http://shop.blog.com by changing the DNS settings, but in this case the shop will not profit from the domain autoriry. So,.. is there a technique (maybe also with DNS?) to host the shop on a different location, but for search engines and visitors it looks like it is on the same server? I hope somebody knows a solution...
Intermediate & Advanced SEO | | Seeders0 -
Create different pages with keyword variations VS. Add keyword variations in 1 page
For searches involving keywords like "lessons", "courses", "classes" I see frequently pages in the top rankings which do not contain the search term in the title tag, despite these terms being quite competitive. It seems that when searching for "classes", google detects that pages about "courses" may be just as relevant. What do you recommend? option 1: creating 10 pages optimized on 10 different keyword variations, each with a significant part of unique content or option 2: one page and dropping throughout the page 10 keyword variations in body and headlines Given that keywords are all synonyms and website has already high domain authority in the niche. thanks
Intermediate & Advanced SEO | | lcourse0 -
Different pages ranking for search terms, often irrelevant.
Website: www.templatemonster.com
Intermediate & Advanced SEO | | templatemonster
Problem: Positions dropped while pages which were ranking previously disappeared from top 100 and now different - often completely irrelevant - pages are ranking. Examples:
Search term: Joomla Templates
Previous Position: 8
Current Position: 35
Previously Ranked Page: http://www.templatemonster.com/joomla-templates.php
Currently Ranked Page: http://www.templatemonster.com/logo-templates.php Similar situation with the following search terms: virtuemart templates, virtuemart themes, prestashop templates, prestashop themes, magento themes, zencart templates, zencart themes, zen cart templates, zen cart themes When: according to the Google Analytics (drop in visitors stats) this happened on July, 2nd Preconditions: we had 45 minutes downtime on July 2-nd - but could this 45 mins have had such disastrous results?
No redirects or canonical URL were used which could lead to such change of ranking page.
No changes in the site's informational structure and design.
In webmaster tools (inbound links report) we saw a website yesterday which had over 800,000 links pointing to our domain - http://moviebestwatch.com/ - and today this site is NOT found in Webmaster Tools report! Also, site is down, domain is quite new (how could it have possibly developed 800,000 pages in such a short time?) and whois is privacy protected. Is this some dirty trick from competitors - could it have possibly influenced our positions? Still, what I completely fail to understand - how could a page like http://www.templatemonster.com/logo-templates.php be the top ranking page for 'Joomla templates' if there is: not a single mention of the word 'Joomla' on the page (or source code), i.e. the page is completely irrelevant to the search term not a single link with 'Joomla templates' anchor text pointing to that page, neither external nor internal PS. No similar changes in other search engines noticed. Also, the pages in question have been re-spidered July 4th and cache shows the right pages, i.e. it is not that Googlebot has seen logotypes page instead of Joomla templates page. I checked any possible reason I could think of (see "Preconditions") but still have no clue - what is going on?1 -
Keyword Self-Cannibalization Concern
Right now I have an E-Commerce website that has layered menu properties. I have one page trying to rank for "NextGen Digital Ballast" that is the main category page. However, on that category page I link out to three product pages which would be "NextGen 400W Digital Ballast", "NextGen 600W Digital Ballast" and "NextGen 1000W Digital Ballast". The on page ranking factors tools is saying I may need to consider making adjustments because of the potential self-cannibalization, but I wanted to get some feedback to see what others thought about that. Thanks.
Intermediate & Advanced SEO | | JerDoggMckoy2