Good technical parameters worst load time.
-
I have recently created a page and added expires headers, nonconfigured e-tags and gzip to htaccess code and just after that according to pingdom tools my page load time has doupled although my yslow ponts went from 78 to 92. I always get a lite bit lost with this technical issue. I mean obviously a site should not produce worse results with adding these parameters and this increase in page load time should rather be due to bandwith usage. I suppose I should leave this stuff in the htacces. Than what is an accurate way to know if you have done a real improvement to your site or your load time has really went up?
This question is more up to date with css sprites as I always read that sometimes spriting every picture is a waste of resources. How can you decide when to stop?
-
MY page is a basic html page. I have already rewrited code, there are a handful of dom elemnts, css files sprited etc. PAge load time went from 230 milisec to 500 when I implemented the new features.
-
I think this should be the case, my page is a basic html at around 200kb. Thanks for the answer.
-
It takes time to compress and decompress a page, for a litwieght page compression can actualy take longer.
If you have a heavy page then compression can be a good thing, but if your page is light then it can work against you.
On a windows server you can tell it how big a file has to be before it is compressed, The default is 256b. Thats should tell you somthing
You can also cache the compressed your static files so that compression is not needed the next time.
-
Can you share the load time that you got before and after working on those technical parameters? In our website, we usually use the webmaster tools and at the same time compare it to our competitor. For example we are in the hotel business so we try to compare our site performance to the biggest hotel chains.
But then in my opinion once you worked on those technical parameters, there are still other aspect of your site that you need to check to increase the load performance.
1. Check the size of your page. Initially, our site loads around 10 secs and this is because of our layout and we use a lot of images. First step is we compress all our jpgs without degrading the quality of it. The second part is a reconstruct the layout or our scripts to reduce the DOM elements in our site. I notice a difference on the load time when our DOM elements is less than 1000
2. For sprite images. It will be better to create a sprite image upon development instead of spriting them on the fly which is why I think they said that it is a waste of resources. I use this site for spriting our images. http://spritegen.website-performance.org/
3. You need to minify and combined all your css and javascript.
I also follow all those rules in YSLOW and Page Speed and I can see a significant improvement in our page load time. Without using a CDN our site now loads around 4-5 secs and with CDN, I think around 3-4 secs.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Technical SEO - Where to begin?
Hi all, I'm looking to learn more about technical SEO. My background was digital marketing/PR where I learned the importance of links, of anchor text, of page speed, of improving UX signals, of SSL, utilising things like Google My Business etc. However, I find I am chasing my tail when it comes to things like understanding JS/CSS/log file analysis etc. I've tried reading so many articles on the subjects and I just find it so damn confusing. AnugalarJS/BackboneJS. Fetching & rendering, URL parameters...etc. I know from my own experiments that JS pages struggle to rank and I've created two very similar pages, one without JS, one with JS (which had far more links) and the non-JS page ranked far higher. So, I suppose I'm asking for some help with how to begin learning this stuff. I find the articles on Moz, Search Engine Land etc to be a bit confusing...maybe I'm not technically minded enough! Cheers, Rhys
Technical SEO | | SwanseaMedicine0 -
URL Parameters as pagination
Hi guys, due to some changes to our category pages our paginated urls will change so they will look like this: ...category/bagger/2?q=Bagger&startDate=26.06.2017&endDate=27.06.2017 You see they include a query parameter as well as a start and end date which will change daily. All URLs with pagination are on noindex/follow. I am worrying that the products which are linked from the category pages will not get crawled well when the URLs on which they are linked from change on a daily basis. Do you have some experience with this? Are there other things we need to worry about with these pagination URLs? cheers
Technical SEO | | JKMarketing0 -
How can I Improve Loading Speed? - Parker Dubrule Lawyers
Parker Dubrule Lawyers' website at parkerdubrulelawyers.com seemed to be loading quite slow this morning (>5 Seconds). I added a lazyload plugin, minified JS and CSS, and ensured that the images were optimized---all of this seemed to help and brough it down to under 2 seconds. We are looking at more reliable hosting options for our clients---ones that are inherently faster possibly without these plugins being added to the mix. Does anyone have insight on a safe, secure, and fast hosting/server option to enhance the experience from the get go? All of the websites that we build are in Worpress. Your help is much appreciated! Thanks!
Technical SEO | | Web3Marketing870 -
Parameter Handling - Nourls Question
We're trying to make sense of Google's new parameter handling options and I seem unable to find a good answer to an issue regarding the NoUrl option. For ex. we have two Urls pointing to the same content: http://www.propertyshark.com/mason/ny/New-York-City/Maps/Manhattan-Apartment-Sales-Map?zoom=1&x=0.518&y=0.3965 http://www.propertyshark.com/mason/ny/New-York-City/Maps/Manhattan-Apartment-Sales-Map?zoom=2&x=0.518&y=0.3965 Ideally, I would want Google to index only the main Url without any parameters, so http://www.propertyshark.com/mason/ny/New-York-City/Maps/Manhattan-Apartment-Sales-Map To do this, I would set the value No Urls for the zoom, x and y parameters. By doing this do we still get any SEO value from back links that point to the URLs with the parameters, or will Google just ignore them?
Technical SEO | | propertyshark0 -
Updating rss feed times without changing content
my question is like the title reads If I have an rss feed in an xml file and from time to time I update the pubdate and time. Will this have a positive effect on my website in terms of the rss aggregators coming to my site thinking that it was recently updated and creating links to these pages or will they be able to determine that there is nothing new by comparing it to the old page that they may have stored. thus doing nothing or maybe even hurting the website.
Technical SEO | | mickey112 -
On a dedicated server with multiple IP addresses, how can one address group be slow/time out and all other IP addresses OK?
We utilize a dedicated server to host roughly 60 sites on. The server is with a company that utilizes a lady who drives race cars.... About 4 months ago we realized we had a group of sites down thanks to monitoring alerts and checked it out. All were on the same IP address and the sites on the other IP address were still up and functioning well. When we contacted the support at first we were stonewalled, but eventually they said there was a problem and it was resolved within about 2 hours. Up until recently we had no problems. As a part of our ongoing SEO we check page load speed for our clients. A few days ago a client who has their site hosted by the same company was running very slow (about 8 seconds to load without cache). We ran every check we could and could not find a reason on our end. The client called the host and were told they needed to be on some other type of server (with the host) at a fee increase of roughly $10 per month. Yesterday, we noticed one group of sites on our server was down and, again, it was one IP address with about 8 sites on it. On chat with support, they kept saying it was our ISP. (We speed tested on multiple computers and were 22MB down and 9MB up +/-2MB). We ran a trace on the IP address and it went through without a problem on three occassions over about ten minutes. After about 30 minutes the sites were back up. Here's the twist: we had a couple of people in the building who were on other ISP's try and the sites came up and loaded on their machines. Does anyone have any idea as to what the issue is?
Technical SEO | | RobertFisher0 -
Is there any issue with using the same structured data property multiple times on the same page?
Im working on implementing structured data properties into my product detail pages. (http://schema.org/Book) My site sells books and many books have both a 13 digit ISBN # and a 10 Digit ISBN. Should I apply the itemprop "isbn" to both of them or just the one with higher search volume? Some books also have multiple authors, how should I handle that?
Technical SEO | | myork07240 -
My website pages (www.ommrudraksha.com) is getting good rank slowly. But no good sales ?
My website has been doing good slowly. I have been using seomoz recommendations. And it is a great help to see that my pages are slowly coming to the first page. I am also running PPC on google. I see there are many visitors to my website. But i do not get good conversion - or not getting customer buying products. My website : www.ommrudraksha.com My target keyword is : rudraksha
Technical SEO | | Ommrudraksha0