Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Page speed tools
-
Working on reducing page load time, since that is one of the ranking factors that Google uses. I've been using Page Speed FireFox plugin (requires FireBug), which is free. Pretty happy with it but wondering if others have pointers to good tools for this task.
Thanks...
-
If you're using wordpress or a CMS of some type, look for posts and tools related to that particular software. The SmushIt tool is available as a wordpress plugin, and it will compress pictures on upload, and will compress (on demand) pictures in the media library.
I use Irfanview for batch resizing images (in the same directory).
-
Have you considered using a CDN? This will cache your site in multiple locations so end users site experience will be better.
-
YSlow for firebug is very useful, don't forget to run SmushIt (Tools in YSlow) to have optimized images. Spites + GZip + 1 single packed CSS + 1 single packed JS + expires headers are very effective. Use media domain if you have a lot of pictures on the page.
See the real load time in the Net tab of firebug as Gary said.
-
I have gone through this same process recently.
FireBug's NET tab, the page speed plugin and Yahoo YSlow plugins were the way to go.
I had a page load time of about 10 seconds before optimising. I reduced this to about 6 seconds by doing the following:
- Using different domain names (e.g. static.mysite.com, static2.mysite.com, static3.mysite.com) to increase parallel downloads of CSS/images etc.
- Gzip
- Reducing size of all images
I then managed to decrease this further to about 1.5 seconds by changing server hosts.
-
decrease size of your pictures by 30% in most cases you wont even see the difference, but it helps a lot with load time.
Also check if you have any errors: http://validator.w3.org/ and http://jigsaw.w3.org/css-validator/
Good Luck.
-
yslow for firefox
and good old Google Webmaster Tools
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Finding issue on Gtmetrix speed test and google speed test
Hey, when i have tested my website https://www.socprollect-mea.com/ on GT metrix. When i test morning, it shown page speed of 3.8 seconds and in noon it shown 2 seconds and later i check, it is defeclecting. speed on the Google page speed test as well. What kind of error is this and does anyone have the solution for this issue?
On-Page Optimization | | nazfazy0 -
Page Title Length
Hi Gurus, I understand that it is a good practice is to use 50-60 characters for the a page title length. Google appends my brand name to the end of each title (15 characters including spaces) it index. Do I need to count what google adds as part of the maximum recommended length? i.e.
On-Page Optimization | | SunnyMay
is the maximum 50-60 characters + the 15 characters brand name Google adds to the end of the title or 50-60 including the addition? Many thanks!
Lev0 -
Why are http and https pages showing different domain/page authorities?
My website www.aquatell.com was recently moved to the Shopify platform. We chose to use the http domain, because we didn't want to change too much, too quickly by moving to https. Only our shopping cart is using https protocol. We noticed however, that https versions of our non-cart pages were being indexed, so we created canonical tags to point the https version of a page to the http version. What's got me puzzled though, is when I use open site explorer to look at domain/page authority values, I get different scores for the http vs. https version. And the https version is always better. Example: http://www.aquatell.com DA = 21 and https://www.aquatell.com DA = 27. Can somebody please help me make sense of this? Thanks,
On-Page Optimization | | Aquatell1 -
Should I optimize my home-page or a sub-page for my most important keyword
Quick question: When choosing the most important keyword set that I would like to rank for, would I be better off optimizing my homepage, or a sub page for this keyword. My thinking goes as follows: The homepage (IE www.mysite.com) naturally has more backlinks and thus a better Google Page Rank. However, there are certain things I could do to a subpage (IE www.mysite.com/green-widgets-los-angeles ) that I wouldn't want to do to the homepage, which might be more "optimal" overall. Option C, I suppose, would be to optimize both the homepage, and a single sub-page, which is seeming like a pretty good solution, but I have been told that having multiple pages optimized for the same keywords might "confuse" search engines. Would love any insight on this!
On-Page Optimization | | Jacob_A2 -
Punctuation at the Start of Page Titles
one of my clients appears to be using an exclamation mark (e.g. "! Graphic Prints By Mirrorin - Fun Childrens Graphic Prints") and to be completely honest, I have no idea if this is bad practice or if it wont have any affect from an SEO point of view? Any help would be appreciated because it is site wide, therefore if it is an issue I would like to be able to get it sorted asap! Thanks
On-Page Optimization | | ZaddleMarketing0 -
URL for location pages
Hello all We would like to create clean, easy URLs for our large list of Location pages. If there are a few URLs for each of the pages, am I right when I'm saying we would like this to be the canonical? Right now we would like the URL to be: For example
On-Page Optimization | | Ferguson
Domain.com/locations/Columbus I have found some instances where there might be 2,3 or more locations in the same city,zip. My conclusion for these would be: adding their Branch id's on to the URL
Domain.com/locations/Columbus/0304 Is this an okay approach? We are unsure if the URL should have city,State,zip for SEO purposes?
The pages will have all of this info in it's content
BUT what would be best for SEO and ranking for a given location? Thank you for any info!0 -
Best Practice for Deleting Pages
What is the best SEO practice for deleting pages? We have a section in our website with Employee bios, and when the employee leaves we need to remove their page. How should we do this?
On-Page Optimization | | Trupanion0 -
Avoiding "Duplicate Page Title" and "Duplicate Page Content" - Best Practices?
We have a website with a searchable database of recipes. You can search the database using an online form with dropdown options for: Course (starter, main, salad, etc)
On-Page Optimization | | smaavie
Cooking Method (fry, bake, boil, steam, etc)
Preparation Time (Under 30 min, 30min to 1 hour, Over 1 hour) Here are some examples of how URLs may look when searching for a recipe: find-a-recipe.php?course=starter
find-a-recipe.php?course=main&preperation-time=30min+to+1+hour
find-a-recipe.php?cooking-method=fry&preperation-time=over+1+hour There is also pagination of search results, so the URL could also have the variable "start", e.g. find-a-recipe.php?course=salad&start=30 There can be any combination of these variables, meaning there are hundreds of possible search results URL variations. This all works well on the site, however it gives multiple "Duplicate Page Title" and "Duplicate Page Content" errors when crawled by SEOmoz. I've seached online and found several possible solutions for this, such as: Setting canonical tag Adding these URL variables to Google Webmasters to tell Google to ignore them Change the Title tag in the head dynamically based on what URL variables are present However I am not sure which of these would be best. As far as I can tell the canonical tag should be used when you have the same page available at two seperate URLs, but this isn't the case here as the search results are always different. Adding these URL variables to Google webmasters won't fix the problem in other search engines, and will presumably continue to get these errors in our SEOmoz crawl reports. Changing the title tag each time can lead to very long title tags, and it doesn't address the problem of duplicate page content. I had hoped there would be a standard solution for problems like this, as I imagine others will have come across this before, but I cannot find the ideal solution. Any help would be much appreciated. Kind Regards5