Javascript to manipulate Google's bounce rate and time on site?
-
I was referred to this "awesome" solution to high bounce rates.
It is suppose to "fix" bounce rates and lower them through this simple script. When the bounce rate goes way down then rankings dramatically increase (interesting study but not my question).
I don't know javascript but simply adding a script to the footer and watch everything fall into place seems a bit iffy to me.
Can someone with experience in JS help me by explaining what this script does?
I think it manipulates the reporting it does to GA but I'm not sure. It was supposed to be placed in the footer of the page and then sit back and watch the dollars fly in.
-
Stephen,
Thanks for the explanation - I just had a client ask me about this script. Based on your explanation, this script will change your bounce rate. This is because once the event is triggered, the visit will no longer be considered a bounce, even if the user only visits one page. So it's an artificial/false decrease in bounce rate, not a "fix" as others claim.
I wrote a short blog post on this (and referenced your description)!
~Adam
-
Thanks for the encouragement Martin.
As it turns out, with the help of the two previous answers, the script is actually based on a valid script adjustment that might actually help some people in their reports but the what my client thought was that this was an easy/quick way to get more traffic. The article they found was saying this would dramatically change results in GA and then directly effect their site's ranking in the SERPs.
They had "proof" in the form of some GA screenshots so I needed more information on what the script actually does. I was able to let my client know what exactly this was and recommend not doing it unless there was a problem in the GA reports that they wanted fixed.
Thanks again for your reply.
-
Dont do it - just improve your content. You know it's wrong to try and cheat the system. Think about what would happen if you banned from the results.
Look i dont mean to be harsh - but i allways balance risks against rewards. In this situation - the risk is to high.
-
Thanks for that link.
The site (link in the previous reply) my client referred me to was manipulating the way they were reporting the results. The closer I looked at it, I realized that it was a little spike but then it went right back down. Knowing them they just paid a bunch of people to visit the site.
This stuff is annoying and gives us SEO's a bad name.
-
The code was from this site http://millionairevolution.com/cut-bounce-rate-by-80/ and looking at the dates and analytics shown on the page this is nothing more than a misrepresentation of the facts and data.
I knew Google doesn't use data from GA but the data graph was showing a contradiction and I didn't know exactly what the script was doing.
-
First, Google Analytics reporting does not, to my knowledge, influence SERP rankings. Altering the data collected through Google Analytics should not affect SEO indicators.
Second, this is from here: http://briancray.com/posts/time-on-site-bounce-rate-get-the-real-numbers-in-google-analytics/
Once this code is installed, your site will update Google Analytics every 10 seconds under the Event Category "Time", the Event Action "Log", and the Event Value will be based on the pattern of 0:10, 0:20, 0:30, 0:40, 0:50, 1:00, 1:10, etc.
The script does not change your bounce rate, it just gives you additional information.
-
You're correct that it's a GA hack. Avoid it.
Google has publicly stated that they don't use your site-specific GA metrics to influence organic search rankings. E.g., they're not taking data from your GA profile, and feeding that to the Search Quality team to determine if your site should rank better or worse. They have MANY better ways to accurately track anonymous user interactions with sites at scale (e.g. Chrome).
The only thing that you'll accomplish with this code is making all of your own internal metrics turn to garbage. Accurate metrics are important. If you bounce rate is high, knowing that allows you to take action to improve your site and reduce it.
The more people who stay on your site for more than 1 pageview, the more money your business is likely to make. Improve your bounce rate to improve the profitability of your website, not for some supposed correlation between bounce rate and organic search ranking.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Spam URL'S in search results
We built a new website for a client. When I do 'site:clientswebsite.com' in Google it shows some of the real, recently submitted pages. But it also shows many pages of spam url results, like this 'clientswebsite.com/gockumamaso/22753.htm' - all of which then go to the sites 404 page. They have page titles and meta descriptions in Chinese or Japanese too. Some of the urls are of real pages, and link to the correct page, despite having the same Chinese page titles and descriptions in the SERPS. When I went to remove all the spammy urls in Search Console (it only allowed me to temporarily hide them), a whole load of new ones popped up in the SERPS after a day or two. The site files itself are all fine, with no errors in the server logs. All the usual stuff...robots.txt, sitemap etc seems ok and the proper pages have all been requested for indexing and are slowly appearing. The spammy ones continue though. What is going on and how can I fix it?
Technical SEO | | Digital-Murph0 -
Google Seeing Way More Pages Than My Site Actually Has
For one of my sites, A-1 Scuba Diving And Snorkeling Adventures, Google is seeing way more pages than I actually have. It sees almost 550 pages but I only have about 50 pages in my XML. I am sure this is an error on my part. Here is the search results that show all my pages. Can anyone give me some guidance on what I did wrong. Is it a canonical url problem, a redirect problem or something else. Built on Wordpress. Thanks in advance for any help you can give. I just want to make sure I am delivering everything I can for the client.
Technical SEO | | InfinityTechnologySolutions0 -
Google Cache can't keep up with my 403s
Hi Mozzers, I hope everyone is well. I'm having a problem with my website and 403 errors shown in Google Webmaster Tools. The problem comes because we "unpublish" one of the thousands of listings on the site every few days - this then creates a link that gives a 403. At the same time we also run some code that takes away any links to these pages. So far so good. Unfortunately Google doesn't notice that we have removed these internal links and so tries to access these pages again. This results in a 403. These errors show up in Google Webmaster Tools and when I click on "Linked From" I can verify that that there are no links to the 403 page - it's just Google's Cache being slow. My question is a) How much is this hurting me? b) Can I fix it? All suggestions welcome and thanks for any answers!
Technical SEO | | HireSpace1 -
Page Indexing increase when I request Google Site Link demote
Hi there, Has anyone seen a page crawling increase in Google Web Master Tools when they have requested a site link demotion? I did this around the 23rd of March, the next day I started to see page crawling rise and rise and report a very visible spike in activity and to this day is still relatively high. From memory I have asked about this in SEOMOZ Q&A a couple of years ago in and was told that page crawl activity is a good thing - ok fine, no argument. However at the nearly in the same period I have noticed that my primary keyword rank for my home page has dropped away to something in the region of 4th page on Google US and since March has stayed there. However the exact same query in Google UK (Using SEOMOZ Rank Checker for this) has remained the same position (around 11th) - it has barely moved. I decided to request an undemote on GWT for this page link and the page crawl started to drop but not to the level before March 23rd. However the rank situation for this keyword term has not changed, the content on our website has not changed but something has come adrift with our US ranks. Using Open Site Explorer not one competitor listed has a higher domain authority than our site, page authority, domain links you name it but they sit there in first page. Sorry the above is a little bit of frustration, this question is not impulsive I have sat for weeks analyzing causes and effects but cannot see why this disparity is happening between the 2 country ranks when it has never happened for this length of time before. Ironically we are still number one in the United States for a keyword phrase which I moved away from over a month ago and do not refer to this phrase at all on our index page!! Bizarre. Granted, site link demotion may have no correlation to the KW ranking impact but looking at activities carried out on the site and timing of the page crawling. This is the only sizable factor I can identify that could be the cause. Oh! and the SEOMOZ 'On-Page Optimization Tool' reports that the home page gets an 'A' for this KW term. I have however this week commented out the canonical tag for the moment in the index page header to see if this has any effect. Why? Because as this was another (if not minor) change I employed to get the site to an 'A' credit with the tool. Any ideas, help appreciated as to what could be causing the rank differences. One final note the North American ranks initially were high, circa 11-12th but then consequently dropped away to 4th page but not the UK rankings, they witnessed no impact. Sorry one final thing, the rank in the US is my statistical outlier, using Google Analytics I have an average rank position of about 3 across all countries where our company appears for this term. Include the US and it pushes the average to 8/9th. Thanks David
Technical SEO | | David-E-Carey0 -
Keyword Variants - Low conversion rates - Is a site redesign required?
Hi Guys, I'm Chris & i'm a noob to SEO. Thanks for taking the time to read this and for offering your support! It is greatly appreciated! I work for a small family run company called custom designed cables and we manufacture bespoke cable and bespoke retractable cables. I posted a couple of days ago about how to move forward with our site. Our website has been targeting specific keywords and their variants across the relevant pages to somewhat reasonable effect. Since signing up to SEOMOZ, i have been trying to narrow down the keywords that we target per page and also tried to improve the content on the pages with the help of the on page optimisation tool. I have a few questions that i would like somebody to help me with please if it isn't to much trouble. Firstly, for our keywords, there are a lot of variants. For example, we braid cables. Here are some of the variants of this keyword: cable braid, cable braiding, cable screen, cable screening, cable shielding etc etc. The way i have gone in the past has been to try and target all of those keywords on the relevant page. Now, writing content has been the issue as whilst trying to write good informative information, squeezing in all of the keywords has been a problem and more than likely affected the readability of the page. From what i have learnt since i signed up, this is not a good practice and therefore i have tried to narrow it down somewhat yet i don't really want to lose potential customers finding us by only targeting one or two of the keywords. This is a similar situation for most of our keywords on the majority of our pages. What is the best way to approach this? If i was to write a page per keyword, i don't believe that it would look very good as we could end up with over 100 pages with say 5 or 6 of them talking about the same subject which then leads onto the problem of writing good content. It would be difficult to write good content for pretty much the same thing across such a wide number of pages. If this is the solution though then i would more than happily tackle it. What would be the best way to move forward? The next issue that I have is that since i have been modifying the pages with the help of SEOMOZ, the number of enquries that we have received have fell off the charts. We used to average around 50 enquiries a month from the website and since i have modified the site, i'd say we've had probably 20 since the end of February. The funny thing is though, we have also averaged around 200 hits more per month since the changes. So the hits have increased yet the enquiries have died. I was wondering if anybody was willing to take a look at our site: http://www.customdesignedcable.co.uk to give me some general feedback as to why this may be happening and what your overall opinions of the site may be with regards to the layout, look and general feel etc because i am beginning to wonder whether or not the design is what could be causing us to convert so little of our new found visitors. If anyone could also provide me with actionable feedback with regards to our keyword targeting for our pages that would also really really help! I am currently considering re-designing the site in it's entirety and i am interested in your opinions on whether or not this would be a good way to go. Any help that you can give me would be greatly appreciated guys! Thanks again! Chris (Just another website/seo noob desperately attempting to avoid the sack)
Technical SEO | | Chris_CDC0 -
For Google + purposes, should the author's name appear in the Meta description or title tag of my web site just as you would your key search phrase?
Relative to Cyrus Shepard's article on January 4th regarding Google's Superior SEO strategy, if I'm the primary author of all blog articles and web site content, and I have a link showing authorship going back to Google Plus, is a site wide link from the home page enough or should that show up on all blog posts etc and editorial comment pages etc? Conversely, should the author's name appear in the Meta description or title tag of my web site just as you would your key search phrase since Google appears to be trying to make a solid connection with my name, and all content?
Technical SEO | | lwnickens0 -
It's imposible to keep track of rankings?
Hello, here something interesting I'm Using Rank Tracker from SEOMOZ And from the link-assistant's Rank Tracker as Well... I need to track Google.com and Google.co.ve (venezuela) so I did... i got my keyword an here are my results. 1 Keyword A at google.com (united states) Rank Tracker SEOMOZ = pos 6 Rank Tracker OTHER = pos 6 Manual Query on google.com = 9 (I used the exact url seomoz tells me its using) 2 Keyword A at Google.co.ve Rank Tracker SEOMOZ = pos 8 Rank Tracker OTHER = pos 7 Manual query on google.co.ve = pos 8 So.... Why it's that?, so far I think that google.com for me down here (it actually says "Español") it's a different index? for latinamerica? only spanish pages? maybe it's because there's a couple of minutes between looking with one tool and the other... any help, would be great... Dan
Technical SEO | | daniel.alvarez0 -
A question about RSS feeds and nofollow's
With the nofollow tag used very widely on the internet these days I was just wondering about how an RSS feed might help me find a way around it. Basically my question is this : I post a comment on a blog, it's approved and my comment together with my link(nofollow tag applied) is there. Now when the blogs RSS feed updates, does this nofollow tag get applied to the feed? As far as I can tell it does not - but I'm not too clue'd up on how the feed is generated. Anyone want to help me understand how it works and if what I'm suggesting would be 'a way around the nofollow tag' ? Thanks 🙂
Technical SEO | | DanHill0