Estimating the number of LRD I need to outrank competitor
-
I just ran a SERP/keyword diffculty report for a keyword I want one of my pages to rank
Also, I just conducted the on-the-page-optimization and now I am going to start buidling links.
=> I would like to estimate how many linking root domains I need to overrank one of my competitor. These are the MOZ data:
1. My page:
Page Linking Root domains: 0
Root Domain Linking Root Domains: 1512. Competitor:
Page linking root domains: 1
Root Domain Linking Root Domains: 5,786I don't really know on which metric (Page or domain LRD) to rely on in order to make an estimation and I would be glad for some help!
To simplyfy the problem, assume that all toher factors (code, on-the-page keyword use., social etc.) are equal for both sites. Can I just get 2LRD to that page in order to likely outrank my competitor or do I need around 5000 more links poiting to my site?
I think an answer to this question could help a lot of users here, since I saw similar questions/difficulties regarding the use of page LRD vs. root domain LRD
P.S. Non of the pages of my website do currently rank in the top 100 for that keyword.
-
**I am trying to get an rough estimation anyway.... **
That's good. Lots of people are selling the $1000/month package to people who need $5000/month to be effective and don't know it. You are trying to figure it out and that is very important.
Economists call this "ceteris paribus" analysis.
There's an awful lot of variables here. That's what makes it interesting.
-
Thanks EGOL for the detailed answer and sorry for my late reply- just coming back from vacation
It is clear to me that the question cannot just be easily answered by a number. I understand this and I wish my clients/supervisors would do that too - but that's a different story...
I am trying to get an rough estimation anyway, assuming that all other factors (on-the-page factors, words on page, quality of the content, social signals, age of the domian etc.) of the competitor's page/site are exactly the same as on my page/site and these factors would not move. Economists call this "ceteris paribus" analysis.
P.S. I've been in the business now for a while, but just starting to use SEOMoz (we used other tools before) - the ressources on this website and especially this question section is just simply awesome!!
-
Thank you Robert. I put a little extra effort into it because I figured that you would like it.
-
Ayup - and to accomplish that means being willing/able to do some business analysis, not just site analysis. Which moves into the realm of web marketing optimization, not just SEO. Which is where the real value in this whole process lies IMO
-
EGOL
This may be the best answer I have seen to any question. The beauty of it goes on and on in that anyone in SEO dealing with clients must always ask themselves: Can the client handle the truth and/or understand the explanation?
The explanation is what you just gave and the truth is that no one knows for sure and it will often cost you a lot to get there. The other difficulty for SEO clients is that we live in a world of instant gratification. I have clients who spend a lot with me and regularly ask what they are getting (yes, they get reports, etc.). Interestingly, these same clients will mention they have just authorized a $250,000 TV ad campaign. (Trust me we are a bit cheaper.)
They understand the tv ad no matter what the results; even with excellent results (more and more customers) they have a hard time understanding the SEO.
Egol,
Thanks a bunch for a truly excellent, thought provoking answer.
Robert
-
We compete against pages on about.com and amazon.com that are thin content and have zero off-page assets. About.com has skimpy information about our topic and nothing to sell. Amazon has a couple things to sell an zero informative information.
They outrank us for the short tail.
We have the best and deepest informative content on the web in our niche, informative video, youtube channel, lots of articles, printable resources, one of the best selections of products on the web, an email address that answers questions almost 24/7/365 and a phone where you can talk to someone who uses and tests most of the products regularly and has handled, inspected, photographed, described, every product that we sell.
This is one of the problems with google giving huge authority to the brands. Mom and pop know an awful lot more and give immediate, helpful, informative, caring service... but they get no respect in the SERPs.
I understand our position in the rankings. Praise be that I don't have to explain it to a client!
-
Or to determine whether the costs of attempting success are just too high to be worth it.
I would really respect an SEO who told me something like this.
-
Egol's explained the true scope of what you're asking, but to give a specific response to your specific question:
No, just having one more incoming link to your page than your competitor has to his will never guarantee you'll outrank him.
Search engine algorithms use hundreds of signals when ranking a page, and the number of incoming links to that page is only one measure.(And it's more than just wht else is on the page, like code, keyword use etc)
Even if we just restrict ourselves to considering the links to a single page, additional issues like the authority of the sources of the incoming links plays a huge part. Not all links are of equal value.
But more importantly, things like the authority of the whole domain have huge impact as well. That's why SEOMoz goes out of it's way to compute scores for domain authority as well as page authority. A weak page on a strong site will frequently outrank a strong page on a weaker site. That's one of the big frustrations/challenges of nearly all small site owners.
Bottom line, neither Page nor Domain LRD metrics are sufficient in themselves for assessing the work ahead of you.
Make sense?
Paul
-
What Egol has just described is why SEO is only part science and the rest - a significant portion - is still an art. It's also why automated implementation tools have never been successful.
It takes juggling a wide range of constantly changing factors, many with very subjective values, to deliver success. Or to determine whether the costs of attempting success are just too high to be worth it.
Website owners always HATE to hear the accurate response, but the honest response to so many (most?) SEO questions is:
It Depends.
And the good, ethical SEOs will tell you that right up front, as Egol has done. It would be so much easier if SEO, Conversion Rate Optimization etc were linear pursuits - do X and the result will be Y. But they're not, which is why good SEOs will beat those who simply follow (and promise results from ) formulas.
Frustrating huh?
Paul
-
Anyone who gives you a number for this question is full of beans.
This is one of the most difficult questions in SEO and most of the people who are charging clients for SEO can not answer it.
Lots of SEOs have clients on a $500/month plan when the target that they are attacking needs many times that much to become competitive.
Let me give just a couple reasons why this question is so difficult and why nobody here can give you an answer with the information that you provided.
====================
Your question is making a straight comparison.... it assumes that you are racing a stationary target. The target is not stationary. This question is really like this...
Two cars are driving on a road.... Car B leaves when Car A is already 100 miles down the road and traveling at a rate of 50 miles per hour. How fast will Car B have to drive to overtake Car A?
The answer is a velocity AND a time. And the question as I stated it assumes no acceleration.
If Car B drives 51 miles per hour... will you have the patience and budget for that long of a wait? Do you have the resources needed to drive 70 and not be stopped by Google?
=====================
Another example...
George Pickett was a southern general at the Battle of Gettysburg who was ordered by Robert E. Lee to run the Union Army off of Cemetery Ridge. To do that Pickett's troops would need to cross a mile-wide open field in broad daylight under an absolute hail of enemy fire.
The field was one mile wide, it was up a slight incline - if they were slow in crossing they would be mowed down...
The Union had hundreds of troops dug in and positioned on the ridge and ready to open fire - if Pickett had only a few troops they would be mowed down....
So Pickett needed to order an enormous number of troops across that field and order them to run their asses off to engage the Union Army quickly - or die in the middle of the field.
If they failed in getting a enough troops... if they failed in getting enough speed... if they failed in getting enough determination and courage... then they would be mowed down.
Your job is similar to Pickett's... you must get an enormous number of links... you must get them quickly... and you must get them before your budget runs out... and you better hope that your budget is big enough.
Pickett knew that his field was a mile wide.. he knew how many troops that he had. He could see Union troops on the other side of the field at the top of the ridge.
Pickett's generals complained when they were told the battle plan... but Pickett pointed at the Ridge.
This battle was lost on bad math.
But your problem is even more difficult... the field is getting wider as you cross it because the target is moving away from you.
==================
And your problem is also more difficult because links have different values. One link from the Pope's site is worth a thousand from pedestrian sites.
Furthermore... different pages compete with different amounts of vigor. I might have a fifty word page but my competitor might have 2000 words, ten images, a video, and tables of data. Big difference.
===================
So, how is an SEO to know what is needed?
-
You run the keyword difficulty tool and get a feel for the numbers (which contain much confusion).
-
You go out to those SERPs and visit the sites and get a feel for their authority and content quality.
-
Then you use open site explorer to see if the Pope is on their side.
-
Then you decide if you are up to beating their content, beating their numbers, beating their quality.... and if you can do that quickly enough before the conditions of engagement change or your budget runs out.
After all of that you decide on taking the gamble or not.
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What are Keywordbasket.com & Keyword-suggest-tool.com. Competitors have Tons of these
My client is in the Custom Scale Model industry. They build Custom Aircraft Models, Custom Ship Models and more. Their competitor's Backlinks reports have a Ton of metrics coming from: http://www.keywordbasket.com and https://www.keyword-suggest-tool.com There is line after line of these entries in the backlink report. As much as 50 line entries. what exactly is happening here? Are they manually using these tools to generate search entries and caching this data in the form of Backlinks? Please explain. is this a Gray Hat Tactic?
Moz Pro | | chachakhota0 -
New to Moz, need some probably basic answers about Keywords, Linking, Competitors and General SEO
Hi, So I have quite a lot of data colelcted about my site now, regarding keyword research, page crawling and competitor research ect. But I find myself second guessing myself about what I have done and what to do next. I have done basic research for as many relevant keywords I could think of to my site, including branded and non branded terms. If the main competitive keywords for my niche are very competitive, shall I start doing more research for long tail keywords and only try to rank for them? Does is matter how many keywords I am doing research for? Does is matter how many keywords I try to optimise for each webpage? Are the amount of branded keywords I am researching skewing my results? As they are all ranked #1, but nearly all of the non branded keywords are much further down the list... Once I have decided what keywords are worth trying to ranking for for each page, are the techniques to actually rank more highly for them - Title, H1 Tag, Description, Meta Data, Fresh Content and using the keywords on the page? Or are there more techniques I haven't heard of? Under Keyword Rankings - I noticded that some of my keywords are directing to specific pages, like "Cavity Waxes" is directing to the URL ending in .com/cavity-waxes - How do you assign the keywords im researching to specific URLs? - Or does Moz do it automatically? As most of my keywords seem to be unassigned to any URL, is that because they are not ranking highly enough? How do I best use the data collected through Moz? Good practices? Techniques? Tips and Tricks? What is the best practice for finding potential link partners and asking them for mutual linking? Techniques for finding partners that are likely to link with us, but still provides link juice. I must apologise for this long-winded set of questions, but these are troubling me! Any help would be greatly appreciated, Kind regards, Max Johnson
Moz Pro | | BiltHamber10 -
Can we scan for Competitor keywords with Moz Pro?
Hi All, Can we scan for Competitor keywords with Moz Pro? Wanted to know if we can check all keywords a site is ranking for... Thanks
Moz Pro | | jomin740 -
External Followed Links History, number of links go down
I was reviewing Historical Domain Analysis and found that in last 2 month we lost almost 10000 external followed links. What this could be? is this real or just question seomoz crawling? 30voy1g.jpg
Moz Pro | | ctam0 -
This Rookie needs help! Duplicate content pages dropped significantly.
So I am pretty new to SEO Moz. I have an e-commerce site and recently did a website redesign. However, not without several mistakes and issues. That said, when SEO Moz did a crawl of my site, the results showed A LOT of Duplicate Content Pages on my site due to my having one item in many variations. It was almost over whelming and because the number of pages was so high, I have been trying to research ways to correct it quickly. The latest crawl from yesterday shows a drastic drop in the number of duplicate content pages and a slight increase in pages with too long page titles (which is fixable). I am embarrassed to give the number of duplicate pages that were showing but, just know, it's been reduced to a third of the amount. I am just wondering if I missed something and should I be happy or concerned? Has there been a change that could have caused this? Thanks for helping this rookie out!
Moz Pro | | AvenueSeo0 -
Wild fluctuation in number of pages crawled
I am seeing huge fluctuations in the number of pages discovered the crawl each week. Some weeks the crawl discovers > 10,000 pages and other weeks I am seeing 4-500. So, this week for example I was hoping to see some changes reflected for warnings from last weeks report (which discovered > 10,000 pages). However, the entire crawl this week was 448 pages. The number of pages discovered each week seems to go back and forth between these two extremes. The more accurate count would be nearer the 10,000 mark than the 400 range. Thanks. Mark
Moz Pro | | MarkWill0 -
Seomoz staff help needed
I have this new client new domain new website, www.ard.uk.com when seo moz crawls the site it says it has 24 million links mozrank of over 7, looks amazing but impossible can someone as seomoz have a look please
Moz Pro | | francesco-2850160