Is SEOMoz only good for "ideas"?
-
Perhaps I've learned too much about the technical aspects of SEO, but nowhere have I found scientific studies backing up any claims made here, or a useful answer to a discussion I recently started.
Maybe it doesn't exist. I do enjoy Whiteboard Friday's. They're fantastic for new ideas. This site is great. But I take it there are no proper studies conducted that examine SEO, rather just the usual spin of "belief from authority".
No?
-
Exactly. And so, would it not be greatly beneficial knowledge to all of us to know if and when a limit is reached where this strategy is no longer effective?
For example, there are many PR8 sites with literally hundreds of PR6 pages that allow dofollow commenting. We can alter anchor text and the deeplink to gain links from these PR6 pages. The question is when does this strategy become ineffective? Let's say our site has 100k pages. Should we spend our time getting a link from every available PR6 page from the same domain? Or is there a diminishing value?
Having some sort of a study that's tried and proven to show if a persistent benefit exists, and when it wears off, would be invaluable to practical SEO, and the results of a study such as this are highly unlikely to change within a year.
Surely you'd like to see something like this too?
I do understand the need to keep SEO in-line with Matt Cutt's objectives, however the reality is that Matt Cutts objectives and what works are two different things. There would be no such thing as off-site SEO at all if Google worked the way it meant to. The thing is, is that it doesn't, and that is why off-site SEO exists.
Instead of people giving hogwash answers, we should be demanding these sorts of useful studies. That is just my opinion anyway.
-
OK... for your question... maybe a little bit more if all links go to the same URL. However, if the links go to different URLs you might get a lot more.
-
And when Schema.org is fully implemented, and in turn integrated into ranking factors, it's going to go through the roof as far as factors go.
-
I wasn't asking to pin point anything to decimal places. I was only after peoples view on whether its "a little bit more" or "probably nothing" etc. That way I can see who here actually knows anything about SEO.
-
Oh - and let's not forget that unless we have the exact same data set as any search engine, we move even further away from the mark of accepted scientific best practices methodology.
This is really important. In the past SEOs could count all of the onpage factors and count their links and count their anchor text. Now as Google starts using things like social data, analytics data from Google.com and other information that SEOs can not count or even see - that is when the ability to reverse engineer begins to disappear - and google becomes less likely to be manipulated.
-
I like how EGOL summed it up in regard to the fact that search engines won't reveal their methods.
They claim hundreds of factors, yet when those are cross-related, that leads to exponential sub-factors (and thus why Google and Bing like to tout 10,000 factors behind the hundreds).
We live in a correlation industry. Any true scientific analysis to reverse engineer the actual factors is by nature going to very likely miss something, and it could be quite significant in how much the results are actually false flag.
Where it gets more complex is that no two situations (in any truly competitive landscape) are exactly alike, and thus the need to replicate for verification is an even more elusive task.
Then add in that hundreds of changes occur to search algorithms throughout the year (some small, some big), and now we're talking about a barrier to true scientific evaluation.
Oh - and let's not forget that unless we have the exact same data set as any search engine, we move even further away from the mark of accepted scientific best practices methodology.
On a final note, the amount of time, computing power and analysis required in most situations, is more likely cost-prohibitive since the results of such effort can not be recouped. So that leaves it to an entity that has the financial, technical, and academic willingness to take on such a task without expectation of compensation.
Rand gets hammered all the time for referring to Moz's process with the tag line "correlation is not causation", even though we work in a correlation industry.
-
Hi Steven,
While I don't feel it's necessary to defend SEOmoz, or SEO for that matter, it is impossible to know exactly what do to implement perfect SEO techniques and rank in the top spot for each industry.
However, SEOmoz is one of the few places I have watched webinars that collect vertical data, and process the statistics to made heads and tails of it. I do remember reading a Google Best Practices Guide from 2010 that did something similar, but not as well as SEOmoz.
The more I read and learn from this site, the more I realize that SEO is not abut tricking Google and other search engines. It is about relying on the best practices to create high quality, relevant websites that Google will appreciate. I continue to believe that if SEOers stay consistent in their practices, Google will end up changing their algorithms to suit SEOmoz's standards. Search engines will be chasing our sites, and our clients sites based on the excellent product we are helping produce.
SEO can be an extremely frustrating if you are looking for black and white answers. I find the best practices approach to be much less frustrating, and easier to manage in regards to expectations and results.
Edit
I also wanted to mention that I wish people would have more ideas to consider in reference to best practices and techniques. This is how SEO will evolve, and rarely do people have a chance to be in on the ground floor with cool stuff like this.
-
Can you tell me the diminishing value of a sitewide link vs a single homepage link for example? And how is it you know the answer to this?
Nobody is going to tell you that a sitewide link is worth 2.76788756 times the value of a homepage link.
The answer isn't that simple. Sites come in different sizes, they have different linkages, they have different navigation structures and they have different numbers of links hitting homepages and internal pages. Maybe it even matters if these links are in the footer, the top navigation or some other location.
There is a word that is used when people try to reverse engineer something that is really complex..... that is fuzzy logic. I think that this term has huge application in SEO.
But my personal opinion is when we deal with questions of fuzzy logic, we make the most effective use of our time when we accept answers such as "a little bit more"..... "probably nothing".... and "maybe a little less". Then we simply apply them, and move on rather than trying to get into proving theorems and attempting to calculate out to five decimal points.
-
Hi Ryan,
You're quite right. The site does offer some useful tools and interviews. In fact for that reason alone I will be retaining my membership. I wasn't really after a secret handshake, but more sharing of analysis data.
-
Steven, it sounds like you joined to learn the secret handshake. If that is the case then, from your point of view, you will be disappointed.
EGOL and Marcus both shared excellent perspectives on the SEOmoz site's offerings. You have also looked around the site. Based on your replies I think you have examined the site and have not missed any major components.
What SEOmoz does offer members is:
-
Tools: Site Explorer, MOZbar and other tools to examine websites. Yes, there are similar tools out there and also some of these tools can be improved. I would like to see all of SEOmoz's tools improved to be the best in the industry. They aren't there yet.
-
Original interviews: you mentioned Whiteboard Fridays. The recent interview with Duane Forrester from Bing offered fantastic insight into upcoming changes at Bing. Being part of a network which talks to industry leaders and asks the right questions is very rewarding.
-
Active community: if you do get stuck or otherwise have a question, SEOmoz offers a place you can go to for help. From what I have seen, other communities are not very active nor do they offer the quality of feedback these forums provide.
The idea seems to be centered around providing a place for those interested in SEO the tools, information and discussion area to do their job better. If that is not what you are looking for, then I would suggest you simply enjoy your free month then try something else.
-
-
Marcus,
I appreciate the good natured will of your post, and I thank you for that. However as I stated to Egol, I am not by any means new to SEO. Please don't let my small post count here deceive you.
The links you provided regarding "asking the experts" is what is known in science as an "appeal to authority", which has no relevance whatsoever to evidence. Just because some prominent scientists, or SEO's, give their opinion about something, even if it is in the vast majority, does not constitute evidence. Those, myself included, who know more than many of the names mentioned within those lists, care not for opinions, but replicatable tests. This is how any knowledge is truly understood.
I do very much respect the ideas presented here, and the community, and in fact have learned some non-technical SEO related things. I especially applaud the Whiteboard Friday section. However I'm not really after trends, or patterns, but rather a discussion regarding specific questions, such as the one brought up in a previous discussion I created.
Can you tell me the diminishing value of a sitewide link vs a single homepage link for example? And how is it you know the answer to this? That is what I want studies demonstrating, so we can analyze this in detail and work out finer points of the algorithm which, despite common theory, doesn't change as much as everyone believes.
-
Egol I agree with you wholeheartedly, though I'm not as much of a newbie as you seem to infer.
The information is indeed kept secret to maintain a competitive advantage, though I thought this site may have revealed some of its own studies. My primary reason for joining here wasn't to learn as much as it was to compare analysis of various SEO algorithms, and I cannot seem to do that here.
One thing I'm sure you're aware given the dynamic, changing nature of SEO is that an understanding of even some temporary aspect, be it for only a few months until it is changed, can bring in huge profits.
Not all fields of SEO are like this however. Some aspects of SEO which haven't changed for years are not even discussed, such as the topic raised in a previous discussion I created. We can study these aspects with near the same level of detail we can study the electron configuration of atoms, it simply requires observation and deduction. This is what I was hoping for.
-
Hey Steven
Your kidding right? The "usual spin"?
It's hard to know exactly what you are looking for here but if there are any studies about SEO then it's SEOMoz and their partners from within the industry that are doing them.
Here are a few that just jump to mind:
1. http://www.seomoz.org/article/search-ranking-factors
The ranking factors article takes so much into consideration. A panel of 130 or so experts and correlation-based analysis from the vast amount of data gathered through the link scape index and other means.
2. http://www.davidmihm.com/local-search-ranking-factors.shtml
This is an article looking at the myriad local search ranking factors. Again it uses a panel and a whole bunch of data to best give an overview of local search ranking factors.
3. http://www.distilled.net/blog/
If you want more general data and mathematical analysis of search then the guys at distilled publish some great information.
4. SEOMoz Tools
You want data? Then the tools here will give you that. Granted, it's not perfect and it won't SEO your site for you but it will give you some metrics to work with.
The blog at SEO book pulls no punches and is often a great place to go for some cutting analysis of what is right and wrong in search.
But, psst, come here, wan't to know the real secret?
Hard work - that's it, plain and simple. SEO is a closed box. The engines don't publish their algorithms and even if they did give us a comprehensive overview of everything covering all areas of search it would likely be out of date before we have finished reading it.
I kind of feel your pain, you want exact answers, specific things that you can do to succeed but SEO just does not work like that. There are way to many variables and for every industry, every site, every country, every city things can be a little different.
If you want to 'examine seo' then you need to do it from your perspective. If you want analysis it has to be done within the context of whatever you are trying to achieve. There are plenty of great SEO's on this board and if you want some help understanding what you can do to rank better then post some more details and I am sure you will get some help, I will certainly take a look.
I hope that helps a little, even if it was not exactly what you may have been looking for.
Marcus
-
Search engines such as Google don't reveal how they rank websites and they modify their methods continuously. One of their most important goals is to avoid manipulation.
As a result nobody knows exactly how google ranks websites. So SEOs must have a mind that is comfortable dealing with a changing uncertainty.
A few people do scientific studies on how rankings work but a lot of that information is kept secret to maintain a competitive advantage - but it has a limited value over time because search engines change their methods.
In my opinion there are four good sources of information....
-
Basic search optimization guides such as the Beginner's Guide to SEO
-
Surveys of web professionals such as the Search Engine Ranking Factors.
-
Forums such as this one where people ask questions and share ideas.
-
Personal records that SEOs keep about what changes they made to their website and the results that occur.
The bottom line is you will rarely know EXACTLY what to do. But you must draw information from what you get from 1,2,3 and 4 to place your best bet.
The primary caution is.... don't be suckered into accepting the answer that you "want to hear" because usually the "most difficult to pull off" is the one that works best.
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Can't get Google to index our site although all seems very good
Hi there, I am having issues getting our new site, https://vintners.co indexed by Google although it seems all technical and content requirements are well in place for it. In the past, I had way poorer websites running with very bad setups and performance indexed faster. What's concerning me, among others, is that the crawler of Google comes from time to time when looking on Google Search Console but does not seem to make progress or to even follow any link and the evolution does not seem to do what google says in GSC help. For instance, our sitemap.xml was submitted, for a few days, it seemed like it had an impact as many pages were then visible in the coverage report, showing them as "detected but not yet indexed" and now, they disappeared from the coverage report, it's like if it was not detected any more. Anybody has any advice to speed up or accelerate the indexing of a new website like ours? It's been launched since now almost two months and I was expected, at least on some core keywords, to quickly get indexed.
Technical SEO | | rolandvintners1 -
Product Subdomain Outranking "Marketing" Domains
Hello, Moz community! I have been puzzling about what to do for a client. Here is the challenge. The client's "product"/welcome page lives at www.client.com this page allows the visitor to select the country/informational site they want OR to login to their subdomain/install of the product. Google is choosing this www.client.com url as the main result for client brand searches. In a perfect world, for searchers in the US, we would get served the client's US version of the information/marketing site which lives at https://client.com/us, and so on for other country level content (also living in a directory for that country) It's a brand new client, we've done geo-targeting within the search console, and I'm kind of scared to rock the boat by de-indexing this www.client.com welcome screen. Any thoughts, ideas, potential solutions are so appreciated. THANKS! Thank you!
Technical SEO | | SimpleSearch0 -
Redundant categorization - "boys" and "girls" category. Any other suggestions than implementing filtering?
One of our clients (a children's clothing company) has split their categories (outwear, tops, shoes) between boys and girls - There's one category page for girls outwear, and one category for boys outwear. I am suspecting that this redundant categorisation is diluting link juice and rankings for the related search queries. Important points: The clothes themselves are rather gender-neutral, girl's sweaters don't differ that much from the boy's sweaters. Our keyword research indicates that norwegians' search queries are also pretty gender neutral - people are generally searching after "children's dresses", "shoes for kids", "snowsuits", etc. So these gender specific categories are not really reflective of people's search behavior. I acknowledge that implementing a filter for "boys" and "girls" would be the best way to solve this redundant categorization, but that would simply be to expensive for our client. I'm thinking that some sort of canonicalisation would be the best approach to solve this issue. Are there any other suggestions or comments to this?
Technical SEO | | Inevo0 -
Many "spin-off" sites - 301 or 401/410?
Hi there, I've just started a new job with a rental car company with locations all over New Zealand and Australia. I've discovered that we have several websites along the lines of "rentalcarsnewzealand", "bigsaverentals" etc that are all essentially clones of our primary site. I'm assuming that these were set up as some sort of "interesting" SEO attempt. I want to get rid of them, as they create customer experience issues and they're not getting a hell of a lot of traffic (or driving bookings) anyway. I was going to just 301 them all to our homepage - is this the right approach? Several of the sites are indexed by Google and they've been linked up to a number of sites - the 301 move wouldn't be to try to derive any linkjuice or anything of that nature, but simply to get people to our main site if they do find themselves clicking a link to one of those sites. Thanks very much for your advice! Nicole
Technical SEO | | AceRentalCars0 -
"Links to your site" in google webmaster tools not showing any data
Hello All I have a very strange query regarding the "Links to your site" section in webmaster's account my account does not show the Link data after so many days (more then 30 days) of verification. Can you please help me out how can I get my data in the webmaster's account?
Technical SEO | | barnesdorf
Please note I have verified the account using Google Analytic verification process. (does this affect?) I have seen this issue in my two websites which I have verified by Google Analytics. Please help me out.0 -
Backlinks go to "example.com" our homepage is "example.com/default.html" am I losing internal link power?
Hey everyone! Thanks again for everybodies contributions to my questions over the last few months. As the title states, our homepage is at "example.com/default.html" but everybody that backlinks to us (as expected) to "example.com" does that mean that I am probably losing a lot of the power of my links??
Technical SEO | | TylerAbernethy0 -
International Websites: rel="alternate" hreflang="x"
Hi people, I keep on reading and reading , but I won't get it... 😉 I mean this page: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=189077&topic=2370587&ctx=topic On the bottom of the page they say: Step 2: Use rel="alternate" hreflang="x" Update the HTML of each URL in the set by adding a set of rel="alternate" hreflang="x" link elements. Include a rel="alternate" hreflang="x" link for every URL in the set, like this: This markup tells Google's algorithm to consider all of these pages as alternate versions of each other. OK! Each URL needs this markup. BUT: Do i need it exactly as written above, or do I have to put in the complete URL of the site, like: The next question is, what happens exactly in the SERPS when I do it like this (an also with Step1 that I haven't copied here)? Google will display the "canonical"-version of the page, but wehen a user from US clicks he will get on http://en-us.example.com/**page.htm **??? I tried to find other sites which use this method, but I haven't found one. Can someone give me an example.website??? Thank you, thank you very much! André
Technical SEO | | waynestock0 -
SeoMoz robot is not able to crawl my website.
Hi, SeoMoz robot crawls only two web pages of my website. I contacts seomoz team and they told me that the problem is because of Javascript use. What is the solution to this? Should I contact my webdesign company and ask them to remove Javascript code?
Technical SEO | | ashish2110