How does using a CMS (i.e. Wordpress/Drupal) affect backlinks and SEO?
-
So I need to build a website with over 100 pages in it. Elements of the design will probably be moved around and or tested so I need to use a CMS. It's pretty much a review site so while the content will remain static I'd like to employ A/B testing to mess with conversion rates. Wordpress has a plugin for that even.
So I'm just wondering, since CMS pages are pretty much created on spot and not retrieved from a library, how this affects backlinks and anchor text? How exactly does the external website point to yours if the URL is dynamically generated?
Or am I misunderstanding something? Please recommend any extra resources as well if you can.
-
Sorry, saw the follow-up, but I think the overall thread has you covered. The only real issue with CMS URLs is that you can sometimes have multiple versions pointing to the same page, and this creates duplicate content. There are plug-ins for WordPress that can help with that.
The only exception would be something like an AJAX-style URL, where the page content could change without the URL ever changing (Flash has the same issue, for example). You'll rarely see that in a standard CMS, though, and definitely not in WordPress.
-
Thank you very much CMC-SD, Jared and goodlegaladvice for all your help.
@CMC-SD: As promised, I stole your analogy (Now I realize it was an analogy and not a metaphor, I think) and I tried to explain CMS to my girlfriend who knows nothing about computers. Unfortunately it did not come out as elegantly as you put it and we ended up eating bison burgers instead.
-
Ditto to that Jared. Great explanation. And now I'm hungry.
-
Oh, okay, I definitely misunderstood. You're asking about the back-end rewriting process that makes a pretty URL point to the corresponding ugly URL which in turn points to the page. That's way back-end. Unlike a 301 redirect, it's invisible to the spider. The spider need never know that a URL like http://www.domain.com/?p=123 even exists. While it's crawling, it sees a link to http://www.domain.com/page1.html, follows the link, and sees the HTML for that page. That's all.
-
@CMS-SD: Great metaphor! I'm going to steal it But I already knew that about CMS's xD. In fact my confusion was about what follows from that... If the pages are created dynamically and not retrieved from the webserver itself, how do would a backlink even REFER it??
I actually found this SEO blog touching on the subject matter: http://www.seomoz.org/blog/url-rewrites-and-301-redirects-how-does-it-all-work
So, pretty much this is how it works: A page is linked through the URL that is randomly generated by a CMS, but the webserver rewrites the URL that points to the original URL. Pretty much the same thing. And google indexes that URL plus the html on the page. Is that about right? That is why I should not worry at all.
-
Thanks! That's what happens when a creative writing major learns php.
-
This is probably the most well constructed, and humorous explanation on this that I have ever read. Bravo.
-
No. What "indexing" means is creating a database of URLs and the HTML that those URLs point to. If your site has been "indexed," it means Google has discovered your URLs and taken note of the HTML that can be found at those URLs.
-
I think you are misunderstanding something, yes.
On a website with a CMS, the URL is not "dynamically generated." The page is dynamically generated. Here's what that means. Whenever you type http://www.domain.com/page1.html into your browser, you are telling your browser to go to that website and pull up the HTML that corresponds to that URL. URL stands for "uniform resource locator," meaning directions to the location of a resource. If you have an old-fashioned website, the URL points to an HTML file that you created, either by typing everything yourself of using a WYSIWYG editor. If you have a CMS, the URL essentially instructs your website to build the corresponding HTML page on the fly.
It's like ... okay, imagine that you walk into a bakery and ask for a chocolate chip cookie. They could either pull a pre-baked chocolate chip cookie off the shelf and hand it to you, or walk in the back and bake you one cookie from the ingredients in the kitchen. When we're talking about baked goods, option 1 is almost always better than option 2 because it's orders of magnitude faster and more efficient. The benefits that option 2 offers aren't worth the extra time and lost efficiency. But when we're talking about websites, that's no longer the case. The server can construct an HTML document almost instantaneously. Your browser gets the HTML just as fast as it would if it asked for a static HTML page.
In fact, your browser really has no idea that this is all happening. Here's another food metaphor. You walk into a fast food joint and order a hamburger. The cashier walks into the kitchen, and a minute later, walks out with your hamburger. Did the cashier pull the hamburger off a shelf of hamburgers that have been sitting under a hotlight for hours? Or did the cashier ask the cook to prepare a fresh hamburger just for you? Assuming the hamburger tastes great either way, you have no way of knowing. In this metaphor, the customer is the surfer, the cashier is the browser, and the kitchen is the server your website is hosted on. Either your server has a bunch of pre-made pages sitting around waiting for someone to "order" them, or your server has a clever program that makes the pages only when they're needed. That clever program, the CMS, is like the short-order cook.
The thing to remember is, the search engine spiders are customers, just like the surfer. They don't know what's going on in the kitchen. They don't care. They "typed in" a URL and got some HTML back. They now know that that URL produces that HTML. They remember that. When they see a link to that URL, they know it's pointing to that HTML.
Clear as mud?
-
Ahhh, so Google indexes URLs and not the pages themselves? D'oh.
-
"So I'm just wondering, since CMS pages are pretty much created on spot and not retrieved from a library, how this affects backlinks and anchor text? How exactly does the external website point to yours if the URL is dynamically generated?"
Firstly, different CMS's create pages differently. CMS just means content management, which means the platform just provides a gui for you to add content or make changes. If you are using WP and creating pages, then these pages wil be indexed as any other page, and links pointing to it would simply target the page's URL.
Wordpress uses permalinks and Drupal uses pathauto to redirect platform generated links into SEO friendly one. They use an internal redirect and the resulting URL is indexed in Google. Therefore, you simply treat the resulting URL as the "real" url, and external links to it work fine.
-
right no difference I took a whole site with statics and changed it over to a cms with all rewrites everything works great kept the urls the same though kept the .htm
-
I was under the impression that URL rewrites just change the way the URL is displayed on the browser but not the URL itself. I really need to learn more about the backend stuff.
So it would make no difference if the backlink contained an absolute path?
-
It works the same as a static page except its easy to manage your content....
You usually also use a url re-write agent that can change your urls to say what ever you want. In fact most of the web is now on a cms.
Backlinks, Anchor text is all the same....
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do elements that are displayed when scrolled impact SEO?
Hi, We are wanting to implement Animate.css and Wowjs on our site and were concerned about the SEO impacts. Basically when the page is loaded, if the element is not within the viewport then the HTML tag (i.e. div tag) have a style="visibility: hidden" and once the element is within the viewport it will change to have style="visibility: visible". Would having the style="visibility: hidden" negatively impact SEO?
Web Design | | KendallHershey0 -
How Progressive Enhancement Will be Helpful for SEO?
We have bundle of webpages where we load the content dynamically with the help of Ajax. Since we, need to implement Ajax crawl scheme for making Google to read those Ajax dynamic content we planned to go with hashbang URL's (!#) by creating HTMl snapshots. But last week Google withdrawn their support on crawling the Ajax crawling scheme we are planning to go with progressive enhancement approach as stated by Google in a press release. So, I just want to know what is meant by progressive enhancement and how we can implement in the case of webpages where we load the content dynamically with the help of Ajax? Please advice me on this.
Web Design | | Prabhu.Sundar1 -
Redirects Not Working / Issue with Duplicate Page Titles
Hi all We are being penalised on Webmaster Tools and Crawl Diagnostics for duplicate page titles and I'm not sure how to fix it.We recently switched from HTTP to HTTPS, but when we first switched over, we accidentally set a permanent redirect from HTTPS to HTTP for a week or so(!).We now have a permanent redirect going the other way, HTTP to HTTPS, and we also have canonical tags in place to redirect to HTTPS.Unfortunately, it seems that because of this short time with the permanent redirect the wrong way round, Google is confused as sees our http and https sites as duplicate content.Is there any way to get Google to recognise this new (correct) permanent redirect and completely forget the old (incorrect) one?Any ideas welcome!
Web Design | | HireSpace0 -
Any alternative techniques to display tabbed content without using Javascript / JSON and be SEO Friendly?
John Mueller's input in the EGWMH hangout suggests that Google MAY ignore expandable content served by Javascript. Are there any alternative techniques to display tabbed content without using Javascript / JSON and be SEO Friendly? I do however view these as good for website interactivity and UX - and see many examples of websites performing well and ranking highly whilst using these techniques - are there any Google friendly ways to serve content on a page so that search bots can recognise and choose to crawl / consume the content as legitimate fodder?
Web Design | | Fergclaw0 -
Please help me articulate why broken pagination is bad for SEO...
Hi fellow Mozzers. I am in need of assistance. Pagination is and has been broken on the Website for which I do SEO in-house...and it's been broken for years. Here is an example: http://www.ccisolutions.com/StoreFront/category/audio-technica This category has 122 products, broken down to display 24 at a time across paginated results. However, you will notice that once you enter pagination, all of the URLs become this: http://www.ccisolutions.com/StoreFront/IAFDispatcher Even if you hit "Previous" or "Next" or your browser back button, the URL stays: http://www.ccisolutions.com/StoreFront/IAFDispatcher I have tried to explain to stakeholders that this is a lost opportunity. That if a user or Google were to find that a particular paginated result contained a unique combination of products that might be more relevant to a searcher's search than the main page in the series, Google couldn't send the searcher to that page because it didn't have a unique URL. In addition, this non-unique URL most likely is bottle-necking the flow of page authority internally because it isn't unique. This is not to mention that 38% of our traffic in Google Analytics is being reported as coming from this page...a problem because this page could be one of several hundred on the site and we have no idea which one a visitor was actually looking at. How do I articulate the magnitude of this problem for SEO? Is there a way I can easily put it in dollars and cents for a business person who really thinks SEOs are a bunch of snake oil salesmen in the first place? Does anyone have any before and after case studies or quantifiable data that they would be willing to share with me (even privately) that can help me articulate better how important it is to address this problem. Even more, what can we hope to get out of fixing it? More traffic, more revenue, higher conversions? Can anyone help me go to the mat with a solid argument as to why pagination should be addressed?
Web Design | | danatanseo0 -
What is the best information architecture for developing local seo pages?
I think I have a good handle on the external local seo factors such as citations but I'd like to determine the best IA layout for starting a new site or adding new content to a local site. I see lots of small sites with duplicate content pages for each town/area which I know is poor practice. I also see sites that have unique content for each of those pages but it seems like bad design practice, from a user perspective, to create so many pages just for the search engines. To the example... My remodeling company needs to have some top level pages on its site to help the customers learn about my product, call these "Kitchen Remodeling" and "Bathroom Remodeling" for our purposes. Should I build these pages to be helpful to the customer without worrying too much about the SEO for now and focus on subfolders for my immediate area which would target keywords like "Kitchen Remodeling Mytown"? Aside from my future site, which is not a priority, I would like to be equipped to advise on best practices for the website development in situations where I am involved at the beginning of the process rather than just making the local SEO fit after the fact. Thanks in advance!
Web Design | | EthanB0 -
Meta author. Is it relevant for website design company in its seo?
We don't usually add the meta author in the websites that we develop. I wonder if it would have any positive effect in our seo. We usually add a link in the footer like this "Diseño Web Vigo "(Website Development in Vigo). I am worry about this links. I'm not sure if they are positive because they are in the footer and so the link appears in all of the pages. Besides all these websites we develop are hosted in two different servers, and google could easily think that it is manipulative thing. What do you think? Thanks!!! 🙂
Web Design | | teconsite.com0 -
Seo for flash home page?
I have a client who insists their home page be a single graphic image of their logo with php menus linking to all of the other pages up top. This appears to me to be an seo nightmare. They seem to be unwilling to have anything to do with changing the appearance and want to rely on seo for "all the other pages" on the site. What's an SEO to do in this situation? Is it possible to have a flash image that lands on a "homepage" for google to crawl, rather than a single image? What's the best seo approach here?
Web Design | | peaceland0