How does using a CMS (i.e. Wordpress/Drupal) affect backlinks and SEO?
-
So I need to build a website with over 100 pages in it. Elements of the design will probably be moved around and or tested so I need to use a CMS. It's pretty much a review site so while the content will remain static I'd like to employ A/B testing to mess with conversion rates. Wordpress has a plugin for that even.
So I'm just wondering, since CMS pages are pretty much created on spot and not retrieved from a library, how this affects backlinks and anchor text? How exactly does the external website point to yours if the URL is dynamically generated?
Or am I misunderstanding something? Please recommend any extra resources as well if you can.
-
Sorry, saw the follow-up, but I think the overall thread has you covered. The only real issue with CMS URLs is that you can sometimes have multiple versions pointing to the same page, and this creates duplicate content. There are plug-ins for WordPress that can help with that.
The only exception would be something like an AJAX-style URL, where the page content could change without the URL ever changing (Flash has the same issue, for example). You'll rarely see that in a standard CMS, though, and definitely not in WordPress.
-
Thank you very much CMC-SD, Jared and goodlegaladvice for all your help.
@CMC-SD: As promised, I stole your analogy (Now I realize it was an analogy and not a metaphor, I think) and I tried to explain CMS to my girlfriend who knows nothing about computers. Unfortunately it did not come out as elegantly as you put it and we ended up eating bison burgers instead.
-
Ditto to that Jared. Great explanation. And now I'm hungry.
-
Oh, okay, I definitely misunderstood. You're asking about the back-end rewriting process that makes a pretty URL point to the corresponding ugly URL which in turn points to the page. That's way back-end. Unlike a 301 redirect, it's invisible to the spider. The spider need never know that a URL like http://www.domain.com/?p=123 even exists. While it's crawling, it sees a link to http://www.domain.com/page1.html, follows the link, and sees the HTML for that page. That's all.
-
@CMS-SD: Great metaphor! I'm going to steal it But I already knew that about CMS's xD. In fact my confusion was about what follows from that... If the pages are created dynamically and not retrieved from the webserver itself, how do would a backlink even REFER it??
I actually found this SEO blog touching on the subject matter: http://www.seomoz.org/blog/url-rewrites-and-301-redirects-how-does-it-all-work
So, pretty much this is how it works: A page is linked through the URL that is randomly generated by a CMS, but the webserver rewrites the URL that points to the original URL. Pretty much the same thing. And google indexes that URL plus the html on the page. Is that about right? That is why I should not worry at all.
-
Thanks! That's what happens when a creative writing major learns php.
-
This is probably the most well constructed, and humorous explanation on this that I have ever read. Bravo.
-
No. What "indexing" means is creating a database of URLs and the HTML that those URLs point to. If your site has been "indexed," it means Google has discovered your URLs and taken note of the HTML that can be found at those URLs.
-
I think you are misunderstanding something, yes.
On a website with a CMS, the URL is not "dynamically generated." The page is dynamically generated. Here's what that means. Whenever you type http://www.domain.com/page1.html into your browser, you are telling your browser to go to that website and pull up the HTML that corresponds to that URL. URL stands for "uniform resource locator," meaning directions to the location of a resource. If you have an old-fashioned website, the URL points to an HTML file that you created, either by typing everything yourself of using a WYSIWYG editor. If you have a CMS, the URL essentially instructs your website to build the corresponding HTML page on the fly.
It's like ... okay, imagine that you walk into a bakery and ask for a chocolate chip cookie. They could either pull a pre-baked chocolate chip cookie off the shelf and hand it to you, or walk in the back and bake you one cookie from the ingredients in the kitchen. When we're talking about baked goods, option 1 is almost always better than option 2 because it's orders of magnitude faster and more efficient. The benefits that option 2 offers aren't worth the extra time and lost efficiency. But when we're talking about websites, that's no longer the case. The server can construct an HTML document almost instantaneously. Your browser gets the HTML just as fast as it would if it asked for a static HTML page.
In fact, your browser really has no idea that this is all happening. Here's another food metaphor. You walk into a fast food joint and order a hamburger. The cashier walks into the kitchen, and a minute later, walks out with your hamburger. Did the cashier pull the hamburger off a shelf of hamburgers that have been sitting under a hotlight for hours? Or did the cashier ask the cook to prepare a fresh hamburger just for you? Assuming the hamburger tastes great either way, you have no way of knowing. In this metaphor, the customer is the surfer, the cashier is the browser, and the kitchen is the server your website is hosted on. Either your server has a bunch of pre-made pages sitting around waiting for someone to "order" them, or your server has a clever program that makes the pages only when they're needed. That clever program, the CMS, is like the short-order cook.
The thing to remember is, the search engine spiders are customers, just like the surfer. They don't know what's going on in the kitchen. They don't care. They "typed in" a URL and got some HTML back. They now know that that URL produces that HTML. They remember that. When they see a link to that URL, they know it's pointing to that HTML.
Clear as mud?
-
Ahhh, so Google indexes URLs and not the pages themselves? D'oh.
-
"So I'm just wondering, since CMS pages are pretty much created on spot and not retrieved from a library, how this affects backlinks and anchor text? How exactly does the external website point to yours if the URL is dynamically generated?"
Firstly, different CMS's create pages differently. CMS just means content management, which means the platform just provides a gui for you to add content or make changes. If you are using WP and creating pages, then these pages wil be indexed as any other page, and links pointing to it would simply target the page's URL.
Wordpress uses permalinks and Drupal uses pathauto to redirect platform generated links into SEO friendly one. They use an internal redirect and the resulting URL is indexed in Google. Therefore, you simply treat the resulting URL as the "real" url, and external links to it work fine.
-
right no difference I took a whole site with statics and changed it over to a cms with all rewrites everything works great kept the urls the same though kept the .htm
-
I was under the impression that URL rewrites just change the way the URL is displayed on the browser but not the URL itself. I really need to learn more about the backend stuff.
So it would make no difference if the backlink contained an absolute path?
-
It works the same as a static page except its easy to manage your content....
You usually also use a url re-write agent that can change your urls to say what ever you want. In fact most of the web is now on a cms.
Backlinks, Anchor text is all the same....
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google is a definition engine / Question answering engine
But what do you do to rank when you sell a product and people don't have questions and none of the companies that rank on your keywords answers questions... Let me give you a few example. Title tag for example : People want to know what is it and then have question about it What is the maximum number of characters ? How do you write a good one ? Etc... How to remove stains is another example : Please want to know how to remove all the different types of stains, (ketchup, grease etc..) But what about when you are a online business and want to sell usb keys , rent bicycles https://www.spinlister.com or even sell a software to do A/B testing on your website https://bit.ly/2a6cBuF Can someone explain me how those people mentioned rank without giving definition or answering questions... because according to me they don't do any of that on the pages that I mentioned. I look forward to your reply. Thank you,
Web Design | | seoanalytics0 -
Do we still have this Page Rank / Link juice / Link equity? So this dilution concept?
Hi all, As per the traditional or standard SEO rules, we have this link juice and dilution concept. Many websites have changed their linking structure with this with the beleif "the more number of pages, the PR will get diluted". Then many websites avoided more number of pages from homepage to avoid link juice dilution. Even we followed same. But I just wonder it's still the same way Google handles websites and rankings as per the links. And many websites even avoid more number of 2nd tier/hierarchy pages to avoid link dilution. I have gone through our competitors where they been employing lot of top level pages like 2nd tier/hierarchy pages but still doing good at rankings. Please share your views and suggestions on this. Thanks
Web Design | | vtmoz0 -
Needs clarification: How "Disallow: /" works?
Hi all, I need clarification on this. I have noticed we have given "Disallow: /" in one of our sub-directory beside homepage. So, how it going to work now? Will this "Disallow: /" at sub-directory level going to disallow only that directory or entire website? If it is going to work for entire website; we have already given one more Disallow: / at homepage level blocking few folders. How it is going to handle with two Disallow: / commands? Thanks
Web Design | | vtmoz0 -
Does too much inline CSS impact SEO rankings
Hello, Does implementing a lot of inline CSS have a negative impact on SEO rankings? I imagine it could affect page speed, but any other issues I might run in to?
Web Design | | STP_SEO1 -
What are the downsides and/or challenges to putting page paths (www.example.com/pagepath) on a different server?
Hi, Our company is organized into three different segments and our development team recently needed to switch a portion of the business to subdomain because they wanted to move to a different server platform. We are now seeing the impact of moving this segment of the business to a subdomain on the main domain. SEO is hurting and our MOZ score has dropped significantly. One fix they are debating is moving everything back to one domain, but place segments of the business on different page paths and hosting specific paths on different servers. I.e. the main domain could be www.example.com hosted in one location and then www.example.com/segment1 would be hosted on a different server. They are hoping to accomplish this using some sort of proxy/caching redirection solution. The goal of this change would be to recapture our domain strength. Is this something that is a good option or no? If not, what are the challenges and issues you see arising from doing something like that as I don't know of any other site set up like this. Thanks in advance.
Web Design | | bradgreene0 -
Community Discussion: UX & SEO – Your experience?
We've been looking at the relationship between SEO & UX a bit more closely lately on the blog. Our good pal Cyrus started the wheels turning with a tweet: https://twitter.com/CyrusShepard/status/748296076411625473 ...and that morphed into a Whiteboard Friday idea, which was filmed and posted here: https://moz.com/blog/ux-vs-seo-whiteboard-friday We shared the story of one site that enjoyed rapid growth and that subsequently battled with managing that UX/SEO relationship on Thursday. And it's hard, right? UX and SEO teams often operate independently of one another, and may make decisions that affect one another's work. Sometimes it's a "hindsight is 20/20" situation. Sometimes the answer is so radical and impactful that you may want to settle for a "safe" alternative. I'd imagine many of you have encountered some big issues with user experience and search optimization in your day-to-day over the years. What's the most difficult situation you've encountered with this? How did you resolve it? (I'd bet money on there being some really creative solutions out there :). Is there a particularly challenging situation you're struggling with now that you'd want to share & crowdsource ideas for?
Web Design | | FeliciaCrawford3 -
My Site Is Using A Lot of Hosting Bandwidth. Suggestions?
My website http://www.socialseomanagement.com/ is using tons of bandwidth. I received a message from the hosting company saying I exceeded my monthly bandwidth and it has only been a few days. Can anyone take a look and make suggestions? Thanks
Web Design | | JChronicle0 -
Flash and SEO
When I search the seomoz site for "flash and seo" I only see older articles, 2008. Is there anything newer you can point me to for a current discussion on the impact on SEO of flash sites?
Web Design | | panderson3210