Reverse proxy a successful blog from subdomain to subfolder?
-
I have an ecommerce site that we'll call confusedseo.com. I created a WordPress blog and CNAME'd it to blog.confusedseo.com. Since then, the blog has earned a PageRank of 3 and a decent amount of organic traffic.
I am considering a reverse proxy to forward blog.confusedseo.com to confusedseo.com/blog/. As I understand it, this will greatly help the "link juice" of the root domain. However, I'm concerned about any potential harm done to the existing SEO value of the blog. What, if anything, should I be doing to ensure that the reverse proxy doesn't hurt my "juice" rather than help it?
-
Hey, I have a question in this:
We have setup a seperate Google Analytics ID and Google Search Console Property for the sub-domain and then if we are using reverse proxy to keep it under sub-directory.
So what happens to the GA tracking and Google Search Console in this case?
You can read my full question here:
-
Hi there,
Im investigating the same reverse proxy solution for my eCommerce blog. was your implementation successful?
-
Canonical will pass link juice almost exactly like 301s will, so there's no harm in going that route. Matt Cutts explains that in this video: http://www.youtube.com/watch?v=zW5UL3lzBOA
You sound like you're good to go. You've got duplicate content worked out, and you've got a plan to retain link juice (canonical).
-
Since the subdomain does still exist live, someone doing a reverse proxy would need to take some steps to mitigate duplicate content issues. The first would be to set up the new permalinks and rel canonical tags via Wordpress and Yoast's SEO plugin (which rocks, btw). Then you would need to do the robots.txt/GWT steps that you quoted. If there's anything else that needs doing, I am definitely all ears before I attempt this.
-
Ah! I misunderstood the bit about reverse proxying. In that case... to be perfectly honest, I'm not sure.
When you setup a reverse proxy, what happens to the sub-domain? Does it go away or does it still exist live? If it remains live, you'd end up with a duplicate content issue.
EDIT >> I found this at the source you linked to (which answers my question) -->
"The next thing you can do is add a robots.txt file to the sub-domain that stops robots from indexing it. As Reverse Proxying keeps the requested URL the /blog/ URLs will use the robots.txt from the main domain rather than the sub-domain.
The final (and most extreme) thing you can do is to register Google Webmaster Tools for the sub-domain and remove it from the index. If you are doing this, you need to do it in conjunction with robots.txt."
-
Thanks for your response, Philip. My research indicates that a 301 redirect on a location that is being reverse proxied would result in an infinite loop. (source) I haven't tested it to confirm, though. Is that true?
-
You need to setup 301 redirects for ALL of the pages and posts on the blog sub-domain to their new locations in the sub-folder. This is very important. Without the proper redirects in place, you will lose all value from links pointing to the blog sub-domain, plus all the history, authority, and rankings that the pages have earned.
As for your reasoning to move it from a sub-domain to a sub-folder, I'm not sure you'll receive any sort of link juice boost on your root domain from doing this. Maybe someone else can prove me wrong/correct me...
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Pros and Cons of using rel=next on blog posts
Hi there, at the bottom of my website's blog posts the rel=prev and rel=next tags are used on links that point to the previous article that was posted and the next article that was posted. Often these articles are not 'linked' in terms of their content or message. Is this the correct use of rel=next/prev and if not what are the possible negative effects. Many thanks.
Technical SEO | | Bee1590 -
Subdomain question
Hi guys, I have a subdomain on my site that i want to completely remove from the index. I tried already everything to remove it but it is special situation so the only choice i have left is to remove it from Search Console in "Remove URLs" feature. So my question is: if i remove my root subdomain (example: http://subdomain.mydomain.com/) via "Remove URLs" feature in Webmaster Console, will it remove all the URLs coming from that particular domain as well? I also want to make sure that my root domain will stay untouched and be functioning normally. Thank you for advice!
Technical SEO | | odmsoft0 -
Effects of significant cross linking between subdomains
A client has noticed in recent months that their traffic from organic search has been declining, little by little. They have a large ecommerce site with several different categories of product - each product type has its own subdomain. They have some big megamenus going on, and the end result is that if you look in their Webmaster Tools for one of their subdomains, under Links to your Site, it says they have nearly 22 million links from their own domain! Client is wondering if this is what is causing the decline in traffic and wondering whether to change the whole structure of their site. Interested to hear the thoughts of the community on this one!
Technical SEO | | helga730 -
Wordpress Blog Blocked by Metarobots
Upon receiving my first crawl report from new pro SEOMoz acc (yaay!) I've found that the wordpress blog plugged into my site hasn't been getting crawled due to being blocked by metarobots. I'm not a developer and have very little tech expertise, but a search dug up that the issue stemmed from the wordpress site settings > privacy > Ask search engines not to index this site option being selected. On checking the blog "Allow search engines to index this site" was selected so I'm unsure what else to check. My level of expertise means I'm not confident going into the back end of the site and I don't have a tech guy on site to speak to. Has anyone else had this problem? Is it common and will I need to consult a developer to get this fixed? Many thanks in advance for your help!
Technical SEO | | paj19790 -
Need specifics about mod_proxy for blog domain and 301s
I am getting the IT staff to move our blog from "blog." to "/blog" using mod_proxy for apache, but I had a couple of questions about this I was hoping someone here might be able to help with. Is it correct that just setting up mod_proxy will make the blog available at both URLs? the "blog." subdomain and the "/blog" folder? If so, what is the best way to 301 redirect all traffic from "blog." to "/blog"? I assume this could be handled with a blanket 301 style rewrite, but I wanted to get some other opinions before getting with my IT guys to do it. I am technical enough to talk about this, but not do it myself, so experienced opinions are appreciated. Thanks!
Technical SEO | | SL_SEM0 -
Best Practice to Remove a Blog
Note: Re-posting since I accidentally marked as answered Hi, I have a blog that has thousands of URL, the blog is a part of my site. I would like to obsolete the blog, I think the best choices are 1. 404 Them: Problem is a large number of 404's. I know this is Ok, but makes me hesitant. 2. meta tag no follow no index. This would be great, but the question is they are already indexed. Thoughts? Thanks PS A 301 redirect to the main page would be flagged as a soft 404
Technical SEO | | Bucky0 -
Robots.txt for subdomain
Hi there Mozzers! I have a subdomain with duplicate content and I'd like to remove these pages from the mighty Google index. The problem is: the website is build in Drupal and this subdomain does not have it's own robots.txt. So I want to ask you how to disallow and noindex this subdomain. Is it possible to add this to the root robots.txt: User-agent: *
Technical SEO | | Partouter
Disallow: /subdomain.root.nl/ User-agent: Googlebot
Noindex: /subdomain.root.nl/ Thank you in advance! Partouter0 -
Subdomain Removal in Robots.txt with Conditional Logic??
I would like to see if there is a way to add conditional logic to the robots.txt file so that when we push from DEV to PRODUCTION and the robots.txt file is pushed, we don't have to remember to NOT push the robots.txt file OR edit it when it goes live. My specific situation is this: I have www.website.com, dev.website.com and new.website.com and somehow google has indexed the DEV.website.com and NEW.website.com and I'd like these to be removed from google's index as they are causing duplicate content. Should I: a) add 2 new GWT entries for DEV.website.com and NEW.website.com and VERIFY ownership - if I do this, then when the files are pushed to LIVE won't the files contain the VERIFY META CODE for the DEV version even though it's now LIVE? (hope that makes sense) b) write a robots.txt file that specifies "DISALLOW: DEV.website.com/" is that possible? I have only seen examples of DISALLOW with a "/" in the beginning... Hope this makes sense, can really use the help! I'm on a Windows Server 2008 box running ColdFusion websites.
Technical SEO | | ErnieB0