Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Merge 2 websites into one, using a non-existing, new domain.
-
I need to merge https://www.WebsiteA.com and https://www.WebsiteB.com to a fresh new domain (with no content) https://www.WebsiteC.com. I want to do it the best way to keep existing SEO juice.
Website A is the companies home page and built with Wordpress
Website B is the company product page and built with Wordpress
Website C will be the new site containing both website A and B, utilizing Wordpress also.
What is the best way to do this? I have research a lot and keep hitting walls on how to do it.
It's a little trickier because it's two different domains going to a brand new domain.
Thanks
-
Thanks Andy!
I'm still a little confused on how I will be copying data over from Site A to Site C. Is it just a matter of creating an empty Wordpress shell with no template and copying the data from Site A to it? (And how is that usually done?) Then would I redirect all Site A's pages to Site C's pages, just with the URL changed? I.E www.SiteA.com/large-dogs redirect to www.SiteC.com/large-dogs
I don't want all of Site B either, just about 10 pages from it, would I manually copy them over also, and how would that be done?
Thanks
-
I did just this type of thing a little over a year ago and organic traffic is up over 300% now. We made the change mainly to improve the structure of the website(s), with more logical organization and better internal linking. We did do the move all at once (thousands of pages) but it took a lot of behind-the scenes planning to be ready for that.
First came the decisions about what sections and categories made sense for our site. (Using the URL structure to guide users around the site makes it easier for them to find what they are looking for and interlinking between related posts as appropriate is also good—and this helps a lot with search engines.)
Then came the organization of posts into their new categories. To make things easier, we kept the individual path names the same (so www.siteA.com/old-category/old-post-string became www.siteC.com/new category/old-post-string) and uploaded them into their new categories when the time came.
We also used this time to do a limited content review (posts with the most traffic) and we updated a lot of these. We made the choice to keep most of our old posts, even though in our market they can get outdated quickly, to conserve any links we may have acquired. (The main site that we were directing to the new site was pretty old and had picked up a lot of links over time.)
We could have done a more complete content review before the changeover, but in part we wanted to see how these posts did under the new structure—we did get renewed life out of some of them, and we further updated and optimized those.
In conjunction with the export of the old sites to the new one, we made sure to 301 redirect all of the old posts to their counterparts on the new site. For the posts we chose not to bring over, we 301 redirected them to a related post in the same category.
We still occasionally come across things that need to be fixed—old posts that need redirecting/updating or 404 errors that need to be tracked down (one big issue we found was a lot of old pages had old links with hard paths to the old website root domains, causing a bunch of nasty internal not found errors—not good!) but overall we are happy with the change. (Up 308%!)
-
Hi,
One way to do this is to decide which site is going to be the main site (site A) sat on C and copy this data over. If you are bringing in site B, then this can sit at another structure level - you will end up with this...
Site A --> Site C, main pages
Site B --> Site C, product pagesThat then brings in everything from both sites to the new domain.
You then want to redirect both of the old sites to the new one, but don't redirect everything to the root. That isn't a good use of 301 mapping. You need to be mapping on a page level so that you will see...
www.sitea.com/about us -301- www.sitec.com/aboutus
www.siteb.com/newproducts/hammers -301- www.sitec.com/newproducts/hammersThere can be differences in the URL's - you don't need to stick with the same structure as the other sites if it doesn't make sense, but always map pages to something very similar.
Page level is the only way to go if you want to maintain a seamless transition for users as well.
Also, don't expect to hit a switch and do this all at once. You can do this over a period of time because to the user, they will just be redirected to the new pages. You will retain more link juice like this.
This is quite a lengthy process and I am sure I have missed the in-between bits, but this is the basis of what you want to be doing.
Others might chip in with other suggestions for you.
-Andy
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
GoogleBot still crawling HTTP/1.1 years after website moved to HTTP/2
Whole website moved to https://www. HTTP/2 version 3 years ago. When we review log files, it is clear that - for the home page - GoogleBot continues to only access via HTTP/1.1 protocol Robots file is correct (simply allowing all and referring to https://www. sitemap Sitemap is referencing https://www. pages including homepage Hosting provider has confirmed server is correctly configured to support HTTP/2 and provided evidence of accessing via HTTP/2 working 301 redirects set up for non-secure and non-www versions of website all to https://www. version Not using a CDN or proxy GSC reports home page as correctly indexed (with https://www. version canonicalised) but does still have the non-secure version of website as the referring page in the Discovery section. GSC also reports homepage as being crawled every day or so. Totally understand it can take time to update index, but we are at a complete loss to understand why GoogleBot continues to only go through HTTP/1.1 version not 2 Possibly related issue - and of course what is causing concern - is that new pages of site seem to index and perform well in SERP ... except home page. This never makes it to page 1 (other than for brand name) despite rating multiples higher in terms of content, speed etc than other pages which still get indexed in preference to home page. Any thoughts, further tests, ideas, direction or anything will be much appreciated!
Technical SEO | | AKCAC1 -
URL Structure On Site - Currently it's domain/product-name NOT domain/category/product name is this bad?
I have a eCommerce site and the site structure is domain/product-name rather than domain/product-category/product-name Do you think this will have a negative impact SEO Wise? I have seen that some of my individual product pages do get better rankings than my categories.
Technical SEO | | the-gate-films0 -
Correct linking to the /index of a site and subfolders: what's the best practice? link to: domain.com/ or domain.com/index.html ?
Dear all, starting with my .htaccess file: RewriteEngine On
Technical SEO | | inlinear
RewriteCond %{HTTP_HOST} ^www.inlinear.com$ [NC]
RewriteRule ^(.*)$ http://inlinear.com/$1 [R=301,L] RewriteCond %{THE_REQUEST} ^./index.html
RewriteRule ^(.)index.html$ http://inlinear.com/ [R=301,L] 1. I redirect all URL-requests with www. to the non www-version...
2. all requests with "index.html" will be redirected to "domain.com/" My questions are: A) When linking from a page to my frontpage (home) the best practice is?: "http://domain.com/" the best and NOT: "http://domain.com/index.php" B) When linking to the index of a subfolder "http://domain.com/products/index.php" I should link also to: "http://domain.com/products/" and not put also the index.php..., right? C) When I define the canonical ULR, should I also define it just: "http://domain.com/products/" or in this case I should link to the definite file: "http://domain.com/products**/index.php**" Is A) B) the best practice? and C) ? Thanks for all replies! 🙂
Holger0 -
Multilingual Website - Sub-domain VS Sub-directory
Hi Folks - Need your advice on the pros and cons of going with a sub-domain vs a sub-directory approach for a multi lingual website. The best would be a ccTLD but that is not possible now, so I would be more interested in knowing your take on these 2 options. Though, I have gone through http://www.stateofsearch.com/international-multilingual-sites-criteria-to-establish-seo-friendly-structure/ and this somewhat vouches for a sub-directory, but what would you say'?
Technical SEO | | RanjeetP0 -
What is the best method to block a sub-domain, e.g. staging.domain.com/ from getting indexed?
Now that Google considers subdomains as part of the TLD I'm a little leery of testing robots.txt with something like: staging.domain.com
Technical SEO | | fthead9
User-agent: *
Disallow: / in fear it might get the www.domain.com blocked as well. Has anyone had any success using robots.txt to block sub-domains? I know I could add a meta robots tag to the staging.domain.com pages but that would require a lot more work.0 -
Using hyphenated sub-domains or non-hyphenated sub-domains? What is the question! I Any takers?
For our corporate business level domain, we are exploring using a hyphenated sub-domain foir a project. Something like www.go-figure.extreme.com I thought from a user perspective it seems cluttered. The domain length might also be an issue with the new Algorithm big G has launched in recent past. I know with past experience, hyphenated domains usually take longer to index, as they are used by spammers more frequently and can take longer to get out of the supplementary index. Our company site has over 90 million viewers / year, so our brand is well established and traffic isn't an issue. This is for a corporate level project and I didn't have the answer! Will this work? anyone have any experience testing this. Any thoughts will help! Thanks, Rob
Technical SEO | | RobMay0 -
Multiple Domains, Same IP address, redirecting to preferred domain (301) -site is still indexed under wrong domains
Due to acquisitions over time and the merging of many microsites into one major site, we currently have 20+ TLD's pointing to the same IP address as our "preferred domain:" for our consolidated website http://goo.gl/gH33w. They are all set up as 301 redirects on apache - including both the www and non www versions. When we launched this consolidated website, (April 2010) we accidentally left the settings of our site open to accept any of our domains on the same IP. This was later fixed but unfortunately Google indexed our site under multiple of these URL's (ignoring the redirects) using the same content from our main website but swapping out the domain. We added some additional redirects on apache to redirect these individual pages pages indexed under the wrong domain to the same page under our main domain http://goo.gl/gH33w. This seemed to help resolve the issue and moved hundreds of pages off the index. However, in December of 2010 we made significant changes in our external dns for our ip addresses and now since December, we see pages indexed under these redirecting domains on the rise again. If you do a search query of : site:laboratoryid.com you will see a few hundred examples of pages indexed under the wrong domain. When you click on the link, it does redirect to the same page but under the preferred domain. So the redirect is working and has been confirmed as 301. But for some reason Google continues to crawl our site and index under this incorrect domains. Why is this? Is there a setting we are missing? These domain level and page level redirects should be decreasing the pages being indexed under the wrong domain but it appears it is doing the reverse. All of these old domains currently point to our production IP address where are preferred domain is also pointing. Could this be the issue? None of the pages indexed today are from the old version of these sites. They only seem to be the new content from the new site but not under the preferred domain. Any insight would be much appreciated because we have tried many things without success to get this resolved.
Technical SEO | | sboelter0 -
Starting a new product, should we use new domain or subdomain
I'm working with a company that has a high page rank on it's main domain and is looking to launch a new business / product offering. They are evaluating either creating a subdomain or launching a brand new domain. In either case, their current site will link contextually to the new site. Is there one method that would be better for SEO than the other? The new business / product is related to the main offering, but may appeal to different / new customers. The new business / product does need it's own homepage and will have a different conversion funnel than the existing business.
Technical SEO | | gallantc0