Multiple doamin with same content?
-
I have multiple websites with same content such as
http://www.example.org and so on. My primary url is http://www.infoniagara.com and I also placed a 301 on .org.
Is that enough to keep away my exampl.org site from indexing on google and other search engines?
the eaxmple.org also has lots of link to my old html pages (now removed). Should i change that links too? or will 301 redirection solve all such issues (page not found/crawl error) of my old webpages?
i would welcome good seo practices regarding maintaining multiple domains
thanks and regards
-
You want your redirect rules on the server, not client site. In Apache you can do this with mod_write and the .htaccess file like so.
To add the www
RewriteEngine on
RewriteCond %{HTTP_HOST} !^www.
RewriteRule ^(.*)$ http://www.%{HTTP_HOST}/$1 [R=301,L]To remove the www:
RewriteEngine On
RewriteCond %{HTTP_HOST} ^www.(.+)$ [NC]
RewriteRule ^(.*)$ http://%1/$1 [R=301,L]In IIS they have a rewrite command too. I've not used it myself but this should help: http://www.petermoss.com/post/How-to-redirect-non-www-domain-to-www-domain-requests-in-IIS-7.aspx
-
Anyway I am lucky being with a group of average team mates.
-
I would recommend either asking a new Q&A as to how an IIS redirect works, or checking Google. I lack experience working in that environment. I was spoiled by working with a very talented team who had performed all those changes and I never needed to learn any aspects of IIS.
-
Hi,
Thanks. As you said i was checking server side redirection. My site server is II7. I tried to make a server side rediretion but couldn't. I found a java script redirection and created a redirection. See the page; http://www.infoniagara.com/d-bed-roses.html
I think this too is not a correct redirection, is it?
thanks
-
Glad to be of help. You are always free to reach out here at the SEOmoz Q&A. If you feel a need to reach me specifically my contact information is in my user profile.
-
Ohh thanks so much Ryan. Let me learn the server side redirection and other aspects related to it.
Thanks once again for your consideration and time. I hope I can approach you in future too.
best regards
-
The javascript code you shared is not a proper redirect.
A proper redirect happens on the server. Instead of loading the original target page, the server instead will load the redirect page instantly along with a header response code of 301 which tells search engines the content has moved to a new URL.
If you use javascript in the manner you shared, the original page will load with a 200 "all ok" header code, and then 3 seconds later the javascript will trigger and load the new page with a 200 header code. All the backlinks will still be applied to the original page and not the redirected page.
The exact method of performing a redirect varies based on your server setup. If you have a LAMP server with cPanel, there is a Redirect tool which you can use.
-
hi,
It's really helpful and now I understand it.
I know that you are one of the true masters of SEO and I think you can clarify one more doubt. It is also regarding the issue of redirection. I want to know that a script i use for redirectining old pages to newpages is right or not?
This script is working properly and I want to know that what type of redirection is this and is it a proper redirection to get the backlink juice? (I add this script to the body just after the
header.)
thanks in advance
-
A 301 redirect is the proper solution and superior to the canonical. It is fine to have the canonical too, but add the redirect.
When you ask "should I do it for each page", understand a single redirect can forward all non-www traffic on your site to it's www equivalent. If you are unsure how to perform the redirect, simply ask your host. Most sites are on managed hosting and it is a very common and easy request.
-
hi dear Ryan,
thanks so much for your time and valuable suggestions. As you said, i am aware of this problem and therefore i added a canonical url to the homepage. should i make a 301 from non-www to www url? should i do it for each page?
thanks & regards
-
In short, you should not use duplicate content across various domains. Doing such will likely negatively affect rankings for either site. Try doing a search that would naturally return the results for a duplicated web page. You will likely find one page ranks well, while the other page ranks significantly lower due to the duplication.
I checked your .org site and it's pages are properly 301 redirected to the .com site. This change would cause any valid pages listed on the .org site to disappear from Google's index. It may take a month from the date the 301 was implemented for Google to crawl and update the entire site.
One point I would add is I suggest you perform a Google site:http://www.infoniagara.org search. Notice you do have a lot of search results for the .org site still present. Those pages are properly redirected to the .org site but they receive 404 errors. If the pages are really gone and there is no equivalent, that is fine and these results should disappear from Google's index over time. If there are similar pages on your site, you should 301 redirect these pages to them.
Another issue, your .com site appears in both the www and non-www form. If you take a URL, remove the "www" then the page appears normally showing the non-www URL. This is a problem which needs to be fixed as it is dividing your backlink juice. Pick one version of your URL, the www or non-www version, and 301 the other version.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Is this considered duplicate content?
Hi Guys, We have a blog for our e-commerce store. We have a full-time in-house writer producing content. As part of our process, we do content briefs, and as part of the brief we analyze competing pieces of content existing on the web. Most of the time, the sources are large publications (i.e HGTV, elledecor, apartmenttherapy, Housebeautiful, NY Times, etc.). The analysis is basically a summary/breakdown of the article, and is sometimes 2-3 paragraphs long for longer pieces of content. The competing content analysis is used to create an outline of our article, and incorporates most important details/facts from competing pieces, but not all. Most of our articles run 1500-3000 words. Here are the questions: NOTE: the summaries are written by us, and not copied/pasted from other websites. Would it be considered duplicate content, or bad SEO practice, if we list sources/links we used at the bottom of our blog post, with the summary from our content brief? Could this be beneficial as far as SEO? If we do this, should be nofollow the links, or use regular dofollow links? For example: For your convenience, here are some articles we found helpful, along with brief summaries: <summary>I want to use as much of the content that we have spent time on. TIA</summary>
White Hat / Black Hat SEO | | kekepeche1 -
Help finding website content scraping
Hi, I need a tool to help me review sites that are plagiarising / directly copying content from my site. But tools that I'm aware, such as Copyscape, appear to work with individual URLs and not a root domain. That's great if you have a particular post or page you want to check. But in this case, some sites are scraping 1000s of product pages. So I need to submit the root domain rather than an individual URL. In some cases, other sites are being listed in SERPs above or even instead of our site for product search terms. But so far I have stumbled across this, rather than proactively researched offending sites. So I want to insert my root domain & then for the tool to review all my internal site pages before providing information on other domains where an individual page has a certain amount of duplicated copy. Working in the same way as Moz crawls the site for internal duplicate pages - I need a list of duplicate content by domain & URL, externally that I can then contact the offending sites to request they remove the content and send to Google as evidence, if they don't. Any help would be gratefully appreciated. Terry
White Hat / Black Hat SEO | | MFCommunications0 -
Are online tools considered thin content?
My website has a number of simple converters. For example, this one converts spaces to commas
White Hat / Black Hat SEO | | ConvertTown
https://convert.town/replace-spaces-with-commas Now, obviously there are loads of different variations I could create of this:
Replace spaces with semicolons
Replace semicolons with tabs
Replace fullstops with commas Similarly with files:
JSON to XML
XML to PDF
JPG to PNG
JPG to TIF
JPG to PDF
(and thousands more) If somoene types one of those into Google, they will be happy because they can immediately use the tool they were hunting for. It is obvious what these pages do so I do not want to clutter the page up with unnecessary content. However, would these be considered doorway pages or thin content or would it be acceptable (from an SEO perspective) to generate 1000s of pages based on all the permutations?1 -
Indexing content behind a login
Hi, I manage a website within the pharmaceutical industry where only healthcare professionals are allowed to access the content. For this reason most of the content is behind a login. My challenge is that we have a massive amount of interesting and unique content available on the site and I want the healthcare professionals to find this via Google! At the moment if a user tries to access this content they are prompted to register / login. My question is that if I look for the Google Bot user agent and allow this to access and index the content will this be classed as cloaking? I'm assuming that it will. If so, how can I get around this? We have a number of open landing pages but we're limited to what indexable content we can have on these pages! I look forward to all of your suggestions as I'm struggling for ideas now! Thanks Steve
White Hat / Black Hat SEO | | stever9990 -
Publishing the same article content on Yahoo? Worth It? Penalties? Urgent
Hey All, I am currently working for a company and they are publishing exactly the same content on their website and yahoo. In addition to this when I put the same article's title it gets outranked by Yahoo. Isn't against Google guidelines? I think Yahoo also gets more than us since they are on the first position. How do you think should the company stop this practice? Please need urgent responses for these questions. Also look at the attachment and look at the snippets. We have a snippet (description) like the first paragraph but yahoo somehow scans the content and creates meta descriptions based on the search queries. How do they do That?
White Hat / Black Hat SEO | | moneywise_test0 -
How do you optimize a page with Syndicated Content?
Content is syndicated legally (licensed). My questions are: What is the best way to approach this situation? Is there any a change to compete with the original site/page for the same keywords? Is it okay to do so? Will there be any negative SEO impact on my site?
White Hat / Black Hat SEO | | StickyRiceSEO0 -
IP-Based Content on Homepage?
We're looking to redesign one of our niche business directory websites and we'd like to place local content on the homepage catered to the user based on IP. For instance, someone from Los Angeles would see local business recommendations in their area. Pretty much a majority of the page would be this kind of content. Is this considered cloaking or in any way a bad idea for SEO? Here are some examples of what we're thinking: http://www.yellowbook.com http://www.yellowpages.com/ I've seen some sites redirect to a local version of the page, but I'm a little worried Google will index us with localized content and the homepage would not rank for any worthwhile keywords. What's the best way to handle this? Thanks.
White Hat / Black Hat SEO | | newriver0