I've screwed up. Domain pointers I forgot about. Think I am getting dinged by google.
-
Hey all. I setup some domain pointers for a client 8 years ago and now think they are hurting them. I am afraid google thinks it duplicate content. They are pointers so you can get to the same page using other domain names. Is my best approach to do a 301 redirect on them? The client is on a shared host so I have to use the web.config file. The site is pretty small so doing it for the 10+ pages is not that big of a deal. My question is this? When should I drop those pointers from the website altogether?
-
Hi Doug
If you have duplicate content then you could add a cross-domain canonical on all the pages from site 2 to site 1. Then when it's dropped away just 301 everything,
That means you'd still get direct traffic to it but Google would rank the main site 1 and drop site 2 because all the canonicals would reference site 1.
You just put the 301s in the .htaccess file
I wouldn't do it this way - I'd just make sure all the content was on site 1 then 301 but I understand you might be nervous.
Regards
Nigel
-
So I wanted to answer this question if someone else has problems with URL redirecting in IIS and they don't have console level access.
I added the following to my web.config file and this took care of all my problems.
Basically it checks to see if https is being used, if not it redirects to the primary domain using https://www.domain1.com. The second condition checks to see if it is the exact domain name. If it is anything else, it redirects to the primary https://www.domain1.com
**This fixes HTTP to HTTPs, non-WWW to WWW, and other domain pointers to the correct one. **
<rewrite><rule name="Force canonical hostname and SSL" stopprocessing="true"><match url=".*"><conditions logicalgrouping="MatchAny"><add input="{HTTPS}" pattern="off"><add input="{HTTP_HOST}" pattern="^www.domain1.com$" negate="true"></add></add></conditions>
<action type="Redirect" url="https://www.domain1.com{URL}" appendquerystring="true" redirecttype="Permanent"></action></match></rule></rewrite> -
Thanks Nigel.
There are 3 domains setup and all point to the same website. I checked with analytics and one of the domains is not even on the radar, but the other accounts for about 6% of traffic for the last year. I think the problem is google is counting this as duplicate content (same page, but 2 different domains to get there - i.e. domainone.com/page1, domaintwo.com/page1).
Domain one is the main one.
I am just super nervous about this as this is a real business that makes money and I do not want to screw it up.
My plan is to do the following:
1. Update to Https
2. Add rel=canonical to each page using absolutes (https://www.domainone.com/page1).
3. Add 301s using web.config file to point any request for the other domains to the primary.
4. Add 301 to non "WWW" request to point them to WWW version.
5. Update sitemapPray I don't make a mistake in my redirects and make the google gods mad
-
Hi Doug
Firstly I doubt there is any value in the backlinks to those old domains so I guess the only way you will get a click is if someone assumes a domain name and types it in, however
You haven't said if there is any content on those pages. If there is, then delete it and 301 to the relevant pages.
1. Use analytics to see how many entries came from those domains.
2. If it's low to zero and there are no backlinks to them then just delete them altogether.I doubt they are doing much damage to be honest unless they have duplicate content on them.
The cleanest way of course is to have one domain name but big companies still use domains that point at them. It happens a lot with TLD variants e.g. diy.co.uk points to diy.com
Regards
Nigel
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How get google reviews on search results?
Hi, We have good google reviews. (4,8) Can we get this rating stars also on our organic search results ? Best remco
Intermediate & Advanced SEO | | remcoz0 -
When rebranding, what's the best thing to do with the new domain before rebranding?
A. Do nothing
Intermediate & Advanced SEO | | Maxaro.nl
B. Redirect to legacy site (current domain)
C. Create a placeholder with information about the rebranding
D. Other... What do you think is best?0 -
Is it possible to rank in google mexico when you don't have a local site?
Hello, someone is asking me why we don't rank in google mexico search engine. I mentioned we don't have a google mexico site, but have a USA site, so we may rank, but not as well as if we had the mexico site. IS there anyway to improve rankings or tips? THanks! Laura Robinson
Intermediate & Advanced SEO | | lauramrobinson321 -
Content From One Domain Mysteriously Indexing Under a Different Domain's URL
I've pulled out all the stops and so far this seems like a very technical issue with either Googlebot or our servers. I highly encourage and appreciate responses from those with knowledge of technical SEO/website problems. First some background info: Three websites, http://www.americanmuscle.com, m.americanmuscle.com and http://www.extremeterrain.com as well as all of their sub-domains could potentially be involved. AmericanMuscle sells Mustang parts, Extremeterrain is Jeep-only. Sometime recently, Google has been crawling our americanmuscle.com pages and serving them in the SERPs under an extremeterrain sub-domain, services.extremeterrain.com. You can see for yourself below. Total # of services.extremeterrain.com pages in Google's index: http://screencast.com/t/Dvqhk1TqBtoK When you click the cached version of there supposed pages, you see an americanmuscle page (some desktop, some mobile, none of which exist on extremeterrain.com😞 http://screencast.com/t/FkUgz8NGfFe All of these links give you a 404 when clicked... Many of these pages I've checked have cached multiple times while still being a 404 link--googlebot apparently has re-crawled many times so this is not a one-time fluke. The services. sub-domain serves both AM and XT and lives on the same server as our m.americanmuscle website, but answer to different ports. services.extremeterrain is never used to feed AM data, so why Google is associating the two is a mystery to me. the mobile americanmuscle website is set to only respond on a different port than services. and only responds to AM mobile sub-domains, not googlebot or any other user-agent. Any ideas? As one could imagine this is not an ideal scenario for either website.
Intermediate & Advanced SEO | | andrewv0 -
Why is my site not getting crawled by google?
Hi Moz Community, I have an escort directory website that is built out of ajax. We basically followed all the recommendations like implementing the escaped fragment code so Google would be able to see the content. Problem is whenever I submit my sitemap on Google webmastertool it always 700 had been submitted and only 12 static pages had been indexed. I did the site query and only a number of pages where indexed. Does it have anything to do with my site being on HTTPS and not on HTTP? My site is under HTTPS and all my content is ajax based. Thanks
Intermediate & Advanced SEO | | en-gageinc0 -
How to link my websites with each other - to avoid google penalities and get some value
I have good high DA PA websites hosted on same IP, added in same google analytics and GWT account. So i think google knows that owner is the same. How should we link them with each other to get some value? Put nofollow? With what anchor (Money keyword or domain name)? But whats the point? We cannot make natural link building profile with our own website nofollow links, i assume they will not count. What can you suggest? Maybe it is better not to link at all?
Intermediate & Advanced SEO | | bele0 -
I'm pulling my hair out trying to figure out why google stopped crawling.. any help is appreciated
This is going to be kind of long, simply because there is a background to the domain name that is not typical to anybody in the world really and I'm not sure if its possible that it was penalized or ranked lower because of that or not. Because of that I'm going to include it with the hopes that giving the full picture some nice soul in the world who has more knowledge in this than me see's something or knows something and can point me in the right direction. Our site has been around for a few years, at one point the domain was seized by homeland security ICE, and then they had to give it back in Dec. which sparked a lot of the SOPA PIPA stuff and we became the poster child so to speak. The site had previously been up since 2008, but due to that whole mess the site was down for 13 months on the dreaded seized server with a scary warning graphic and site title which caused quite obviously a bunch of 404 errors and who knows what else damage to anything we'd had before that as far as page rank and incoming links. we had a lot of incoming links from high quality sites. We were advised upon getting the domain back to pretty much scrap all the old content that was on the site prior and just start fresh.. which we did. Googlebot started crawling slowly, but then as we started getting back into the swing of things people started linking to us,some with high page rank, we were getting indexed quite frequently and ranking high on search results in our niche.. Then something happened on March 4th, we had arguably our best day with google traffic, we'd been linked back by places like Huff Post etc for content in our niche.. and the next day literally it was a freefall. Darn near nothing. I've attached a screen shot from webmaster tools so you can see how drastic it was. I went crazy, trying to figure out what was wrong, searching obsessively through webmaster tools looking for any indication of a problem, searched the site on google site:dajaz1.com and what comes up is page 2 page 3 page 45 page 46. It's also taken to indexing our category and tag pages and even our search pages. I've now set those all to noindex follow but when I look at where the googlebots are at on the site, they're on the categories, pages, author pages, and tags. Some of our links are still getting indexed, but doing a search just of our site name and we're ranking below many of the media sites that have written about our legal issues, when a month ago we were at least top result for our own name. I've racked my brain trying to figure out the issue. I've disabled plugins, I'm on fetch as google bot all the time making sure our stuff is at least coming out as 200 (we had 2 days where we were getting 403 errors due to a super-cache issue, but once fixed googlebot returned like it never left) I've literally watched 1000 videos, read 100 forums, added in SEO plugins, tried to optimize the site to the point I'm worried I'm over doing it.. and still they've barely begun to crawl. As you can see there is some activity in the last 2-3 days, but even submitting a new site map once I changed the theme out of desperation it's only indexed 16. I've looked for errors all through webmaster tools and I can't find anything to tell me why that happened, how to fix it, and how to get googlebot to like us again. I'm pulling my hair out here. The links we have incoming are high quality links like huffington post , spin, complex, etc. Those haven't slowed down at all, we do outgoing links to sites we trust and are high quality as well. I've got interns working on how they're writing titles and such, I've gone through and attempted to fix duplicate pages and titles.. I've been going through and re-writing meta description tags What am I missing? I'm pulling my hair out trying to figure out what the issue is. Eternally grateful for any help provided. jnzb6.png
Intermediate & Advanced SEO | | malady0 -
Are sites that leave out www from domain at a disadvantage to domains with www in url
I know this has been discussed but was wondering what would be the best approach from an SEO perspective. I quite like the idea of setting up websites with domains without www but always worry that setting up domains without www has a disadvantage because user are use to referring to sites with the www included. Thus one of my fears are that users would link back using www version which will mean even if you do a 301 redirect that some of the link juice would be lost. I know some famous sites have used this convention such as http://searchenginewatch.com/ so think it would be possible but still concerned that for new sites it would be better to rather stick to conventions. What are your opinions about this?
Intermediate & Advanced SEO | | SABest0