Www v.s non www
-
The canonical URLs (and all our link building efforts) is on the www version of the site.
However, the site is having a massive technical problem and need to redirect some links (some of which are very important) from the www to the non www version of the site (for these pages the canonical link is still the www version).
How big of a SEO problem is this?
Can you please explain the exact SEO dangers?
Thanks!
-
Thanks for all your responses - I will use this as the basis of my answer to the technical team.
-
I'm endorsing Stephen's idea, because if you really have no choice, I think it's a good potential alternative. THB's comments (which I thumbed up) are very important, though.
If you really have no choice, I do think the 302 is safer here - the canonical tag should override it. There is some risk, though, and it's definitely not ideal.
I'm not clear on the problem, but could you return a 503? It basically says "We've got a temporary problem - come back later" and, if it really is temporary, Google won't de-index the pages. If you're talking a couple of days, this may be a better solution. If you're talking a few weeks, you may have to take Stephen's advice. You might want to pull in expert help, though, because my gut reaction is that there's a better way to fix what's broken here.
-
Hehe.
Generally speaking, and I've actually come across this quite a bit lately, it's better to just put your efforts towards fixing the technical issues than to try and manipulate the site using redirects and canonical tags. But it's easy to say when it's not my technical problem, nor my money/time on the line to fix it! However, that is always the best-case scenario in my opinion.
-
Agreed. It's a problem waiting to bite you in the proverbials....
-
I worry about setting up a canonical tag that points to a URL Google can't access (as it's just being redirected via 302 back to the non-www version anytime it will try and read the canonical URL). And since a canonical tag is kinda sorta like a 301, you'd ultimately be 301'ing (kinda sorta) back to the www version, only to have a 302 header sent, 302'ing Google back to the non-www. And endless loop, so-to-speak. I'm not sure how Google would handle this.
How about just working 24/7 to resolve the "technical problem" that is causing this? I know, easy for me to say
-
I'm no expert on this but I think you'll be fine IF you:
1 - 302 redirect (temporary redirect) to the non-www page
2 - Add a rel canonical on the non-www page giving the WWW version link credit.
When you've fixed your tech issues remove the 302 redirect.
I THINK google will play nice on this.
Hope that helps
Steve
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Huge number of crawl anomalies and 404s - non- existent urls
Hi there, Our site was redesigned at the end of January 2020. Since the new site was launched we have seen a big drop in impressions (50-60%) and also a big drop in total and organic traffic (again 50-60%) when compared to the old site. I know in the current climate some businesses will see a drop in traffic, however we are a tech business and some of our core search terms have increased in search volume as a result of remote-working. According to search console there are 82k urls excluded from coverage - the majority of these are classed as 'crawl anomaly' and there are 250+ 404's - almost all of the urls are non-existent, they have our root domain with a string of random characters on the end. Here are a couple of examples: root.domain.com/96jumblestorebb42a1c2320800306682 root.domain.com/01sportsplazac9a3c52miz-63jth601 root.domain.com/39autoparts-agency26be7ff420582220 root.domain.com/05open-kitchenaf69a7a29510363 Is this a cause for concern? I'm thinking that all of these random fake urls could be preventing genuine pages from being indexed / or they could be having an impact on our search visibility. Can somebody advise please? Thanks!
Technical SEO | | nicola-10 -
Google tries to index non existing language URLs. Why?
Hi, I am working for a SAAS client. He uses two different language versions by using two different subdomains.
Technical SEO | | TheHecksler
de.domain.com/company for german and en.domain.com for english. Many thousands URLs has been indexed correctly. But Google Search Console tries to index URLs which were never existing before and are still not existing. de.domain.com**/en/company
en.domain.com/de/**company ... and an thousand more using the /en/ or /de/ in between. We never use this variant and calling these URLs will throw up a 404 Page correctly (but with wrong respond code - we`re fixing that 😉 ). But Google tries to index these kind of URLs again and again. And, I couldnt find any source of these URLs. No Website is using this as an out going link, etc.
We do see in our logfiles, that a Screaming Frog Installation and moz.com w opensiteexplorer were trying to access this earlier. My Question: How does Google comes up with that? From where did they get these URLs, that (to our knowledge) never existed? Any ideas? Thanks 🙂0 -
Conversion of URL's for Readability
Reading over Rands latest Post about URL structure I had a quick question about the best way to convert URL's that don't have perfect URL structure... Current the Structure of our E-commerce store has a structure that is not friendly with domain.com/product/zdcd-jobd3d-fdoh what is the easiest way to convert these to read URL's without causing any disruptions with the SERP. Are we talking about a MOD-Rewrite in the CMS.......
Technical SEO | | CMcMullen0 -
Link building to ROOT domain OR to WWW.?
Hello, Here I come with one more 'sensitive' question, hoping that you SEO gurus could give some input on. My title explains pretty much what I'm wondering about, but let me give you some short data. I have from .htaccess file set that all traffic goes to WWW.mydomain.com. I know that it is 'better' for search engines not to have duplicate destinations as that can give decreased page rank because of 'double content'. As for search engines http://domain.com and http://www.domain.com is totally different domains. Now wondering one thing: If I build a several thousands of backlinks at various sources, blogs, directories, web sites etc etc. - shall I link to domain ROOT or shall I include WWW prefix? When looking at Moz Keyword Analysis for my domains, I can see a block about 'Linking Root Domains' and 'Page Linking Root Domains'. But no 'www' variable (sub-domain) there. As I have already set canonical part so everything shows with WWW on my website - what logic shall I use when building backlinks? How will search engine translate the link juice in regards I wrote above? Thanks in advance, great forum!
Technical SEO | | SEOisSEO0 -
Why would an image that's much smaller than our Website header logo be taking so long to load?
When I check http://www.ccisolutions.com at Pingdom, we have a tiny graphic that is taking much longer to load than other graphics that are much bigger. Can anyone shed some light on why this might be happening and what can be done to fix it? Thanks in advance! Dana
Technical SEO | | danatanseo0 -
What's the latest on Title Tags?
What is the latest on what Google is looking for? Keyword one, Keyword two? Sentences with the Keyword in them?
Technical SEO | | netviper0 -
Different IP's in one Server
Hi, I just want to ask if there is no bad effect in SEO if we do have different websites that has different IP address but has shared in only 1 server? Thank you
Technical SEO | | TirewebMarketing0 -
Issue with Joomla Site not showing in SERP's
Site: simpsonelectricnc dot com I'm working on a Joomla website for a local business that isn't ranking at all for any relevant keyword - including the business name. The site is only about six months old and has relatively few links. I realize it takes time to compete for even low-volume keywords, but I think something else may be preventing the site from showing up. The site is not blocked by Robots.txt (which includes a valid reference to the sitemap)
Technical SEO | | CGR-Creative
There is no duplicate content issue, the .htaccess is redirecting all non-www traffic to www version
Every page has a unique title and H1 tag.
The URL's are search-engine friendly (not dynamic either)
XML sitemap is live and submitted to Google WM Tools. Google shows that it is indexing about 70% of the submitted URL's. The site has essentially no domain authority (0.02) according to Moz - I'm assuming this is due to lack of links and short life on the web.
Until today, 98% of the pages had identical meta descriptions. Again, I realize six months is not an eternity - but the site will not even show up for "business name + city,state" searches in Google. In fact, the only way I've seen it in organic results is to search for the exact URL. I would greatly appreciate any help.0