How will google respond to allowing multilingual search terms for a single language website?
-
We would like to set up a website in English language only and promote this in various European countries. As said the website will only be available in English language, but we will keep translations (google translate) in backend. When a user in France then enters search query in French language in browser, a search can be done in French content, but we will present relevant content in English. Does anyone have any experience with that? Will it be allowed given the fact that the result (in English language) will probably not include any of the terms that was searched on (in French language).
-
It depends upon which service you're using to translate.
And after translate, the URL's remains the same or it changes.
If it changes then it will be discovered by Google. -
this is very useful
-
I'm also interested in this topic because im using wordpress right now and i use polylang for all my translations and till now haven't had any problems with google
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What should I do after a failed request for validation (error with noindex, nofollow) in new Google Search Console?
Hi guys, We have the following situation: After an error message in new google search console for a large amount of pages with noindex, nofollow tag, a validation is requested before the problem is fixed. (it's incredibly stupid decision taken before asking the SEO team for advice) Google starts the validation, crawls 9 URLs and changes the status to "Failed". All other URLs are still in "pending" status. The problem has been fixed for more than 10 days, but apparently Google doesn't crawl the pages and none of the URLs is back in the index. We tried pinging several pages and html sitemaps, but there is no result. Do you think we should request for re-validation or wait more time? It there something more we could do to speed up the process?
Intermediate & Advanced SEO | | ParisChildress0 -
How to maximize CTR from Google image search?
I'm getting good, solid growth in my Google SERPs and Google search traffic now, but I do notice that 70% of my high ranking search results are images and the CTR on those is only 3-4%. All my images are illustrative and highly relevant to my travel blog, but I guess that hardly matters unless they get CTR so people see them in context. Has anyone seen or done any good research on what makes people click through on Google Image Search results? What are the key factors? How do you optimize for click-through? Is it better to watermark your images or overlay label them to increase likelihood of click-through? Thanks, Tony FYI the travel blog in question is www.asiantraveltips.com and a relevant Google search where I rank highly is "songkran 2016 phuket".
Intermediate & Advanced SEO | | Gavin.Atkinson0 -
Best way to remove full demo (staging server) website from Google index
I've recently taken over an in-house role at a property auction company, they have a main site on the top-level domain (TLD) and 400+ agency sub domains! company.com agency1.company.com agency2.company.com... I recently found that the web development team have a demo domain per site, which is found on a subdomain of the original domain - mirroring the site. The problem is that they have all been found and indexed by Google: demo.company.com demo.agency1.company.com demo.agency2.company.com... Obviously this is a problem as it is duplicate content and so on, so my question is... what is the best way to remove the demo domain / sub domains from Google's index? We are taking action to add a noindex tag into the header (of all pages) on the individual domains but this isn't going to get it removed any time soon! Or is it? I was also going to add a robots.txt file into the root of each domain, just as a precaution! Within this file I had intended to disallow all. The final course of action (which I'm holding off in the hope someone comes up with a better solution) is to add each demo domain / sub domain into Google Webmaster and remove the URLs individually. Or would it be better to go down the canonical route?
Intermediate & Advanced SEO | | iam-sold0 -
Can SPA (single page architecture) websites be SEO friendly?
What is the latest consensus on SPA web design architecture and SEO friendliness?
Intermediate & Advanced SEO | | Robo342
By SPA, I mean rather than each page having its own unique URL, instead each page would have an anchor added to a single URL. For example: Before SPA: website.com/home/green.html After SPA: website.com/home.html#green (rendering a new page using AJAX) It would seem that Google may have trouble differentiating pages with unique anchors vs unique URLs, but have they adapted to this style of architecture yet? Are there any best practices around this? Some developers are moving to SPA as the state of the art in architecture (e.g., see this thread: http://www.linkedin.com/groups/Google-crawling-websites-built-using-121615.S.219120193), and yet there may be a conflict between SPA and SEO. Any thoughts or black and white answers? Thanks.0 -
How important is the optional <priority>tag in an XML sitemap of your website? Can this help search engines understand the hierarchy of a website?</priority>
Can the <priority>tag be used to tell search engines the hierarchy of a site or should it be used to let search engines know which priority to we want pages to be indexed in?</priority>
Intermediate & Advanced SEO | | mycity4kids0 -
Website layout for a new website [Over 50 Pages & targeting Long Tail Keywords]
Hey everyone, We are designing a new website with over 50 pages and I have a question regarding the layout. Should I target my long tail keywords via blog pages? It will be easier to manage and list and link out to similar articles related to my long tail keywords using a word press blog. For this example - lets suppose the website is www.orange.com and we sells 'Oranges' Am I going about this in the right way? Main Section: Main Section 1 : Home Page - Keyword Targeted - Orange Main Section 2 : Important Conversion page - 'Buy oranges' Long Tail Keyword (LTK) 1: www.orange.com/blog/LTK1 Subsection(SS): www.orange.com/blog/LTK1/SS1 www.orange.com/blog/LTK1/SS1a www.orange.com/blog/LTK1/SS1b Long Tail Keyword (LTK) 2: www.orange.com/blog/LTK2 Long Tail Keyword (LTK) 3: www.orange.com/blog/LTK3 Subsection(SS): www.orange.com/blog/LTK1/SS3 www.orange.com/blog/LTK1/SS3a www.orange.com/blog/LTK1/SS3b All these long tail pages and sub sections under them are built specifically for hosting content that targets these specific long tail keywords. Most of my traffic will come initially via the sub section pages - and it is important for me to rank well for these terms initially. _E.g. if someone searches for the keyword 'SS3b' on Google - my corresponding page www.orange.com/blog/LTK1/SS3b should rank well on the results page. _ For ranking purposes - will using this blog/category structure hurt or benefit me? Instead do you think I should build static pages? Also, we are targeting more than 50 long tail keywords - and building quality content for each of these keywords - and I assume that we will be doing this continuously. So in the long term term which is more beneficial? Do you have any suggestions on if I am going about this the right way? Apologies for using these random terms - oranges, LKT, SS etc in this example. However, I hope that the question is clear. Looking forward to some interesting answers on this! Please feel free to share your thoughts.. Thank you! Natasha
Intermediate & Advanced SEO | | Natashadogres0 -
Top Google News Result = Search Result #11 (first on second page)
Hey all, I've noticed that, in most cases, when we have an article that gets the top spot in Google News results for a given keyword, the search result for that same article will appear in position #11 (the first result on the second page for standard SERP viewing). This is nearly always the case, which suggests its built into Google's algorithm to prevent overlap. Has anyone else experienced this? I haven't seen it discussed previously on Moz or other SEO forums, but it makes sense. Or if you haven't experienced this, I'd love to hear about what you're seeing.
Intermediate & Advanced SEO | | dangaul0 -
Could Profanity Decrease Google Trust in a Website?
Could profanity such as swear words or cursing decrease the trust Google has in a website? For example, abusive comments... etc? I'm not talking about the SafeSearch filter, I'm talking about the overall trust of the website. Your thoughts and suggestions are greatly appreciated.
Intermediate & Advanced SEO | | Peter2640