How do we handle sitemaps in robots.txt when multiple domains point to same physical location?
-
we have www.mysite.net, www.mysite.se, www.mysite.fi and so on. all of these domains point to the same physical location on our webserver, and we replace texts given back to client depending on which domain he/she requested.
My problem is this: How do i configure sitemaps in robots.txt when robots.txt is used by multiple domains? If I for instance put the rows
Sitemap: http://www.mysite.net/sitemapNet.xml
Sitemap: http://www.mysite.net/sitemapSe.xmlin robots.txt, would that result in some cross submission error?
-
Thanks for your help René!
-
yup
-
Yes, I mean GTW of course :).
A folder for each site would definitely make some things easier, but it would also mean more work every time we need to republish the site or make configurations.
Did I understand that googlelink correctly in that if we have verified ownership in GWT for all involved domains cross-site submission in robots.txt was okay? I guess google will think its okay anyway.
-
actually google has the answer, right here: http://www.google.com/support/webmasters/bin/answer.py?hl=en&answer=75712
I always try to do what google recommends even though something might work just as well.. just to be on the safe side
-
you can't submit a sitemap in GA so I'm guessing you mean GWT
Whether or not you put it in the robots.txt shouldn't be a problem. since in each sitemap, the urls would look something like this:
Sitemap 1:<url><loc>http:/yoursite.coim/somepage.html</loc></url>
Sitemap 2:<url><loc>http:/yoursite.dk/somepage.html</loc></url>
I see no need to filter what sitemap is shown to the crawler. If your .htaccess is set-up to redirect traffic from the TLD (top level domain eg .dk .com ex.) to the correct pages. Then the sitemaps shouldn't be a problem.
The best solution would be: to have a web in web. (a folder for each site on the server) and then have the htaccess redirect to the right folder. in this folder you have a robots.txt and a sitemap for that specific site. that way all your problems will be gone in a jiffy. It will be just like managing different 3 sites. even though it isn't.
I am no ninja with .htaccess files but I understand the technology behind it and know what you can do in them. for a how to do it guide, ask google thats what I allways do when I need to goof around in the htaccess. I hope it made sense.
-
Thanks for your response René!
Thing is we already submit the sitemaps in google analytics, but this SEO company we hired wants us to put the sitemaps in robots.txt as well.
The .htaccess idea sounds good, as long as google or someone else dont think we are doing some cross-site submission error (as described here http://www.sitemaps.org/protocol.php#submit_robots)
-
I see no need to use robots.txt for that. use Google and Bings webmaster tools. Here you have each domain registered and can submit sitemaps to them for each domain.
If you want to make sure that your sitemaps are not crawled by a bot for a wrong language. I would set it up in the .htaccess to test for the entrance domain and make sure to redirect to the right file. Any bot will enter a site just like a browser so it needs to obey the server. so if the server tells it to go somewhere it will.
the robots.txt can't by it self, do what you want. The server can however. But in my opinion using bing and google webmaster tools should do the trick.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Keyword in Domain AND Title. Yes or No?
We're working on a new buildout, and this one is really important to us. We've put a lot of resources into it. Before we launch, we want the structure to be just right... and this one question is nagging at me. How to structure urls? Consider these two options. The fictitious domain is "icesurfing.org". Including all 50 states in the keyword, there are nearly one million searches per month for "ice surfing [state]". We have a page for each state to focus on this traffic. But how would you structure the urls and titles? **icesurfing.org/state ** icesurfing.org/ice-surfing-state One concern is that the duplicate keywords in option 2 seem redundant, and a little spammy. When presented with google search, the matching tags are not as clean. Texas - IceSurfing.org Ice Surfing Texas - IceSurfing.org But Yoast automatically suggests option 2. Is this really the best practice? Is there are definitive article on this? THANK YOU!
On-Page Optimization | | RetBit0 -
Multiple and tags
It has recently come to our attention that some of the pages on our site have multiple and tags. I understand that this goes against best practices-- but I was wondering if anyone could tell me specifically how this could be impacting our site's performance and search rankings.
On-Page Optimization | | ZoomInformation0 -
SEO for page that is made up of multiple posts
When a page is made up of lots of individual wordpress posts, how should it be handled from an SEO perspective? Should the individual posts be hidden from the search engines and only the page url which contains all the posts collectivity be indexed? Page in question below: http://teammateworld.com/case-studies/ Thanks in advance.
On-Page Optimization | | Rich_H0 -
Using Robots Meta Tag on Review Form Pages
I have gone over this so many times and I just can't seem to get it straight and hope someone can help me out with a couple of questions: Right now, on my dynamically created pages created by filters (located on the category pages) I am using rel""canonical" to point them to their respective category page. Should I also use the robots meta tag as well? Similarly, each product I have on my site has a review form on it and thus is getting indexed by Google. I have placed the same canonical tag on them as well pointing them to the page with the review form on it. In the past I used robots.txt to block google from the review pages but this didn't really do much. Should I be using the robots meta tag on these pages as well? If I used the robots meta tag should I noindex,nofollow? Thanks in advance, Jake
On-Page Optimization | | jake3720 -
Our sitemap is not indexed well
Hey there, Hope you guys can help. We get the following error: Nested indexing. Another Sitemap index refers to the index of sitemaps. The thing is that we cant find the error they are talking about. Thanks!!!!
On-Page Optimization | | Comunicare0 -
Exact Match Domain + shorter permalink vs. longer permalink?
So here's the scenario... You own the exact match domain for "Acupuncture Intake Form" and want to create a page that targets the phrase, "Acupuncture Intake Form Template"... In terms of SEO, which of the following permalinks does a better job targeting the term "Acupuncture Intake Form Template" : A) www.acupunctureintakeform.com/template/ B) www.acupunctureintakeform.com/acupuncture-intake-form-template/ From a user-friendliness perspective, I can definitely see why the /template/ is more ideal. It's more memorable, easier to link to, and logical. I'm just wondering whether there is some edge that the /acupuncture-intake-form-template/ has that I'm not aware of, or if there's some gray area. I can also think of scenarios where the longer version might be better. For example, if there were several different kinds of acupuncture intake form templates. If that were the case, then /template/ might be too vague (e.g. is it the normal "acupuncture intake form template" or is it the "acupuncture intake form template with diagram"). So in that respect, the longer one might be less concise in certain situations, but in others provide more clarity. (Note: I'm looking for answers to this situation as it applies to any website. I just chose this extremely small niche exact match domain because it's an example that illustrate my questions). Thanks in advance for any answers, insights, or comments! I'm not sure if there's a specific answer, but I'm sure there are some key points to discuss. Michael
On-Page Optimization | | InRakeLLC0 -
Blog outgoing links to a certain domain?
Hi Mozzers, I am working with a website with very decentralized ownership. There are two different languages, each with a different owner. Owner A keeps linking to crap sites, that hurt the entire site. My question is this: Is there a way - through .htaccess or robots.txt - that Google can be asked NOT to crawl the links to external crap sites? The problem is that Owner B cannot control Owner A's html, and thus not implement rel="nofollow" on links. Thanks!
On-Page Optimization | | ThomasHgenhaven0 -
Installing a site on top level domain directory VS deeper directory
How important it is for seo to install a site on a top level directory vs deeper directory? example: www.mysite.com VS www.mysite.com/catalog many eCommerce scripts such as oscommerce and cre loaded will install by default to /catalog and that's what I've been doing for most of my customer. does it mean it will be harder for them to get good seo results? Thanks in advance for the input...
On-Page Optimization | | zigi0