Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Other domains hosted on same server showing up in SERP for 1st site's keywords
-
For the website in question, the first domain alphabetically on the shared hosting space, strange search results are appearing on the SERP for keywords associated with the site.
Here is an example:
A search for "unique company name" shows the results: www.uniquecompanyname.com as the top result.
But on pages 2 and 3, we are getting results for the same content but for domains hosted on the same server. Here are some examples with the domain name replaced:
UNIQUE DOMAIN NAME PAGE TITLE
ftp.DOMAIN2.com/?action=news&id=63
META DESCRIPTION TEXTUNIQUE DOMAIN NAME PAGE TITLE 2
www.DOMAIN3.com/?action=news&id=120
META DESCRIPTION TEXT2UNIQUE DOMAIN NAME PAGE TITLE 2
www.DOMAIN4.com/?action=news&id=120
META DESCRIPTION TEXT2UNIQUE DOMAIN NAME PAGE TITLE 3
mail.DOMAIN5.com/?action=category&id=17
META DESCRIPTION TEXT3ns5.DOMAIN6.com/?action=article&id=27
There are more but those are just some examples.
These other domain names being listed are other customer domains on the same VPS shared server. When clicking the result the browser URL still shows the other customer domain name B but the content is usually the 404 page. The page title and meta description on that page is not displayed the same as on the SERP.As far as we can tell, this is the only domain this is occurring for.So far, no crawl errors detected in Webmaster Tools and moz crawl not completed yet.
-
For those that are interested, we figured out the root of the problem and the fix. Maybe this will help someone down the road. Thanks to all that responded.
As it seems all the domains/accounts hosted on our VPS are top level; so we couldn’t use robots.txt to fix this.The problem:
-We have a handful of websites using a shared IP address – Only sites with SSL cert get a dedicated IP
-Some of these websites have additional A record DNS names (mail.domain1, ns5.domain2, etc) that all resolve to that shared IP address
-Within your browser if you went to one of those records it would pull up the website (account) that is first alphabetically within that shared IP pool
-It’s assumed that a website or multiple websites somewhere in the web must have linked these domain names somewhere on their webpages – so search engines crawled them for the content of that first alphabetical website
-Search engines then indexed the content of that site for these other DNS A records and displayed the alternate URLs in the SERPSThe fix:
Edit Apache’s httpd.conf file on our VPS – adding a new (first listed) virtual directory of that shared IP address to point to the servers cpanel default cgi page instead of listing one of our hosted accounts first.
This page describes what we needed done and what fixed it for us: http://forums.spry.com/cpanel-whm/1568-how-can-i-change-default-page-ip-address-points.html
Thanks again!
-
You should be able to resolve this with robots.txt and .htaccess files, but one more thing I've learned to do is to use a domain name for the primary domain that I do not plan to submit to Google for indexing. For example, I'll use arbitrary color names like silverblue.net as the primary domain and then the domains I intend to index will be virtual domains. This isn't really necessary but it helps me keep all my domain names straight in my head and avoid the issue you described above.
Best,
Christopher -
Thanks for the replies, guys. We are going to try Christopher's suggestion first and we'll update the question when it's resolved!
-
The other domains are ours, yes.
Oh, that's a horse of a different color. If you have multiple domains on one account, one domain will be the primary domain and the other domains will be virtual domains. If you look at the file structure on your server, all the virtual domains will be sub-folders under the primary domain. You need to create a robots.txt file in the primary domain that blocks the spiders from discovering the other domains through the sub-directory structure. Otherwise, all the virtual domains will be duplicated, once via the primary domain and again via the virtual domain.
Best,
Christopher -
The other domains are ours, yes.
I don't think any of them are using Joomla.
-
Are those other domains also yours, by any chance and/or is this a Joomla cms?
-
Wow... that's not so ideal. My guess is that all of those sites are sharing the same IP address, and the server's settings are such that if you hit the IP address with a forward slash (i.e. http://00.00.00.00/domainname/) it's showing content.
My $0.02 is to find a new hosting company.... and one that gives each site it's own IP address instead of doing shared-level hosting with the same IP address in a name-based space.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Subdomain replaced domain in Google SERP
Good morning, This is my first post. I found many Q&As here that mostly answer my question, but just to be sure we do this right I'm hoping the community can take a peak at my thinking below: Problem: We are relevant rank #1 for "custom poker chips" for example. We have this development website on a subdomain (http://dev.chiplab.com). On Saturday our live 'chiplab.com' main domain was replaced by 'dev.chiplab.com' in the SERP. Expected Cause: We did not add NOFOLLOW to the header tag. We also did not DISALLOW the subdomain in the robots.txt. We could have also put the 'dev.chiplab.com' subdomain behind a password wall. Solution: Add NOFOLLOW header, update robots.txt on subdomain and disallow crawl/index. Question: If we remove the subdomain from Google using WMT, will this drop us completely from the SERP? In other words, we would ideally like our root chiplab.com domain to replace the subdomain to get us back to where we were before Saturday. If the removal tool in WMT just removes the link completely, then is the only solution to wait until the site is recrawled and reindexed and hope the root chiplab.com domain ranks in place of the subdomain again? Thank you for your time, Chase
Intermediate & Advanced SEO | | chiplab0 -
Does Google Read URL's if they include a # tag? Re: SEO Value of Clean Url's
An ECWID rep stated in regards to an inquiry about how the ECWID url's are not customizable, that "an important thing is that it doesn't matter what these URLs look like, because search engines don't read anything after that # in URLs. " Example http://www.runningboards4less.com/general-motors#!/Classic-Pro-Series-Extruded-2/p/28043025/category=6593891 Basically all of this: #!/Classic-Pro-Series-Extruded-2/p/28043025/category=6593891 That is a snippet out of a conversation where ECWID said that dirty urls don't matter beyond a hashtag... Is that true? I haven't found any rule that Google or other search engines (Google is really the most important) don't index, read, or place value on the part of the url after a # tag.
Intermediate & Advanced SEO | | Atlanta-SMO0 -
How to combine 2 pages (same domain) that rank for same keyword?
Hi Mozzers, A quick question. In the last few months I have noticed that for a number of keywords I am having 2 different pages on my domain show up in the SERP. Always right next to each other (for example, position #7 and #8 or #3 and #4). So in the SERP it looks something like: www.mycompetition1.com www.mycompetition2.com www.mywebsite.com/page1.html
Intermediate & Advanced SEO | | rayvensoft
4) www.mywebsite.com**/page2.html**
5) www.mycompetition3.com Now, I actually need both pages since the content on both pages is different - but on the same topic. Both pages have links to them, but page1.html always tends to have more. So, what is the best practice to tell Google that I only want 1 page to rank? Of course, the idea is that by combining the SEO Juice of both pages, I can push my way up to position 2 or 1. Does anybody have any experience in this? Any advice is much appreciated.0 -
How can I get a list of every url of a site in Google's index?
I work on a site that has almost 20,000 urls in its site map. Google WMT claims 28,000 indexed and a search on Google shows 33,000. I'd like to find what the difference is. Is there a way to get an excel sheet with every url Google has indexed for a site? Thanks... Mike
Intermediate & Advanced SEO | | 945010 -
Do links to PDF's on my site pass "link juice"?
Hi, I have recently started a project on one of my sites, working with a branch of the U.S. government, where I will be hosting and publishing some of their PDF documents for free for people to use. The great SEO side of this is that they link to my site. The thing is, they are linking directly to the PDF files themselves, not the page with the link to the PDF files. So my question is, does that give me any SEO benefit? While the PDF is hosted on my site, there are no links in it that would allow a spider to start from the PDF and crawl the rest of my site. So do I get any benefit from these great links? If not, does anybody have any suggestions on how I could get credit for them. Keep in mind that editing the PDF's are not allowed by the government. Thanks.
Intermediate & Advanced SEO | | rayvensoft0 -
Why does a site have no domain authority?
A website was built and launched eight months ago, and their domain authority is 1. When a site has been live for a while and has such a low DA, what's causing it?
Intermediate & Advanced SEO | | optimalwebinc0 -
How to prevent 404's from a job board ?
I have a new client with a job listing board on their site. I am getting a bunch of 404 errors as they delete the filled jobs. Question: Should we leave the the jobs pages up for extra content and entry points to the site and put a notice like this job has been filled, please search our other job listings ? Or should I no index - no follow these pages ? Or any other suggestions - it is an employment agency site. Overall what would be the best practice going forward - we are looking at probably 20 jobs / pages per month.
Intermediate & Advanced SEO | | jlane90 -
There's a website I'm working with that has a .php extension. All the pages do. What's the best practice to remove the .php extension across all pages?
Client wishes to drop the .php extension on all their pages (they've got around 2k pages). I assured them that wasn't necessary. However, in the event that I do end up doing this what's the best practices way (and easiest way) to do this? This is also a WordPress site. Thanks.
Intermediate & Advanced SEO | | digisavvy0