Control indexed content on Wordpress hosted blog...
-
I have a client with a blog setup on their domain (example: blog.clientwebsite.com) and even though it loads at that subdomain it's actually a Wordpress-hosted blog. If I attempt to add a plugin like Yoast SEO, I get the attached error message. Their technical team says this is a brick wall for them and they don't want to change how the blog is hosted.
So my question is... on a subdomain blog like this... if I can't control what is in the sitemap with a plugin and can't manually add a sitemap because the content is being pulled from a Wordpress-hosted install, what can I do to control what is in the index?
I can't add an SEO plugin...
I can't add a custom sitemap...
I can't add a robots.txt file...
The blog is setup with domain mapping so the content isn't actually there. What can I do to avoid tags, categories, author pages, archive pages and other useless content ending up in the search engines?
-
That almost looks like... your client doesn't have WordPress actually installed on their sub-domain at all. It looks like they set up a 'something.wordpress.com' site, which WordPress actually hosts - and somehow overlayed their own sub-domain over it (using DNS / name-server shenanigans)
If that is true then, since WordPress hosts the blog, there's not much you can do. If it is a local WordPress install that does exist on your client's actual website instead of being 'framed' in (or something shady like that) - then I haven't seen this error before and it seems really odd. It smacks of someone trying to cut corners with their hosting environment, trying to 'be clever' instead of shelling out for a proper WP install. Clearly there are limitations...
Ok, there's only one other alternative really. This is also technical though and I don't know if it wold be any easier for your dev guys but...
You can send no-index directives to Google without altering the site / web-page coding, as long as you are willing to play around with the (server-level) HTTP headers
There's something called X-Robots which might be useful to you. You need to read this post here (from Google). You need to start reading from (Ctrl+F for): "Using the X-Robots-Tag HTTP header"
As far as I know, most meta-robots indexation tag directives, can also be fired through the HTTP header using X-robots
It's kinda crazy but, it might be your only option
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sitemap For Static Content And Blog
We'll be uploading a sitemap to google search console for a new site. We have ~70-80 static pages that don't really chance much (some may change as we modify a couple pages over the course of the year). But we have a separate blog on the site which we will be adding content to frequently. How can I set up the sitemap to make sure that "future" blog posts will get picked up and indexed. I used a sitemap generator and it picked up the first blog post that's on the site, but am wondering what happens with future ones? I don't want to resubmit a new sitemap each time that has a link to a new blog post we posted.
Technical SEO | | vikasnwu0 -
Database driven content producing false duplicate content errors
How do I stop the Moz crawler from creating false duplicate content errors. I have yet to submit my website to google crawler because I am waiting to fix all my site optimization issues. Example: contactus.aspx?propid=200, contactus.aspx?propid=201.... these are the same pages but with some old url parameters stuck on them. How do I get Moz and Google not to consider these duplicates. I have looked at http://moz.com/learn/seo/duplicate-content with respect to Rel="canonical" and I think I am just confused. Nick
Technical SEO | | nickcargill0 -
Can't get Google to Index .pdf in wp-content folder
We created an indepth case study/survey for a legal client and can't get Google to crawl the PDF which is hosted on Wordpress in the wp-content folder. It is linked to heavily from nearly all pages of the site by a global sidebar. Am I missing something obvious as to why Google won't crawl this PDF? We can't get much value from it unless it gets indexed. Any help is greatly appreciated. Thanks! Here is the PDF itself:
Technical SEO | | inboundauthority
http://www.billbonebikelaw.com/wp-content/uploads/2013/11/Whitepaper-Drivers-vs-cyclists-Floridas-Struggle-to-share-the-road.pdf Here is the page it is linked from:
http://www.billbonebikelaw.com/resources/drivers-vs-cyclists-study/0 -
Duplicate Content Problem!
Hi folks, I have a quite awkward problem. Since a few weeks a get a huge amount of "duplicate content errors" in my MOZ crawl reports. After a while of looking for the error I thought of the domains I've bought additionally. So I went to Google and typed in site:myotherdomains.com The results was as I expected that my original website got indexed with my new domains aswell. That means: For example my original website was index with www.domain.com/aboutus - Then I bought some additional domains which are pointing on my / folder. What happened is that I also get listed with: www.mynewdomains.com/com How can I fix that? I tried a normal domain redirect but it seems as this doesn't help as when I am visiting www.mynewdomains.com the domain doesnt change in my browser to www.myoriginaldomain.com but stays with it ... I was busy the whole day to find a solution but I am kinda desperate now. If somebody could give me advice it would be much appreciated. Mike
Technical SEO | | KillAccountPlease0 -
Acquiring a blog
Hello All, I've recently acquired somebody else's blog and have redirected every post to the relevant page of my website (madegood.org). The content is the same as on the original site, and I have used 301 redirects. The original blog didn't have a particularly high page rank I'm slightly worried that there are now thousands of links coming from one domain, which itself doesn't have much authority. Is there a way that I can tell google that I've acquired the blog, as opposed to just having lots of links from one domain. Thanks Will
Technical SEO | | madegood0 -
Getting More Pages Indexed
We have a large E-commerce site (magento based) and have submitted sitemap files for several million pages within Webmaster tools. The number of indexed pages seems to fluctuate, but currently there is less than 300,000 pages indexed out of 4 million submitted. How can we get the number of indexed pages to be higher? Changing the settings on the crawl rate and resubmitting site maps doesn't seem to have an effect on the number of pages indexed. Am I correct in assuming that most individual product pages just don't carry enough link juice to be considered important enough yet by Google to be indexed? Let me know if there are any suggestions or tips for getting more pages indexed. syGtx.png
Technical SEO | | Mattchstick0 -
Old Blog
I have an old blog that I started long ago and it has tons of content. I'm thinking about migrating it my current blog but am worried about panda and bringing over mediocre content. The content is fine, not bad not good. Should I bring it over or should I just delete the blog?
Technical SEO | | tylerfraser0