Why do some reputable publishers have problems with their microdata?
-
I'm using the Google Structured Data Testing Tool to test: https://search.google.com/structured-data/testing-tool
NY Times and Women's Health being two good examples.
These two reputable publishers don't seem to have the microdata they've implemented recognized. Are they doing something wrong or is there a problem with the tool?
-
Haven't tried that, but that's not a bad idea, thanks. I assumed I wouldn't hear back.
-
Basically what I mean is if the validator made by Google isn't validating, that would indicate that the code should be cleaned up right?
-
Interesting. Could this mean that the structured data is somehow being blocked. Should the Google Structured Data tool be the gold standard? What does it mean if it can't read the structured data?
-
Sure, this is a typical example: https://search.google.com/structured-data/testing-tool#url=http%3A%2F%2Fwww.womenshealthmag.com%2Flife%2Fthis-bride-walked-down-the-aisle-with-the-man-who-got-her-fathers-donor-heart
This link won't even get validated in the tool: http://www.nytimes.com/interactive/2016/08/10/sports/olympics/gymnastics-parents.html?hp&action=click&pgtype=Homepage&clickSource=image&module=photo-spot-regionion=top-news&WT.nav=top-news&_r=0
This one is using schema.org markup but its not being recognized in the structured data testing tool: http://www.msn.com/en-us/sports/olympics/hope-solos-mistake-helps-colombia-tie-usa/ar-BBvsb4J?li=BBnb7Kz
Hell, check SI.com, it doesn't even use RDF, microdata, or JSON. I don't get it.
-
It looks like Women's Health is using RDFa. Why they would be using this instead of JSON-LD or microdata I don't know...
They're not even doing it right according to the google structured data testing tool.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Having possible problems with rankings due to development website
Hi all, I've got an interesting issue and a bit of a technical challenge for you. It's a bit complicated to explain, but please bear with me. We have a client website (http://clientwebsite.com) which we are having a hard time ranking in the past few months. Main keywords simply don't show up in Top100 searches, even though we are constantly building backlinks through Guest Posts, Citations, Media mentions, Profile links etc. Normally, we use ahrefs to look at the client's website backlinks, but just today we used Majestic to look at the backlink profile and one backlink stood out. This is a backlink from a development server (http://developmentwebsite.com) which redirects to http://clientwebsite.com
Intermediate & Advanced SEO | | zakkyg
The developers who were working on the redesign of the client website, put it up on their server and forgot to delete it.
Also, the content inside the development website is almost identical with the client website. We then checked to see if http://developmentwebsite.com is indexed.
It's not. Although, inside the robots file http://developmentwebsite.com/robots.txt there's:
User-agent: *
Allow: /
The funny (and weird thing) is that http://developmentwebsite.com/ and all development website inner pages are not indexed in Google. But if we go to http://developmentwebsite.com/inner-page, it doesn't redirect to the corresponding http://clientwebsite.com/inner-page, it's the same development website page URL and the pages even have links to the client website, but like I said, none of the pages of the development website are indexed, even though crawlers are allowed in the robots.txt's development website. In your opinion, could this be the reason why we are having a hard time to rank the client website? Second question is:
How do we approach in solving this issue?
Do we simply delete the whole http://developmentwebsite.com with all the inner pages?
Or should we do 301 redirrects on a per-page basis?0 -
Content Publishing Volume/Timing
I am working with a company that has a bi-monthly print magazine that has several years' worth of back issues. We're working on building a digital platform, and the majority of articles from the print mag - tips, how-tos, reviews, recipes, interviews, etc - will be published online. Much of the content is not date-sensitive except for the occasional news article. Some content is semi-date-sensitive, such as articles focusing on seasonality (e.g. winter activities vs. summer activities). My concern is whether, once we prepare to go live, we should ensure that ALL historical content is published at once, and if so, whether back-dates should be applied to each content piece (even if dating isn't relevant), or whether we should have a strategy in place in terms of creating a publishing schedule and releasing content over time - albeit content that is older but isn't necessarily time-sensitive (e.g. a drink recipe). Going forward, all newly-created content will be published around the print issue release. Are there pitfalls I should avoid in terms of pushing out so much back content at once?
Intermediate & Advanced SEO | | andrewkissel0 -
Problems in indexing a website built with Magento
Hi all My name is Riccardo and i work for a web marketing agency. Recently we're having some problem in indexing this website www.farmaermann.it which is based on Magento. In particular considering google web master tools the website sitemap is ok (without any error) and correctly uploaded. However only 72 of 1.772 URL have been indexed; we sent the sitemap on google webmaster tools 8 days ago. We checked the structure of the robots.txt consulting several Magento guides and it looks well structured also.
Intermediate & Advanced SEO | | advmedialab
In addition to this we noticed that some pages in google researches have different titles and they do not match the page title defined in Magento backend. To conclude we can not understand if this indexing problems are related to the website sitemap, robots.txt or something else.
Has anybody had the same kind of problems? Thank you all for your time and consideration Riccardo0 -
Next Gen Gallery Crawler Problem
I use the Next gen gallery plugin on my wordpress sites. The moz crawler reports a ton of high importance issues with this plugin because it creates duplicate pages It will have domain.com/page, domain.com/page/gallery, domain.com/page/gallery/1/, domain.com/page/gallery/2/ This is a pretty popular plugin so I am hoping there is some way of fixing this relatively easy. I can imagine i need to set up a rel canonical but there does not seem to be an easy way to do so. Thoughts?
Intermediate & Advanced SEO | | Atomicx0 -
Does anyone have a clue about my search problem?
After three years of destruction, my site still has a problem - or maybe more than one. OK, I understand I had - and probably still have - a Panda problem. The question is - does anyone know how to fix it, without destroying eveything? If I had money, I'd gladly give it up to fix this, but all I have is me, a small dedicated promotions team, 120,000+ visitors per month and the ability to write, edit and proofread. This is not an easy problem to fix. After completing more than 100 projects, I still haven't got it right, in fact, what I've done over the past 2 months has only made things worse - and I never thought I could do that. Everything has been measured, so as not to destroy our remaining ability to generate income, because without that, its the end of the line. If you can help me fix this, I will do anything for you in return - as long as it is legal, ethical and won't destroy my reputation or hurt others. Unless you are a master jedi guru, and I hope you are, this will NOT be easy, but it will prove that you really are a master, jedi, guru and time lord, and I will tell the world and generate leads for you. I've been doing website and SEO stuff since 1996 and I've always been able to solve problems and fix anything I needed to work on. This has me beaten. So my question is: is there anyone here willing to take a shot at helping me fix this, without the usual response of "change domains" "Delete everything and start over" or "you're screwed" Of course, it is possible that there is a different problem, nothing to do with algorithms, a hard-coded bias or some penalizing setting, that I don't know about, a single needle in a haystack. This problem results in a few visible things. 1. Some pages are buried in supplemental results 2. Search bots pick up new stories within minutes, but they show up in search results many hours later Here is the site: http://shar.es/EGaAC On request, I can provide a list of all the things we've done or tried. (actually I have to finish writing it) Some Notes: There is no manual spam penalty. All outgoing links are nofollow, and have been for 2 years. We never paid for incoming links. We did sell text advertising links 3-4 years ago, using text-link-ads.com, but removed them all 2 1/2 years ago. We did receive payment for some stories, 3-4 years ago, but all have been removed. One more thing. I don't write much - I'm a better editor than a writer, but I wrote a story that had 1 million readers. the massive percentage of 0.0016% came from you-know-who. Yes, 16 visitors. And this was an exclusive, unique story. And there was a similar story, with half a million readers. same result. Seems like there might be a problem!
Intermediate & Advanced SEO | | loopyal0 -
Microdata / Schema.org and HTTPS
I have a quick question regarding Microdata / Schema.org files that are not hosted on secure connections. I receive a receive a security error from my e-commerce site because the code references the schema over HTTP instead of HTTPS.<div< span="">itemscope itemtype="http://schema.org/Product"></div<>This is not the first time I have run into this issue. We also use MRSS schema for an RSS feed from yahoo and the same thing happens.<div< span="">xmlns:media="http://search.yahoo.com/mrss"></div<>The problem mainly lies in the fact that these schemas are not hosted over HTTPS. If you add HTTPS to the beginning of both you will get a security error.Just wondering if anyone else has dealt with this or similar issue and what the "best practices" are around this?Is it ok to obtain the schema directly and then host it on our server, over our secure connection?Thanks!
Intermediate & Advanced SEO | | AnthonyMangia0 -
Microdata and dinamic data.
Hi, everybody! We're starting up a local services website in Brazil. Something like redbeacon.com or thumbtack.com, but obviously different. So we are developing our 2.0 version of the site, and I want do put microdata in every provider's pages, to rank people's evaluation about this particular provider, and geographic information about him. Ok, we want to use microdata in several pages, but those are more important: the providers. These data (geo and rank) will be dynamically generated from our database. In Schema.org, I only found information about using static data to build microdata for my intentions. My doubt is: does google and bing and yahoo and etc index dynamic generated data? Is there something about sitemaps.xml or robots.txt that I can do to have my data indexed on search engines? Our front-end is the guy who deal with html and our codemaster uses pure php for coding. Thanks!
Intermediate & Advanced SEO | | ivan.precisodisso0 -
Think I may have found a problem with site. Can you confirm my suspicions?
So I've been wracking my brain about a problem. I had posted earlier about our degrading rank that we haven't been able to arrest. I thought we were doing everything right. Many years ago we had a program that would allow other stores in our niche use our site as a storefront if they couldn't deal with setting up their own site. They would have their own homepage with their own domain but all links from that page would go to our site to avoid duplicate content issues (before I knew about canonical meta tags or before they existed, I don't remember). I just realize that we had dozens of these domains pointing to our site without nofollow meta tags. Is it possible that this pattern looked like we were trying to game Google and have been penalized as some kind of link farm since Panda? I've added nofollow meta tags to these domains. If we were being penalized for this, should this fix the problem?
Intermediate & Advanced SEO | | IanTheScot0