Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Event Schema markup for multiple events (same location/address)?
-
I was wondering if its possible to markup multiple events on the same page for one location/address using the event schema.org markup? I tried doing it on a sample page below:
http://www.rama.id.au/event-schema-test/
Google's schema testing tool shows that its all good (except for warning for offers). Just wanted to know if I am doing it correctly or is there a better solution. Any help would be much appreciated.
Thank you
-
Webmaster tools / search console https://www.google.com/webmasters/tools/home?hl=en
https://search.google.com/structured-data/testing-tool/u/0/
You can use this as a template
Or use this great tool https://jsonld.com/json-ld-generator/
(remember with JSON-LD you must have the content in HTML if you post via JSON-LD)
-
-
Hi Vincent
Again, it may be a matter of creating individual pages for each event w proper Schema, or it is possible to use tags in Schema as well. So, each event would be wrapped by their proper Event tag, with the address information in a meta tag included. Both are rather tedious. You can read more here:
https://schema.org/docs/gs.html#schemaorg_testing
Ideally though, you'd have an individual page for each event.
You could follow Ticketmaster's path and use data-vocabulary.com markup, however, Schema is the standard. If you're wondering what I mean, run the following URL through the Google Structure Data Markup Tester I linked to in my previous comment:
http://www.ticketmaster.com/Chicago-Bulls-tickets/artist/805914
Sorry for not posting links - I am on my phone and I cannot. Will update in the AM.
Hope these help! Good luck!
-
Hello Oleg and Patrick
Thank you so much gentlemen for helping me out. Unfortunately, I cannot wrap each event in it's own using itemscope itemtype="http://schema.org/Event" as each event will then require me to specify the same address multiple times due to the "location" attribute being a required field. Since the address occurs only once on the pageI am bound to use it only once by tying the same address to multiple address. On a side note, how come the Google schema testing tool able to pass my implementation on the sample URL?
Hope to hear from you soon.
Thanks once again.
-
I agree with Oleg here - each event should have it's own page.
That being said, it is possible to markup individual events on the same page, because each event is has it's own unique attributes. Each event will be wrapped in it's own itemscope itemtype="http://schema.org/Event" - so be mindful of that.
You can read more here.
Keep in mind Google and Yandex have structured data markup testing tools.
Hope this helps! Good luck!
-
You should create separate events for each event you have (even if the location is the same)
from https://schema.org/Event --> "Repeated events may be structured as separate Event objects."
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Multiple robots.txt files on server
Hi! I have previously hired a developer to put up my site and noticed afterwards that he did not know much about SEO. This lead me to starting to learn myself and applying some changes step by step. One of the things I am currently doing is inserting sitemap reference in robots.txt file (which was not there before). But just now when I wanted to upload the file via FTP to my server I found multiple ones - in different sizes - and I dont know what to do with them? Can I remove them? I have downloaded and opened them and they seem to be 2 textfiles and 2 dupplicates. Names: robots.txt (original dupplicate)
Technical SEO | | mjukhud
robots.txt-Original (original)
robots.txt-NEW (other content)
robots.txt-Working (other content dupplicate) Would really appreciate help and expertise suggestions. Thanks!0 -
Title Tag vs. H1 / H2
OK, Title tag, no problem, it's the SEO juice, appears on SERP, etc. Got it. But I'm reading up on H1 and getting conflicting bits of information ... Only use H1 once? H1 is crucial for SERP Use H1s for subheads Google almost never looks past H2 for relevance So say I've got a blog post with three sections ... do I use H1 three times (or does Google think you're playing them ...) Or do I create a "big" H1 subhead and then use H2s? Or just use all H2s because H1s are scary? 🙂 I frequently use subheads, it would seem weird to me to have one a font size bigger than another, but of course I can adjust that in settings ... Thoughts? Lisa
Technical SEO | | ChristianRubio0 -
Schema markup for products is missing "price": Is this bad?
Hey guys, So a current client of mine has an e-commerce shop with a few hundred products. They purposely choose to keep the prices off of their website, which is causing errors in Google Webmaster Tools. Basically the error shows: Error: Structured Data > Product (markup: schema.org) Error type: missing price 208 items with error Is this a huge deal? Or are we allowed to have non-numerical prices for schema ie. "call for quote"
Technical SEO | | tbinga1 -
Can Google Read schema.org markup within Ajax?
Hi All, as a local business directory, we also display Openinghours on a business listing page. ex. http://www.goudengids.be/napoli-kontich-2550/
Technical SEO | | TruvoDirectories
At the same time I also have schema.org markup for Openinghours implemented.
But, for technical reasons (performance), the openinghours (and the markup alongside) are displayed using AJAX. I'm wondering if google is able to read the markup. The rich snippet tool and markup plugings like Semantic Inspector can't "see" the markup for openinghours. Any advice here?0 -
Is it a problem to have a homepage with a slug / URL ?
Hi, We are designing a web site for one of our clients, and using a home made CMS. I don't know how this CMS has been built, but anyways, in the end the homepage has a URL format which looks like this : www.mydomain.com/my-custom-url.html. No www.mydomain.com. Is it dangerous for SEO to have a slug/URL directly on the homepage ? Do you have experiences, cases where it has impacted a site negatively ? The main problem I expect is duplicate content (with Google seeing both www.mydomain.com and www.mydomain.com/my-custom-url.html as being different pages) but apparently the CMS is doing a 302 redirect from the root domain to the URL (I told my colleague it should at least be a 301). Sorry if this question seems like basic SEO knowledge, but I really can't find a definitive answer on the subject. Thank you very much 🙂
Technical SEO | | edantadis0 -
Cloudflare Email Address Protecting & SEO
I have a client who uses Cloudflare to protect their email address on their website from spam and when I crawl their site looking for errors it causes it to report multiple broken links on each page. Does anyone know how this impacts SEO? Should I recommend removing Cloudflare?
Technical SEO | | farlandlee0 -
OK to block /js/ folder using robots.txt?
I know Matt Cutts suggestions we allow bots to crawl css and javascript folders (http://www.youtube.com/watch?v=PNEipHjsEPU) But what if you have lots and lots of JS and you dont want to waste precious crawl resources? Also, as we update and improve the javascript on our site, we iterate the version number ?v=1.1... 1.2... 1.3... etc. And the legacy versions show up in Google Webmaster Tools as 404s. For example: http://www.discoverafrica.com/js/global_functions.js?v=1.1
Technical SEO | | AndreVanKets
http://www.discoverafrica.com/js/jquery.cookie.js?v=1.1
http://www.discoverafrica.com/js/global.js?v=1.2
http://www.discoverafrica.com/js/jquery.validate.min.js?v=1.1
http://www.discoverafrica.com/js/json2.js?v=1.1 Wouldn't it just be easier to prevent Googlebot from crawling the js folder altogether? Isn't that what robots.txt was made for? Just to be clear - we are NOT doing any sneaky redirects or other dodgy javascript hacks. We're just trying to power our content and UX elegantly with javascript. What do you guys say: Obey Matt? Or run the javascript gauntlet?0 -
Nofollow and ecommerce cart/checkout pages
Hi!! Another noob question: Should I be nofollowing my site's cart and checkout pages? Or as SEs can't get to the checkout pages without either logging in or completing the form is it something I shouldn't worry about? Have read things saying both. Not sure which is correct. Thank you! Appreciate the help. Lynn
Technical SEO | | hiphound0