Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Using a non-visible H1
-
I have a developer that wants to use style="text-indent:-9999px" to make the H1 non-visible to the user. Being the conservative person I am, I've never tried this before and worry that Search Engines may think this is a form of cloaking. Am I worrying about nothing? And apologies if it's already been covered here. I couldn't find it. Thanks in advance!!!!
-
From Whiteboard Friday - The Biggest SEO Mistakes SEOmoz Has Ever Made
http://www.seomoz.org/blog/whiteboard-friday-the-biggest-seo-mistakes-seomoz-has-ever-made
"3. Recommending People Use H1 Tags with Keywords
This mistake is a little bit more subtle. For years, SEOmoz recommended including keywords in the H1 of pages. After we started doing formal machine learning correlation tests we found out that this tactic didn't actually help very much at all (including the keywords in normal text in bigger fonts worked essentially the same). This was a shame because it meant we wasted time and energy convincing our clients to update their H1s."
-
Using that CSS wouldn't Hide it from the spiders view, it will simply "move" the H1 off the screen.
It is a pretty old "trick".
Lets not forget Heading tags are useful to site visitors to so shouldn't necessarily be hidden to them.
Users will use the headings whilst they Scan read your pages, if they can't quickly identify what the page content is about there is a danger they could simply bounce off... and you will lose them.
As for Search engines penalising you for it, I'm not too sure, is there any research which anybody can point us towards? I dont think they are reading CSS attributes just yet right?
Andy
-
You came to the right place for the validity you seek
I frequently vet things here in the forum and it has proven very helpful in convincing other members of my team to go one way or the other. Also, I completely agree with George's suggestion to use the "alt" attribute if it is indeed an image we are talking about, but it appears we are really talking about a bonafidetag for text with keywords in it.
That being the case. Stick to your guns and insist on it being visible. If you really feel that it disrupts the design...it would be better to leave it out than to make it invisible.
Good luck!
Dana
-
Thanks All! So here's more detail. The home page design was completed. I still think H1 has some reasonable value and it didn't have one so I told him to put a keyword rich H1 in. He felt it disrupted the existing design and executed it as above. So....I thought I would seek "convergent validity" on the subject as a next step.
-
I concur with Dana,
Hiding your H1 tag will not necessarily cause a penalty. However, if you do so you are at risk for a penalty. If a particular savvy competitor comes along and notices you are obfuscating your H1 tags and reports it, then you may get dinged. I doubt that alone would cause a problem, but if that sort of tactic is par for the course for this web developer you may be in trouble.
-
Read up on this Webmaster Tools guideline: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=66353
Note the following from the guideline if there is a very important reason for why your developer wants to use CSS to move the text off the page:
"However, not all hidden text is considered deceptive. For example, if your site includes technologies that search engines have difficulty accessing, like JavaScript, images, or Flash files, using descriptive text for these items can improve the accessibility of your site."
If there is not a very important reason, and even if there is, suggest they populate the ALT attribute of the image with the text instead.
Hope this helps!
-
The general SEO community consensus is that you should:
A. be doing what is best for the user (so not concealing the H1 tag)
B. not doing anything that could make Matt Cutts and the Google team upset
They have advised against attempting to conceal content for SEO gain so I would strongly recommend avoiding it. They have been dealing with these issues now for a LONG time, so presumably their bots can easily pick up on those tricks.
The Google bots can now "see" what is visible on the page. They discount things that are not in the visible content area so the benefit to an offset H1 would likely be none. Also: They're watching you.
-
Personally, I wouldn't do it. Does it work? Maybe. Or, maybe it works for a while and then Googlebot wises up and deindexes you. Is all the work you will have to go through for reconsideration going to justify hiding that tag? I'd say, definitely not.
It's just an
tag...leave it on the page and visible. Listen to your conservative gut and do what you know is the right thing. That's my two cents

-
I have personally created an H1 tag in an image, I didn't see no negative effects. H1 tags are not as important but should be implemented, so even if it had any impact maybe it was minuscule.
H1 tags don't generally have to be visible like in my case, it was an H1 tag for the logo. I'm not sure where you are putting the H1 tag but if its an image I say why not, but if it is a regular text, why not just keep it as an H1 without hiding?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sitemap use for very large forum-based community site
I work on a very large site with two main types of content, static landing pages for products, and a forum & blogs (user created) under each product. Site has maybe 500k - 1 million pages. We do not have a sitemap at this time.
Technical SEO | | CommManager
Currently our SEO discoverability in general is good, Google is indexing new forum threads within 1-5 days roughly. Some of the "static" landing pages for our smaller, less visited products however do not have great SEO.
Question is, could our SEO be improved by creating a sitemap, and if so, how could it be implemented? I see a few ways to go about it: Sitemap includes "static" product category landing pages only - i.e., the product home pages, the forum landing pages, and blog list pages. This would probably end up being 100-200 URLs. Sitemap contains the above but is also dynamically updated with new threads & blog posts. Option 2 seems like it would mean the sitemap is unmanageably long (hundreds of thousands of forum URLs). Would a crawler even parse something that size? Or with Option 1, could it cause our organically ranked pages to change ranking due to Google re-prioritizing the pages within the sitemap?
Not a lot of information out there on this topic, appreciate any input. Thanks in advance.0 -
Merge 2 websites into one, using a non-existing, new domain.
I need to merge https://www.WebsiteA.com and https://www.WebsiteB.com to a fresh new domain (with no content) https://www.WebsiteC.com. I want to do it the best way to keep existing SEO juice. Website A is the companies home page and built with Wordpress Website B is the company product page and built with Wordpress Website C will be the new site containing both website A and B, utilizing Wordpress also. What is the best way to do this? I have research a lot and keep hitting walls on how to do it. It's a little trickier because it's two different domains going to a brand new domain. Thanks
Technical SEO | | jarydcat10 -
Canonical Tag when using Ajax and PhantomJS
Hello, We have a site that is built using an AJAX application. We include the meta fragment tag in order to get a rendered page from PhantomJS. The URL that is rendered to google from PhantomJS then is www.oursite.com/?escaped_fragment= In the SERP google of course doesnt include the hashtag in the URL. So my question, with this setup, do i still need a canonical tag and if i do, would the canonical tag be the escaped fragment URL or the regular URL? Much Appreciated!
Technical SEO | | RevanaDigitalSEO0 -
Having www. and non www. links indexed
Hey guys, As the title states, the two versions of the website are indexed in Google. How should I proceed? Please also note that the links on the website are without the www. How should I proceed knowing that the client prefers to have the www. version indexed. Here are the steps that I have in mind right now: I set the preferred domain on GWMT as the one with www. I 301 redirect any non www. URL to the www. version. What are your thoughts? Should I 301 redirect the URL's? or is setting the preference on GWMT enough? Thanks.
Technical SEO | | BruLee0 -
How do I fix the h1 tag?
No More Than One H1 Tag Easy fix <dl> <dt>Number of H1s</dt> <dd>2</dd> <dt>Explanation</dt> <dd>Best practices for both SEO and accessibility require only a single H1 tag. The H1 is meant to be the page's headline, and thus, multiple H1s are confusing. Consider employing H2, H3 or CSS styles to achieve the same results with text visualization.</dd> <dt>Recommendation</dt> <dd>Remove multiple instances of the H1 tag, so that only one exists on the page.</dd> <dd>I get this error yet it does not tell me how to fix it. I'm not even sure what the H1 tag is?
Technical SEO | | 678648631264
</dd> </dl>0 -
Using video transcripts v captions and avoiding duplicate content?
Part 1: After editing a You Tube transcript, I typically re-upload as a caption file (with time codes)...for SEO does it matter whether you upload as a transcript v. captions? Is one better than the other? Part 2: If you upload a transcript (or caption) to YouTube, then post that video/transcript in your blog, wouldn't you get pinged for duplicate content?
Technical SEO | | vernonmack0 -
404 errors on non-existent URLs
Hey guys and gals, First Moz Q&A for me and really looking forward to being part of the community. I hope as my first question this isn't a stupid one but I was just struggling to find any resource that dealt with the issue and am just looking for some general advice. Basically a client has raised a problem with 404 error pages - or the lack thereof- on non-existent URLs on their site; let's say for example: 'greatbeachtowels.com/beach-towels/asdfas' Obviously content never existed on this page so its not like you're saying 'hey, sorry this isn't here anymore'; its more like- 'there was never anything here in the first place'. Currently in this fictitious example typing in 'greatbeachtowels.com/beach-towels/asdfas**'** returns the same content as the 'greatbeachtowels.com/beach-towels' page which I appreciate isn't ideal. What I was wondering is how far do you take this issue- I've seen examples here on the seomoz site where you can edit the URI in a similar manner and it returns the same content as the parent page but with the alternate address. Should 404's be added across all folders on a site in a similar way? How often would this scenario be and issue particularly for internal pages two or three clicks down? I suppose unless someone linked to a page with a misspelled URL... Also would it be worth placing 301 redirects on a small number of common mis-spellings or typos e.g. 'greatbeachtowels.com/beach-towles' to the correct URLs as opposed to just 404s? Many thanks in advance.
Technical SEO | | AJ2340 -
Should we use Google's crawl delay setting?
We’ve been noticing a huge uptick in Google’s spidering lately, and along with it a notable worsening of render times. Yesterday, for example, Google spidered our site at a rate of 30:1 (google spider vs. organic traffic.) So in other words, for every organic page request, Google hits the site 30 times. Our render times have lengthened to an avg. of 2 seconds (and up to 2.5 seconds). Before this renewed interest Google has taken in us we were seeing closer to one second average render times, and often half of that. A year ago, the ratio of Spider to Organic was between 6:1 and 10:1. Is requesting a crawl-delay from Googlebot a viable option? Our goal would be only to reduce Googlebot traffic, and hopefully improve render times and organic traffic. Thanks, Trisha
Technical SEO | | lzhao0