Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Using a non-visible H1
-
I have a developer that wants to use style="text-indent:-9999px" to make the H1 non-visible to the user. Being the conservative person I am, I've never tried this before and worry that Search Engines may think this is a form of cloaking. Am I worrying about nothing? And apologies if it's already been covered here. I couldn't find it. Thanks in advance!!!!
-
From Whiteboard Friday - The Biggest SEO Mistakes SEOmoz Has Ever Made
http://www.seomoz.org/blog/whiteboard-friday-the-biggest-seo-mistakes-seomoz-has-ever-made
"3. Recommending People Use H1 Tags with Keywords
This mistake is a little bit more subtle. For years, SEOmoz recommended including keywords in the H1 of pages. After we started doing formal machine learning correlation tests we found out that this tactic didn't actually help very much at all (including the keywords in normal text in bigger fonts worked essentially the same). This was a shame because it meant we wasted time and energy convincing our clients to update their H1s."
-
Using that CSS wouldn't Hide it from the spiders view, it will simply "move" the H1 off the screen.
It is a pretty old "trick".
Lets not forget Heading tags are useful to site visitors to so shouldn't necessarily be hidden to them.
Users will use the headings whilst they Scan read your pages, if they can't quickly identify what the page content is about there is a danger they could simply bounce off... and you will lose them.
As for Search engines penalising you for it, I'm not too sure, is there any research which anybody can point us towards? I dont think they are reading CSS attributes just yet right?
Andy
-
You came to the right place for the validity you seek
I frequently vet things here in the forum and it has proven very helpful in convincing other members of my team to go one way or the other. Also, I completely agree with George's suggestion to use the "alt" attribute if it is indeed an image we are talking about, but it appears we are really talking about a bonafide
tag for text with keywords in it.
That being the case. Stick to your guns and insist on it being visible. If you really feel that it disrupts the design...it would be better to leave it out than to make it invisible.
Good luck!
Dana
-
Thanks All! So here's more detail. The home page design was completed. I still think H1 has some reasonable value and it didn't have one so I told him to put a keyword rich H1 in. He felt it disrupted the existing design and executed it as above. So....I thought I would seek "convergent validity" on the subject as a next step.
-
I concur with Dana,
Hiding your H1 tag will not necessarily cause a penalty. However, if you do so you are at risk for a penalty. If a particular savvy competitor comes along and notices you are obfuscating your H1 tags and reports it, then you may get dinged. I doubt that alone would cause a problem, but if that sort of tactic is par for the course for this web developer you may be in trouble.
-
Read up on this Webmaster Tools guideline: http://support.google.com/webmasters/bin/answer.py?hl=en&answer=66353
Note the following from the guideline if there is a very important reason for why your developer wants to use CSS to move the text off the page:
"However, not all hidden text is considered deceptive. For example, if your site includes technologies that search engines have difficulty accessing, like JavaScript, images, or Flash files, using descriptive text for these items can improve the accessibility of your site."
If there is not a very important reason, and even if there is, suggest they populate the ALT attribute of the image with the text instead.
Hope this helps!
-
The general SEO community consensus is that you should:
A. be doing what is best for the user (so not concealing the H1 tag)
B. not doing anything that could make Matt Cutts and the Google team upset
They have advised against attempting to conceal content for SEO gain so I would strongly recommend avoiding it. They have been dealing with these issues now for a LONG time, so presumably their bots can easily pick up on those tricks.
The Google bots can now "see" what is visible on the page. They discount things that are not in the visible content area so the benefit to an offset H1 would likely be none. Also: They're watching you.
-
Personally, I wouldn't do it. Does it work? Maybe. Or, maybe it works for a while and then Googlebot wises up and deindexes you. Is all the work you will have to go through for reconsideration going to justify hiding that tag? I'd say, definitely not.
It's just an
tag...leave it on the page and visible. Listen to your conservative gut and do what you know is the right thing. That's my two cents
-
I have personally created an H1 tag in an image, I didn't see no negative effects. H1 tags are not as important but should be implemented, so even if it had any impact maybe it was minuscule.
H1 tags don't generally have to be visible like in my case, it was an H1 tag for the logo. I'm not sure where you are putting the H1 tag but if its an image I say why not, but if it is a regular text, why not just keep it as an H1 without hiding?
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sitemap use for very large forum-based community site
I work on a very large site with two main types of content, static landing pages for products, and a forum & blogs (user created) under each product. Site has maybe 500k - 1 million pages. We do not have a sitemap at this time.
Technical SEO | | CommManager
Currently our SEO discoverability in general is good, Google is indexing new forum threads within 1-5 days roughly. Some of the "static" landing pages for our smaller, less visited products however do not have great SEO.
Question is, could our SEO be improved by creating a sitemap, and if so, how could it be implemented? I see a few ways to go about it: Sitemap includes "static" product category landing pages only - i.e., the product home pages, the forum landing pages, and blog list pages. This would probably end up being 100-200 URLs. Sitemap contains the above but is also dynamically updated with new threads & blog posts. Option 2 seems like it would mean the sitemap is unmanageably long (hundreds of thousands of forum URLs). Would a crawler even parse something that size? Or with Option 1, could it cause our organically ranked pages to change ranking due to Google re-prioritizing the pages within the sitemap?
Not a lot of information out there on this topic, appreciate any input. Thanks in advance.0 -
302 redirect used, submit old sitemap?
The website of a partner of mine was recently migrated to a new platform. Even though the content on the pages mostly stayed the same, both the HTML source (divs, meta data, headers, etc.) and URLs (removed index.php, removed capitalization, etc) changed heavily. Unfortunately, the URLs of ALL forum posts (150K+) were redirected using a 302 redirect, which was only recently discovered and swiftly changed to a 301 after the discovery. Several other important content pages (150+) weren't redirected at all at first, but most now have a 301 redirect as well. The 302 redirects and 404 content pages had been live for over 2 weeks at that point, and judging by the consistent day/day drop in organic traffic, I'm guessing Google didn't like the way this migration went. My best guess would be that Google is currently treating all these content pages as 'new' (after all, the source code changed 50%+, most of the meta data changed, the URL changed, and a 302 redirect was used). On top of that, the large number of 404's they've encountered (40K+) probably also fueled their belief of a now non-worthy-of-traffic website. Given that some of these pages had been online for almost a decade, I would love Google to see that these pages are actually new versions of the old page, and therefore pass on any link juice & authority. I had the idea of submitting a sitemap containing the most important URLs of the old website (as harvested from the Top Visited Pages from Google Analytics, because no old sitemap was ever generated...), thereby re-pointing Google to all these old pages, but presenting them with a nice 301 redirect this time instead, hopefully causing them to regain their rankings. To your best knowledge, would that help the problems I've outlined above? Could it hurt? Any other tips are welcome as well.
Technical SEO | | Theo-NL0 -
Does image domain name matter when using a CDN?
Has anyone does studies on using a different CDN domain name for images on a site? Here is an example:
Technical SEO | | findwellor http://cdn.mydomain.com/image.jpg> mydomain.com ranks highly and many images show up in Google/Bing image searches. Is there any actual data that says that using your real domain name for the CDN has benefits versus the default domain name provided by the CDN provider? On the surface, it feels like it would, but I haven't experimented with it.
0 -
Rel=Canonical, WWW vs non WWW and SEO
Okay so I'm a bit of a loss here. For what ever reason just about every single Wordpress site I has will turn www.mysite.com into mysite.com in the browser bar. I assume this is the rel=canonical tag at work, there are no 301s on my site. When I use the Open Site Explorer and type in www.mysite.com it shows a domain authority of around 40 and a few hundred backlinks... and then I get the message. Oh Hey! It looks like that URL redirects to XXXXXX. Would you like to see data for <a class="clickable redirects">that URL instead</a>? So if I click to see this data instead I have less than half of that domain authority and about 2 backlinks. *** Does this make a difference SEO wise? Should my non WWW be redirecting to my WWW instead because that's where the domain authority and backlinks are? Why am I getting two different domain authority and backlink counts if they are essentially the same? Or am I wrong and all that link juice and authority passes just the same?
Technical SEO | | twilightofidols0 -
Www vs non-www which is better?
Is it better to have all your pages point to the www version or non www version.
Technical SEO | | bronxpad0 -
404 errors on non-existent URLs
Hey guys and gals, First Moz Q&A for me and really looking forward to being part of the community. I hope as my first question this isn't a stupid one but I was just struggling to find any resource that dealt with the issue and am just looking for some general advice. Basically a client has raised a problem with 404 error pages - or the lack thereof- on non-existent URLs on their site; let's say for example: 'greatbeachtowels.com/beach-towels/asdfas' Obviously content never existed on this page so its not like you're saying 'hey, sorry this isn't here anymore'; its more like- 'there was never anything here in the first place'. Currently in this fictitious example typing in 'greatbeachtowels.com/beach-towels/asdfas**'** returns the same content as the 'greatbeachtowels.com/beach-towels' page which I appreciate isn't ideal. What I was wondering is how far do you take this issue- I've seen examples here on the seomoz site where you can edit the URI in a similar manner and it returns the same content as the parent page but with the alternate address. Should 404's be added across all folders on a site in a similar way? How often would this scenario be and issue particularly for internal pages two or three clicks down? I suppose unless someone linked to a page with a misspelled URL... Also would it be worth placing 301 redirects on a small number of common mis-spellings or typos e.g. 'greatbeachtowels.com/beach-towles' to the correct URLs as opposed to just 404s? Many thanks in advance.
Technical SEO | | AJ2340 -
How to handle sitemap with pages using query strings?
Hi, I'm working to optimize a site that currently has about 5K pages listed in the sitemap. There are not in face this many pages. Part of the problem is that one of the pages is a tool where each sort and filter button produces a query string URL. It seems to me inefficient to have so many items listed that are all really the same page. Not to mention wanting to avoid any duplicate content or low quality issues. How have you found it best to handle this? Should I just noindex each of the links? Canonical links? Should I manually remove the pages from the sitemap? Should I continue as is? Thanks a ton for any input you have!
Technical SEO | | 5225Marketing0 -
What tool do you use to check for URLs not indexed?
What is your favorite tool for getting a report of URLs that are not cached/indexed in Google & Bing for an entire site? Basically I want a list of URLs not cached in Google and a seperate list for Bing. Thanks, Mark
Technical SEO | | elephantseo3