Content Below the Fold
-
Hi
I wondered what the view is on content below the fold?
We have the H1, product listings & then some written content under the products - will Google just ignore this?
I can't hide it under a tab or put a lot of content above products - so I'm not sure what the other option is?
Thank you
-
Hi Becky,
Here is what I found:
The pros and cons of hiding content using JavaScript and CSS (display: none) has been a topic of some debate within the SEO industry, and Google’s comments over time have somewhat added to the confusion.
- **November 2014 **– Google’s John Mueller stated that Google _“may not” _index or rank hidden content. In aGoogle+ Hangout the following month, John repeated this, stating that hidden content would be _“discounted”_and has been for a number of years
- **21 July 2015 **– Google’s Gary Illyes, contributing to a Stack Overflow forum thread , provided clarification of this by stating that this type of content is given “way less weight in ranking”
- **27 July 2015 **– In a separate Stack Overflow thread on the same topic, Gary Illyes again confirmed that _“[Google] will index that but the content’s weight will be lower since it’s hidden” _
So the content will still be indexed, but deemed less important by the crawlers.
-
Yeh it's disappointing.
I've tried having some content behind a tab and some under the products and I am not seeing either one as having much of an effect.
Unless I remove it altogether, I'm not sure what else I can do with it?
-
Hi
Yes I tried different pages and it's still the same. I think it's to do with things we have blocked in robots.txt...
-
I'm not seeing a problem in my GoogleBot simulators, Becky, but the one within your Google Search Console is still the best judge. Have you tried reloading the Fetch as... a couple of times? And tried it on different pages?
-
Yup - Google still says content that can only be seen after a user interaction is given less importance. Kinda stupid, given that things like tabs/accordians are a major usability enhancement, but that's still where we are.
P.
-
Hi
So I did fetch as Google - and I'm seeing the page quite differently if I'm Googlebot vs. visitors.
It just sees a few big images, I can't see it rendering any product listings or content - do I now have a bigger problem?
Thank you
-
Hi
Thank you for the replies. I don't want to hide it, I just can't have it pushing products down the page so they can't be seen..
I thought in Google webmaster guidelines they included a comment to say they will ignore content behind tabs?
Becky
-
Any content below the fold will still be read. Are you trying to hide it but still get the SEO value? If that's the case, I would create a collapsible tab to keep the content on the page but hidden. If you want it to be visible, leave it as is and don't worry about Google not reading it—it will be read.
-
While theoretically logical, Google's own John Mueller stated last week that code to text ratio has absolutely no effect on crawling of a site, and in a followup question, he directly told me text/code ratio has no effect as a ranking factor either.
These used to be very minor considerations back when the search engine crawlers weren't as powerful, but no longer.
Fully agree with Pia that the idea of "above the fold" influencing ranking is nonsense as well. Given that the sweet spot for consistently high-ranking pages is ~2200 words, the idea that only the first paragraph or two are more important is unsupportable.
Hope that helps?
Paul
-
Additionally, do check the content using Fetch as google in Google Search Console / Webmaster tools. It would really help you see how spiders see the content compared to users. This is an important aspect of SEO which a lot of people ignore, you are looking to find that whether the spiders see a structured view of the content and not messy. I hope this helps, if you have further questions, please feel free to ask. Regards, Vijay
-
There's no manipulation whatsoever. In fact, Google encourage website developers and SEOs to optimise/tidy their code and keep a good code-to-content ratio. This is why Google gives us so many tools in order to do so. It makes our sites easier to crawl for Google, and in return Google may even like us more for it!
Just found an article that sums it up quite nicely:
"Essentially what is being stated is a fairly logical conclusion: reduce the amount of code on your page and the content (you know, the place where your keywords are) takes a higher priority. Additionally compliance will, by necessity, make your site easy to crawl and, additionally, allow you greater control over which portions of your content are given more weight by the search engines. The thinking behind validating code for SEO benefits is that, once you have a compliant site, not only will your site be a better user experience on a much larger number of browsers, but you’ll have a site with far less code, that will rank higher on the search engines because of it."
- http://www.hobo-web.co.uk/official-google-prefers-valid-html-css/
But going back to your original post, "above-the-fold is dead", yadda yadda... So long as your content in the source is metaphorically "above the fold" and not drowning in heavy code, on the page itself just worry about giving your users the "experience" that they're looking for. And not how many pixels from the top of the browser your content is. Hope that makes more sense!
-
Great thank you, you read so many conflicting articles that it's difficult to know.
I'll see if we can look at our code, but I'd want to be mindful of not manipulating Google.
Thank you!
-
I feel prioritising elements to be "above the fold" is a bit of an outdated concept these days.
Where is the fold? Different devices and screen resolutions will have different folds, and more websites are being designed now to make the traditional "above the fold" section more visually interesting and designed for user experience, rather than packed full of content.
The higher the content is in the source code itself, the more weight it will have on the page. This doesn't necessarily translate to the "visually higher the content is on the page". Google is going to be reading from top to bottom of your code, so naturally you want the most important content/links to be found first. As long as you meet (or exceed!) the user's expectation of the content upon arrival, and you keep the code tidy in terms of how much Google has to read before it gets to the real valuable content, I doubt Google's going to worry about whether users have to scroll a little to get to it.
-
Hi Becky,
As far as i understand Google will not ignore however Google do treat some part of the page as more important than other. For instance, if you have written a description of the product and some of the description is been hide.
Google, will take that as the important piece of content been displayed for user and least important been hide.
I do not see any point for Google to ignore the fold one. -
Content below the fold is still read, however less value is placed on it. So it is still worth having content that is produced for below the fold as it will still help that page rank.
Show the user what they want to see when they land on the page, majority of the time in doing this you will actually show Google what they need to rank you.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Ranking all but lost after new content pushed
Okay, I thought I was following best practices. In our industry, electronic hardware, we were ranking well for a particular product line (/spacers) but we wanted to do better. We addressed several concerns that Moz found first; duplicate page titles, lack of meta-descriptions and overall lack of targeted keywords. We also took a new approach to add a better structure to our site. Instead of being presented with a list of part numbers we wanted the user to learn more about our products with content. So we added a /products page with content and a product specific page (/spacers) that is almost a definitive buyers guide. We are attempting to answer the questions that we think our customers find most relevant. Well our customers might find it relevant but Google sure didn't. After our deployment of new content our rankings for targeted keywords in Google fell from 10-15 to 80-95 As an open ended question, could somebody explain to me why our ranks fell off a cliff? Homepage: https://www.lyntron.com
Intermediate & Advanced SEO | | jandk4014
New catalog summary page: https://www.lyntron.com/products
New content with focus to rank high: https://www.lyntron.com/spacers TPdn6ym1 -
Best practice with duplicate content. Cd
Our website has recently been updated, now it seems that all of our products pages look like this cdnorigin.companyname.com/catagory/product Google is showing these pages within the search. rather then companyname.com/catagory/product Each product page does have a canaonacal tag on that points to the cdnorigin page. Is this best practice? i dont think that cdnorigin.companyname etc looks very goon in the search. is there any reason why my designer would set the canonical tags up this way?
Intermediate & Advanced SEO | | Alexogilvie0 -
All Thin Content removed and duplicate content replaced. But still no success?
Good morning, Over the last three months i have gone about replacing and removing all the duplicate content (1000+ page) from our site top4office.co.uk. Now it been just under 2 months since we made all the changes and we still are not showing any improvements in the SERPS. Can anyone tell me why we aren't making any progress or spot something we are not doing correctly? Another problem is that although we have removed 3000+ pages using the removal tool searching site:top4office.co.uk still shows 2800 pages indexed (before there was 3500). Look forward to your responses!
Intermediate & Advanced SEO | | apogeecorp0 -
Duplicate content reported on WMT for 301 redirected content
We had to 301 redirect a large number of URL's. Not Google WMT is telling me that we are having tons of duplicate page titles. When I looked into the specific URL's I realized that Google is listing an old URL's and the 301 redirected new URL as the source of the duplicate content. I confirmed the 301 redirect by using a server header tool to check the correct implementation of the 301 redirect from the old to the new URL. Question: Why is Google Webmaster Tool reporting duplicated content for these pages?
Intermediate & Advanced SEO | | SEOAccount320 -
Wordpress and duplicate content
Hi, I have recently installed wordpress and started a blog but now loads of duplicate pages are cropping up for tags and authors and dates etc. How do I do the canonical thing in wordpress? Thanks Ian
Intermediate & Advanced SEO | | jwdl0 -
Penalised for duplicate content, time to fix?
Ok, I accept this one is my fault but wondering on time scales to fix... I have a website and I put an affiliate store on it, using merchant datafeeds in a bid to get revenue from the site. This was all good, however, I forgot to put noindex on the datafeed/duplicate content pages and over a period of a couple of weeks the traffic to the site died. I have since nofollowed or removed the products but some 3 months later my site still will not rank for the keywords it was ranking for previously. It will not even rank if I type in the sites' name (bright tights). I have searched for the name using bright tights, "bright tights" and brighttights but none of them return the site anywhere. I am guessing that I have been hit with a drop x place penalty by Google for the duplicate content. What is the easiest way around this? I have no warning about bad links or the such. Is it worth battling on trying to get the domain back or should I write off the domain, buy a new one and start again but minus the duplicate content? The goal of having the duplicate content store on the site was to be able to rank the category pages in the store which had unique content on so there were no problems with that which I could foresee. Like Amazon et al, the categories would have lists of products (amongst other content) and you would click through to the individual product description - the duplicate page. Thanks for reading
Intermediate & Advanced SEO | | Grumpy_Carl0 -
Lots of optimized content but crappy rankings
Hi, I write content for the site http://www.healthchoices.ca. We were hit by Panda (a different issue that is resolved now) but even before that, I'd write an entire library of good content. An example: the plant sterols library. Here's an article as an example: http://www.healthchoices.ca/article/where-are-plant-sterols-phytosterols-found There are about eight on average in each medical library, and we cover topics from acne to sexual health. The other half of the business is a directory. We have thousands of local health are providers, a basic version with just an address, and a longer version where we optimize the text. Here's an example: http://www.healthchoices.ca/profile/ct-mri-scans/canada-diagnostic-centres We come up buried on the third page, far after directories with zero content or crappy content. What am I missing? I am getting very frustrated as I've been writing this stuff for a long time and nothing seems to come of it. Thanks so much, Erin
Intermediate & Advanced SEO | | erinhealthchoices0 -
Duplicate Content across 4 domains
I am working on a new project where the client has 5 domains each with identical website content. There is no rel=canonical. There is a great variation in the number of pages in the index for each of the domains (from 1 to 1250). OSE shows a range of linking domains from 1 to 120 for each domain. I will be strongly recommending to the client to focus on one website and 301 everything from the other domains. I would recommend focusing on the domain that has the most pages indexed and the most referring domains but I've noticed the client has started using one of the other domains in their offline promotional activity and it is now their preferred domain. What are your thoughts on this situation? Would it be better to 301 to the client's preferred domain (and lose a level of ranking power throught the 301 reduction factor + wait for other pages to get indexed) or stick with the highest ranking/most linked domain even though it doesn't match the client's preferred domain used for email addresses etc. Or would it better to use cross-domain canoncial tags? Thanks
Intermediate & Advanced SEO | | bjalc20110