Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Reasons Why Our Website Pages Randomly Loads Without Content
-
I know this is not a marketing question but this community is very dev savvy so I'm hoping someone can help me. At random times we're finding that our website pages load without the main body content. The header, footer and navigation loads just fine. If you refresh, it's fine but that's not a solution.
- Happens on Chrome, IE and Firefox, testing with multiple browser versions
- Happens across various page types - but seems to be only the main content section/container
- Happens while on the company network, as well as externally
- Happens after deleting cookies, temporary internet files and restarting computer
- We are using a CMS that is virtually unheard of - Bridgeline/Iapps
- Codebase is .net
Our IT/Dev group keeps pushing back, blaming it on cookies or Chrome plugins because they apparently are unable to "recreate the problem". This has been going on for months and it's a terrible experience for the user to have. It's also not great when landing PPC visitors on pages that load with no content. If anyone has ideas as to why this may be happening I would really appreciate it.
I'm not sure if links are allowed, by today the issue happened on this page serversdirect.com/dm/geek-biz
Linking to an image example below
-
Thanks Gazzerman - This helps as well. Yes, I agree the site definitely needs SEO attention. It was not until recently that a more experience team was brought on to 'fix the site'.
-
I took one look at the code and was amazed by how much is inline. Most of it should be in js and css files. I noticed the solutions pages many of them don't even have title tags! What's going on there? That needs seo attention and is a big no no for coding. This leads me to believe the code needs very close inspection. It surly cant take that long to strip a whole bunch of that code out of each page or template.
If a page is slow to load or pulls in external js files as well you can have scripts try to execute before the required code has loaded. This is probably what is happening some of the time and would explain the inconsistent nature of it.
Run the site through a bunch of validators. I am sure you will find a bunch of stuff.
-
Oddly enough it only seemed to happen on the /Servers page - for me at least. The first time I loaded the page, I got a broken image and an X-Out/Close box. That may help you hone in on the problem.
-
Travis - Thanks for your response. This gives me a better idea of where to start looking. Honestly this site was very poorly developed in my opinion. It was before my time and much of it outsourced to junior teams overseas, which is evident when looking at the source code. But I am stuck with it for now.
-
Found the site. I'm able to recreate the problem. Check the image attachment.
I ran it through tools.pingdom.com, a chrome extension called cast_sender.js isn't loading. But something makes me think there's an issue with the CSS that only crops up from time to time, like the JS blocks it every once in a while. There's both a Global.css and a Global.js in the head. There are also folder levels called /Script%20Library/ and /Style%20Library/.
There's a ton of JavaScript on the site, which I don't have time to get into. And honestly, I'm more of a WordPress guy. I managed to make the page load fail a few times out of a dozen or so tries. I think something in the JavaScript is causing the issue.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What type of website is best for seo.
I need a new website for my health insurance business. What type is best for SEO? Many thanks
Web Design | | laurentjb1 -
Does changing content and design of the website gonna affect my all the backlinks i have made till now
i have been working on my link profile for a month now, after learning about 5 step moz methodology i have decided that i would like to change all of the content of my site and taylor it to what my customers need, am i gonna loose all the domain authority if make changes? if it gonna affect, hows that gonna come out
Web Design | | calvinkj0 -
Reason for robots.txt file blocking products on category pages?
Hi I have a website with thosands of products. On the category pages, all the products are linked to with the code “?cgid” in the URL. But “?cgid” is also blocked in the robots.txt file for some reason. So I'm thinking it's stopping all my products getting crawled by Google. Am I right here? Is there any reason why a website would want to limit so many URL's? I'm only here a week and the sites getting great traffic, so don't want to go breaking it!!! Thanks
Web Design | | Frankie-BTDublin0 -
Weird Layout on Initial Website Load?
Whenever I open my site from an uncached source, like google incognito, for a split second it displays purple links and a white background while it loads the rest of the content. I've included a screenshot. Is there any way to fix that.? The site is www.kemprugegreen.com. u8P9q
Web Design | | KempRugeLawGroup1 -
How to prevent development website subdomain from being indexed?
Hello awesome MOZ Community! Our development team uses a sub-domain "dev.example.com" for our SEO clients' websites. This allows changes to be made to the dev site (U/X changes, forms testing, etc.) for client approval and testing. An embarrassing discovery was made. Naturally, when you run a "site:example.com" the "dev.example.com" is being indexed. We don't want our clients websites to get penalized or lose killer SERPs because of duplicate content. The solution that is being implemented is to edit the robots.txt file and block the dev site from being indexed by search engines. My questions is, does anyone in the MOZ Community disagree with this solution? Can you recommend another solution? Would you advise against using the sub-domain "dev." for live and ongoing development websites? Thanks!
Web Design | | SproutDigital0 -
Above the Fold Content - Use of large images
Hi All, Our designers have come to the SEO team to ask if have a large image across the top of the page taking up a large majority of the above the fold real estate will impact our SEO. Our initial thoughts are no as long as we have an optimised H1 visibal to the user landing there which informs them what the page is about. Any thoughts would be appreciated.
Web Design | | J_Sinclair1 -
One Page Guide vs. Multiple Individual Pages
Howdy, Mozzers! I am having a battle with my inner-self regarding how to structure a resources section for our website. We're building out several pieces of content that are meant to be educational for our clients and I'm having trouble deciding how to layout the content structure. We could either layout all eight short sections on a single page, or create individual pages for each section. The goal is obviously to attract new potential clients by targeting these terms that they may be searching for in an information gathering stage. Here's my dilemma...
Web Design | | jpretz
With the single page guide, it would be nice because it will have a lot of content (and of course, keywords) to be picked up by the SERPS but I worry that it is going to be a bit crammed (because of eight sections) for the user. The individual pages would be much better organized and you can target more specific keywords, but I worry that it may get flagged for light content as some pages may have as little as a 150 word description. I have always been mindful of writing copy for searchers over spiders, but now I'm at a more technical crossroads as far as potentially getting dinged for not having robust content on each page. Here's where you come in...
What do you think is the better of the two options? I like the idea of having the multiple pages because of the ability to hone-in on a keyword and the clean, organized feel, but I worry about the lack of content (and possibly losing out on long-tail opportunities). I'd love to hear your thoughts. Please and thank you. Ready annnnnnnnnnnnd GO!0 -
What is the best tool to view your page as Googlebot?
Our site was done with asp.net and a lot of scripting. I want to see what Google can see and what it can't. What is the best tool that duplicates Googlebot? I have found several but they seem old or inaccurate.
Web Design | | EcommerceSite0