News Errors In Google Search Console
-
Years ago a site I'm working on was publishing news as one form of content on the site.
Since then, has stopped publishing news, but still has a q&a forum, blogs, articles... all kinds of stuff.
Now, it triggers "News Errors" in GWT under crawl errors. These errors are
"Article disproportionately short"
"Article fragmented" on some q&a forum pages
"Article too long" on some longer q&a forum pages
"No sentences found"
Since there are thousands of these forum pages and it's problem seems to be a news critique, I'm wondering what I should do about it. It seems to be holding these non-news pages to a news standard:
https://support.google.com/news/publisher/answer/40787?hl=en
For instance, is there a way and would it be a good idea to get the hell out of Google News, since we don't publish news anymore? Would there be possible negatives worth considering?
What's baffling is, these are not designated news urls. The ones we used to have were /news/title-of-the-story per...
https://support.google.com/news/publisher/answer/2481373?hl=en&ref_topic=2481296
Or, does this really not matter and I should just blow it off as a problem.
The weird thing is that we recently went from http to https and The Google News interface still has us as http and gives the option to add https, which I am reluctant to do sine we aren't really in the news business anymore.
What do you think I should do?
Thanks!
-
Update: 5 months later, the problem has long since gone away.
-
Thanks for the answers Matthew & Martijn. So, I'm going to go with what I have in place... 301s to the main forums page to pick up/recycle the incoming links to that section and removing the /news category from where Google News looks for news, but not leave the Google News program altogether right now.
Thanks, again!
-
I'd opt for 301s if there is any link equity on those pages worth pushing to the forum pages. If not, I'd robots.txt them out to save on wasted crawl.
-
HI Matthew,
Thanks for the insight. On your, "I think the important part is that Google doesn't suddenly crawl a bunch of 404s that they think are news pages of some importance" would you robots.txt out the old category they looked at (/news) or leave it and figure out via the mass of 301s and the removal from the url structure we told them to crawl, so that Google figures it out on it's own?
-
Hi there,
It wouldn't be a problem to get out of Google News, but by the same logic, it wouldn't hurt just to leave it alone and let them figure out you're not publishing news anymore. It won't affect your web search rankings, since you're not targeting traffic from the news onebox or news.google.com.
These old pages redirect to forum pages? I think the important part is that Google doesn't suddenly crawl a bunch of 404s that they think are news pages of some importance.
-
Possible, hard to guess what your current set-up looks like. What you could do alternatively is set up a robots.txt with Disallow statement that are only targeting the Google News bot instead of just the general Google Bot.
-
Hi Martijn,
We still have the old /news url still in Google News. I don't think we've ever submitted a news site map.
Is there any downside to deleting the url structure they're currently looking at, which does forward to our forum? Or, would it be better to just get out of Google News altogether, since we don't really have or publish news anymore? Is there a downside to that?
To summarize... delete the url they look to as news or get out of Google News altogether... what do you think?
Thanks!
-
Are you still submitting a Google News Sitemap to Google Search Console? Because usually that's the biggest reason where these errors are coming from because Google is picking up these kind of 'new' pages/articles as news content and then seeing it doesn't match with their guidelines.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
My competitor is ranking above me for a branded search in Google. How can I come back on top?
I work with an organization that is ranking #2 for a branded search term, second to a competitor. They have zero similarity between their names, and we've worked with them to up their SEO game around all major areas (one drawback: SquareSpace is killing their site speed). Their DA is 59, the competitor's DA is 77. What are some smart, specific ways that we can help our client come back out on top?
Intermediate & Advanced SEO | | ogiovetti0 -
SEO Value of Google+?
Hi Mozers, Does having a Google+ page really impact SEO? Thanks, Yael
Intermediate & Advanced SEO | | yaelslater1 -
Search console site verification
I've been going on the assumption that when verifying a website in search console, it's always good to register and verify all variants of the site URL: http https www non-www However, if you create redirects to the preferred URL, is it really necessary to register/virfy of the other three? If so, why?
Intermediate & Advanced SEO | | muzzmoz0 -
Href Lang Errors
Hi Im getting href lang errors in my site audit for having no self referntial tags. I can't see anything wrong? Can anyone enlighten me 🙂 From this page ** http://www.key.co.uk/en/key/0-4-tonne-wll-6mm-single-leg-wire-rope-sling-580a078n**
Intermediate & Advanced SEO | | BeckyKey0 -
I have 6 URL errors in GSC showing a 500 error code. How do I fix?
I am not sure how to fix some errors that are popping up in Google Search Console. The response codes showing are all: 500 error code I need some advice as to how to fix these. What are my options?
Intermediate & Advanced SEO | | pmull0 -
Low on Google ranking despite error-free!?
Hi all, I'm following up on a recent post i've made about our indexing and especially ranking problems in Google: http://moz.com/community/q/seo-impact-classifieds-website Thanks to all good comments we managed to get rid of most of our crawl errors and as a result our high priority /duplicated content decreased from +22k to 270. In short, we created canonical urls, run an xml sitemap, used url parameters in GWT, created h1 and meta description for each ad posted by users etc. I then used google fetch a few times (3 weeks ago and last week) both for desktop and mobile version for re-approval. Nothing really improves in google rankings (all our core keywords are ranked +50)since months now: yet yahoo and bing organic traffic went up and is 3x higher than google's. In the meanwhile we're running paid campagins on facebook and adwords since months already to keep traffic consistent, yet this is eating up our budget, even though our ctr and conversion rates are good. I realize we might have to create more content on-site and through social media, but right now our social media traffic is already around 50% and we are using more of twitter and google+ as well since recently. Our organic traffic is only 14%; with google only a third of that. In the end, I believe this breakdown should look more something like organic 50%-70%, (paid)social,referral and direct traffic. 50%-30%... I can't believe we are hit by a penalty although this looks like it is the case. Especially while yahoo and bing traffic goes up and google does not. Should I wait for a signal once our site is "approved" again through GWT fetch? Or am i missing something that i need to check as well to improve these rankings? Thanks for your help! Ivor ps: ask me for additional stats or info in a pm if needed!
Intermediate & Advanced SEO | | ivordg0 -
Google Tag Manager
Has anyone used Google Tag Manager and do you feel it is worth it?
Intermediate & Advanced SEO | | ChristinaRadisic0 -
"nocontent" class use for Google Custom Search: SEO Ramifications?
Hi all, Have a client that uses Google Custom Search tool which is crawling, indexing and returning millions of irrelevant results for keywords that are on every page of the site. IT/Web dev. team is considering adding a class attribute to prohibit Google Custom Search from indexing bolierplate content regions. Here's the link to Google's custom search help page: http://support.google.com/customsearch/bin/answer.py?hl=en&answer=2364585 "...If your pages have regions containing boilerplate content that's not relevant to the main content of the page, you can identify it using the nocontent class attribute. When Google Custom Search sees this tag, we'll ignore any keywords it contains and won't take them into account when calculating ranking for your Custom Search engine. (We'll still follow and crawl any links contained in the text marked nocontent.) To use the nocontent class attribute, include the boilerplate content in a tag (for example, span or div) like this: Google Custom Search also notes:"Using nocontent won't impact your site's performance in Google Web Search, or our crawling of your site, in any way. We'll continue to follow any links in tagged content; we just won't use keywords to calculate ranking for your Custom Search engine."Just want to confirm if anyone can forsee any SEO implications the use of this div could create? Anyone have experience with this?Thank you!
Intermediate & Advanced SEO | | MRM-McCANN0