Is it bad practice to create pages that 404?
-
We have member pages on our site that are initially empty, until the member does some activity. Currently, since all of these pages are soft 404s, we return a 404 for all these pages and all internal links to them are js links (not links as far as bots are concerned). As soon as the page has content, we switch it to 200 and make the links into regular hrefs.
After doing some research, I started thinking that this is not the best way to handle this situation. A better idea would be to noindex/follow the pages (before they have content) and let the links to these pages be real links.
I'd love to hear input and feedback from fellow Mozzers. What are your thoughts?
-
Yair,
See the infographic on this page regarding rel nofollow tags in links, and when you may want to consider using them. Specifically, see the part about User Generated Content:
http://searchengineland.com/infographic-nofollow-tag-172157
However, Google can decide to crawl whatever they want to crawl, whether it is a nofollowed link, links on a page with a nofollow meta tag, or javascript links. If you really want to keep Google out of those portions of the site you should use the robots.txt disallow statement, as I mentioned in your other thread, or use the X-Robots-Tag as described here.
-
Thanks Everett,
As far as I know, nofollows don't conserve crawl budget. The bots will crawl the link, they just wont transfer any PR.
-
I'm sure Jane meant that it would block indexation of the page.
In my opinion you should probably noindex,follow (robots meta tag) the pages and make the internal links just normal links, possibly with a rel= nofollow link attribute, until the user fills out the profile.
I will go look into your other question as well.
PS: The drawback of this solution is that bots will still be spending crawl budget crawling those URLs if you are linking to them internally. [Edited]
-
Thanks for the clear and concise answer, Jane. You hit the nail right on the head! I appreciate your input.
One question, though. You say that noindex will block bot access to these pages. I'm pretty sure the bots will still crawl the pages (if they find them), just they won't be indexed and presumably they won't be "counted against us" like 404 pages. Is that what you meant?
If you have a minute, maybe you can help me out with this question next: http://moz.com/community/q/internal-nofollows
(Side note: Er_Maqul was referring to the original version of the question (before I edited it) where I had mistakenly written that we nofollow the links.)
-
Hi there,
A 404 certainly isn't the best way to handle a new URL / page before it is populated with content. It is good that Google isn't finding these pages yet (as you state in a later comment), but keep in mind that it could - the pages aren't linked to, but there is never any particular guarantee about what Google will and won't find. It's highly unlikely if you don't link to them, but still - it's not worth taking the risk. As you also say, there's no stopping anyone else from linking to the pages and / or for Google to go on a exploratory mission of its own.
As a note, internal 404s / 410s ("Gone") are perfectly okay if they're appropriate for the situation, i.e. a page has been removed. Not ever removed resource has to be 301ed elsewhere. This isn't the case here though.
To my mind, blocking bot access to these pages while they are empty is a better option, and noindex / follow would achieve this. I believe Er_Maqul has misunderstood what you were saying here - there is no "nofollow" in this situation.
-
Got it, I see.
Well, let's see here. I will state I am no expert in this realm. This is much more of a job for the likes of EGOL or RobertFischer. EGOL in particular with his intimate knowledge of NOINDEX
http://www.mattcutts.com/blog/google-noindex-behavior/
(Scroll down and look at the first post on Matt Cutt's blog there.)
That being said, I still have a few thoughts.
I think you certainly could continue to do what you are doing, I also think that Er_Maquai brought up a point that I touched on as well. Obviously the best scenario is to avoid a 404, and to have original content. Now unfortunately if you don't have have any content to write, or the person doesn't in this case, that set's you up for thin content, and a lot of duplicate content.
It seems to me there is no way to avoid having the extra pages without some sort of script or coding which houses the profiles somewhere other than on a separate page or on a flux capacitor somewhere. So going off that you could create a generic "no profile page" that gets published and use the rel=canonical tag.
I take back my prior statement about "ANY" content. Thin, pointless content, is thin and pointless, and won't benefit you at all. I hope that wasn't interpreted that way.
Again, I think this one is somewhat out of my scope of help, and it might even be worth calling in an SEO professional who specializes in forums for a second opinion. It's like having surgery, gotta go to another Dr to verify your diagnosis.
Sorry I couldn't give a better answer!
-
Wow - thanks for the thorough response, HashtagHustler!
Let me explain a little better...
We get hundreds of signups a day. Each new member has a profile page, which is empty until they do something. Sometimes they never do. So we don't link to the empty pages and they return a 404. As soon as the page has some content, we do link to it and it returns a 200.
Google is not reporting 404s for these pages because they are not linked to. In the pat, when we did link to them, Google reported them as soft 404s.
The current system is working fine.
My question is simply if it makes more sense to allow Google to find these pages (link to them) but noindex them since they do not have content (and are considered soft 404s by Google) or if we should continue doing it the way we are today (which makes me a little uncomfortable since we are creating 1000's of pages - that theoretically may be linked to by other sites - that are 404s)?
-
Yes... all 404 can hurt your SEO campaign, because even if they doesn't count for points (i don't know these), at least the spider ignores a bit more or crawl your page more slowly. Because this, you need to get away all your possible 404 errors.
Think, a empty page can have something. You can use a default text, or use a noindex tag while the page are empty, or simply make a standard profile and link to them all empty pages with a rel=canonical (I think there's the best option). But having more pages, even with low quantity of text, that's better who i doesn't have the pages itself, and, very much better than have 404 errors.
Also, think one more thing. Even more changes have your page, more indexed you're get and low times between indexing. There's another reason to have the pages working, even when they doesn't have any custom info. Because the creation of the page are a change, and customization of them are another change too.
-
Good Morning Yair!
A 404 is a 404 plain and simple. And if Google was able to report it, then they were able to get to it. Which basically means, someone else could. I'm not sure what platform you are using currently using but there are plenty of easy options for a quick fix.
If you wanted a simple option you could just throw up a 302 if you are planning on putting content up SOON. Temporary redirects differ from 301 because they are temporary. Google has to evaluate if they should keep the old page, so in this case, when you are planning on launching the old one (well, technically new one) you might be able to use that to your advantage.
A better idea would be to noindex/follow the pages (before they have content) and let the links to these pages be follow links.
I was unsure exactly what you meant by this. The only difference between follow and no follow links is that you are telling Google not to let your linkjuice carry over to the page you are linking to. You are linking to the page without vetting them. My apologies if you already knew this, but I was slightly confused by your sentence. Google will still go to that page to check it out.
Another option: change your schedule and don't put up pages that far in advance that you aren't planning on publishing. I edit everything offline. Google likes new stuff, especially content. Of course they take longevity into account as well, but as far as making a big splash, putting a website up piecemeal is like having people show up at your birthday saying "Suprise!" when you answer the door when you were the one who invited them. The notion is nice, but it just doesn't have the same effect. Not only in the Google realm, but even more so in Social Media.
My favorite option, and my personal recommendation would be to play the cards you were dealt. Get rid of the 404 and the 200 and embrace that you have a new page! Go on and write a profile piece for the member. Write some sort of Biographical data that can act as a placeholder. At this point it doesn't even need to be stuffed with keywords and amazing seo phrases.
On Second Thought: I'm not sure exactly what your forum is for, and if this issue is specific to a few members or if you are referring to bulk membership, but I have a few ideas on how you might be able to extrapolate some of the signup data into even a simple post to avoid getting a 404! Even if you parsed some of the forms from signup and created a simple little one page that displayed, I think that could help.
At the end of the day, Google loves forums that are strong and authoritative. They also understand that every single person on a forum isn't going to post and isn't going to interact. So depending on what kind of forum you have, and what exactly you are doing, some of the forum issues will just have to be accepted. I think it would be more valuable to clean up negative linking and and analyze internal issues and build form recognition etc. than fix soft 404's coming from a handful of users. Again all of this depends on the size of your pool.
Also, you could just track their usage, if someone is logging in all the time and not posting then fine, I would deal with it because who knows what they are doing in the real world. If someone made an account in 1999, and then threw their computer out of the window in Y2K and never bought another one because they still believe that everything crashed, well then maybe its time to fix that 404.
Hope that helps!
Sorry if my train of thought is a little off this morning.... not enough coffee!
-
Thanks for your quick response, Er.
You are correct about the 404s and I realized that what I wrote in the question was a mistake. We don't have any internal links to these pages (not even nofollow). Until there is content on the page, we make all links to the page into js links. I corrected this in the question now.
Concerning what you said about the pages being useful for SEO even without content: I don't think this is correct. Before we started 404ing the empty profile pages, Webmaster was reporting them as soft 404s. Doesn't this mean that they were hurting us (especially since we have many of them)?
-
Any 404 is bad for SEO. Make a little page for the profile without data, and you can gain a lot of SEO even if your pages doesn't have content.
Even if the links have a nofollow, google follows them to see what is in the other side. For this reason, avoid ALWAYS you can to have a link to a 404 page.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Indexed Answer Box Result Leads to a 404 page?
Hey everyone, One of my clients is currently getting an answer box (people also ask) result for a page that is no longer live. They migrated their site approximately 6 months ago, and the old page is for some reason still indexed in the (people also asked) results. Weird thing is that this page leads to a 404 error. Why the heck is Google showing this? Are there separate indexes for "people also asked" results, and regular organic listings? Has anyone ever seen/experienced something like this before? Any insight would is much appreciated
Intermediate & Advanced SEO | | HSawhney0 -
Link Removal Request Sent to Google, Bad Pages Gone from Index But Still Appear in Webmaster Tools
| On June 14th the number of indexed pages for our website on Google Webmaster tools increased from 676 to 851 pages. Our ranking and traffic have taken a big hit since then. The increase in indexed pages is linked to a design upgrade of our website. The upgrade was made June 6th. No new URLS were added. A few forms were changed, the sidebar and header were redesigned. Also, Google Tag Manager was added to the site. My SEO provider, a reputable firm endorsed by MOZ, believes the extra 175 pages indexed by Google, pages that do not offer much content, may be causing the ranking decline. My developer submitted a page removal request to Google via Webmaster tools around June 20th. Now when a Google search is done for site:www.nyc-officespace-leader.com 851 results display. Would these extra pages cause a drop in ranking? My developer issued a link removal request for these pages around June 20th and the number in the Google search results appeared to drop to 451 for a few days, now it is back up to 851. In Google Webmaster Tools it is still listed as 851 pages. My ranking drop more and more everyday. At the end of displayed Google Search Results for site:www.nyc-officespace-leader.comvery strange URSL are displaying like:www.nyc-officespace-leader.com/wp-content/plugins/... If we can get rid of these issues should ranking return to what it was before?I suspect this is an issue with sitemaps and Robot text. Are there any firms or coders who specialize in this? My developer has really dropped the ball. Thanks everyone!! Alan |
Intermediate & Advanced SEO | | Kingalan10 -
Will Using Attributes For Landing Pages In Magento Dilute Page Rank?
Hello Mozzers! We have an ecommerce site built on Magento. We would like to use attribute filters in our layered navigation for landing page purposes. Each page will have a unique URL, Meta Title and Meta Description. For example: URL: domain.com/art/abstract (category is Art, attribute is Abstract) Title: Abstract Art For Sale Meta: Blah Blah Blah Currently these attribute pages are not being indexed by google as they are set in google parameters. We would like to edit google parameters to start indexing some of the attribute filters that users search for, so they can be used as landing pages. Does anyone have experience with this? Is this a good idea? What are the consequences? Will this dilute Page Rank? Could this destroy the world? Cheers! MozAddict
Intermediate & Advanced SEO | | MozAddict0 -
Targeting local areas without creating landing pages for each town
I have a large ecommerce website which is structured very much for SEO as it existed a few years ago. With a landing page for every product/town nationwide (its a lot of pages). Then along came Panda... I began shrinking the site in Feb last year in an effort to tackle duplicate content. We had initially used a template only changing product/town name. My first change was to reduce the amount of pages in half by merging the top two categories, as they are semantically similar enough to not need their own pages. This worked a treat, traffic didn't drop at all and the remaining pages are bringing in the desired search terms for both these products. Next I have rewritten the content for every product to ensure they are now as individual as possible. However with 46 products and each of those generating a product/area page we still have a heap of duplicate content. Now i want to reduce the town pages, I have already started writing content for my most important areas, again, to make these pages as individual as possible. The problem i have is that nobody can write enough unique content to target every town in the UK via an individual page (times by 46 products), so i want to reduce these too. QUESTION: If I have a single page for "croydon", will mentioning other local surrounding areas on this page, such as Mitcham, be enough to rank this page for both towns? I have approx 25 Google local place/map listings and grwoing, and am working from these areas outwards. I want to bring the site right down to about 150 main area pages to tackle all the duplicate content, but obviously don't want to lose my traffic for so many areas at once. Any examples of big sites that have reduced in size since Panda would be great. I have a headache... Thanks community.
Intermediate & Advanced SEO | | Silkstream0 -
Getting individual website pages to rank for their targeted terms instead of just the home page
Hi Everyone, There is a pattern which I have noticed when trying to get individual pages to rank for the allocated targeted terms when I execute an SEO campaign and would been keen on anyones thoughts on how they have effectively addressed this. Let me try and explain this by going through an example: Let's say I am a business coach and already have a website where it includes several of my different coaching services. Now for this SEO campaign, I'm looking to improve exposure for the clients "business coaching" services. I have a quick look at analytics and rankings and notice that the website already ranks fairly well for that term but from the home page and not the service page. I go through the usual process of optimising the site (on-page - content, meta data, internal linking) as well as a linkbuilding campaign throughout the next couple of month's, however this results in either just the home page improving or the business page does improve, but the homepage's existing ranking has suffered, therefore not benefiting the site overall. My question: If a term already ranks or receives a decent amount of traffic from the home page and not from the page that its supposed to, why do you think its the case and what would you be your approach to try shift the traffic to the individual page, without impacting the site too much?. Note: To add the home page keyword target term would have been updated? Thanks, Vahe
Intermediate & Advanced SEO | | Vahe.Arabian0 -
Duplicate page content and Duplicate page title errors
Hi, I'm new to SeoMoz and to this forum. I've started a new campaign on my site and got back loads of error. Most of them are Duplicate page content and Duplicate page title errors. I know I have some duplicate titles but I don't have any duplicate content. I'm not a web developer and not so expert but I have the impression that the crawler is following all my internal links (Infact I have also plenty of warnings saying "Too many on-page links". Do you think this is the cause of my errors? Should I implement the nofollow on all internal links? I'm working with Joomla. Thanks a lot for your help Marco
Intermediate & Advanced SEO | | marcodublin0 -
Best practice for removing pages
I've got some crappy pages that I want to delete from a site. I've removed all the internal links to those pages and resubmitted new site maps that don't show the pages anymore, however the pages still index in search (as you would expect). My question is, what's the best practice for removing these pages? Should I just delete them and be done with it or make them 301 re-direct to a nicer generic page until they are removed from the search results?
Intermediate & Advanced SEO | | PeterAlexLeigh0 -
Pages On Subfolder Not Ranking
A subdirectory/folder on our website doesn't seem to rank for any keywords where the same type of pages on the same competition level keywords rank perfectly fine. For awhile the pages weren't getting indexed but were crawled regularly. Can't seem to figure the problem out.
Intermediate & Advanced SEO | | bprimeelitellc0