Is it bad practice to create pages that 404?
-
We have member pages on our site that are initially empty, until the member does some activity. Currently, since all of these pages are soft 404s, we return a 404 for all these pages and all internal links to them are js links (not links as far as bots are concerned). As soon as the page has content, we switch it to 200 and make the links into regular hrefs.
After doing some research, I started thinking that this is not the best way to handle this situation. A better idea would be to noindex/follow the pages (before they have content) and let the links to these pages be real links.
I'd love to hear input and feedback from fellow Mozzers. What are your thoughts?
-
Yair,
See the infographic on this page regarding rel nofollow tags in links, and when you may want to consider using them. Specifically, see the part about User Generated Content:
http://searchengineland.com/infographic-nofollow-tag-172157
However, Google can decide to crawl whatever they want to crawl, whether it is a nofollowed link, links on a page with a nofollow meta tag, or javascript links. If you really want to keep Google out of those portions of the site you should use the robots.txt disallow statement, as I mentioned in your other thread, or use the X-Robots-Tag as described here.
-
Thanks Everett,
As far as I know, nofollows don't conserve crawl budget. The bots will crawl the link, they just wont transfer any PR.
-
I'm sure Jane meant that it would block indexation of the page.
In my opinion you should probably noindex,follow (robots meta tag) the pages and make the internal links just normal links, possibly with a rel= nofollow link attribute, until the user fills out the profile.
I will go look into your other question as well.
PS: The drawback of this solution is that bots will still be spending crawl budget crawling those URLs if you are linking to them internally. [Edited]
-
Thanks for the clear and concise answer, Jane. You hit the nail right on the head! I appreciate your input.
One question, though. You say that noindex will block bot access to these pages. I'm pretty sure the bots will still crawl the pages (if they find them), just they won't be indexed and presumably they won't be "counted against us" like 404 pages. Is that what you meant?
If you have a minute, maybe you can help me out with this question next: http://moz.com/community/q/internal-nofollows
(Side note: Er_Maqul was referring to the original version of the question (before I edited it) where I had mistakenly written that we nofollow the links.)
-
Hi there,
A 404 certainly isn't the best way to handle a new URL / page before it is populated with content. It is good that Google isn't finding these pages yet (as you state in a later comment), but keep in mind that it could - the pages aren't linked to, but there is never any particular guarantee about what Google will and won't find. It's highly unlikely if you don't link to them, but still - it's not worth taking the risk. As you also say, there's no stopping anyone else from linking to the pages and / or for Google to go on a exploratory mission of its own.
As a note, internal 404s / 410s ("Gone") are perfectly okay if they're appropriate for the situation, i.e. a page has been removed. Not ever removed resource has to be 301ed elsewhere. This isn't the case here though.
To my mind, blocking bot access to these pages while they are empty is a better option, and noindex / follow would achieve this. I believe Er_Maqul has misunderstood what you were saying here - there is no "nofollow" in this situation.
-
Got it, I see.
Well, let's see here. I will state I am no expert in this realm. This is much more of a job for the likes of EGOL or RobertFischer. EGOL in particular with his intimate knowledge of NOINDEX
http://www.mattcutts.com/blog/google-noindex-behavior/
(Scroll down and look at the first post on Matt Cutt's blog there.)
That being said, I still have a few thoughts.
I think you certainly could continue to do what you are doing, I also think that Er_Maquai brought up a point that I touched on as well. Obviously the best scenario is to avoid a 404, and to have original content. Now unfortunately if you don't have have any content to write, or the person doesn't in this case, that set's you up for thin content, and a lot of duplicate content.
It seems to me there is no way to avoid having the extra pages without some sort of script or coding which houses the profiles somewhere other than on a separate page or on a flux capacitor somewhere. So going off that you could create a generic "no profile page" that gets published and use the rel=canonical tag.
I take back my prior statement about "ANY" content. Thin, pointless content, is thin and pointless, and won't benefit you at all. I hope that wasn't interpreted that way.
Again, I think this one is somewhat out of my scope of help, and it might even be worth calling in an SEO professional who specializes in forums for a second opinion. It's like having surgery, gotta go to another Dr to verify your diagnosis.
Sorry I couldn't give a better answer!
-
Wow - thanks for the thorough response, HashtagHustler!
Let me explain a little better...
We get hundreds of signups a day. Each new member has a profile page, which is empty until they do something. Sometimes they never do. So we don't link to the empty pages and they return a 404. As soon as the page has some content, we do link to it and it returns a 200.
Google is not reporting 404s for these pages because they are not linked to. In the pat, when we did link to them, Google reported them as soft 404s.
The current system is working fine.
My question is simply if it makes more sense to allow Google to find these pages (link to them) but noindex them since they do not have content (and are considered soft 404s by Google) or if we should continue doing it the way we are today (which makes me a little uncomfortable since we are creating 1000's of pages - that theoretically may be linked to by other sites - that are 404s)?
-
Yes... all 404 can hurt your SEO campaign, because even if they doesn't count for points (i don't know these), at least the spider ignores a bit more or crawl your page more slowly. Because this, you need to get away all your possible 404 errors.
Think, a empty page can have something. You can use a default text, or use a noindex tag while the page are empty, or simply make a standard profile and link to them all empty pages with a rel=canonical (I think there's the best option). But having more pages, even with low quantity of text, that's better who i doesn't have the pages itself, and, very much better than have 404 errors.
Also, think one more thing. Even more changes have your page, more indexed you're get and low times between indexing. There's another reason to have the pages working, even when they doesn't have any custom info. Because the creation of the page are a change, and customization of them are another change too.
-
Good Morning Yair!
A 404 is a 404 plain and simple. And if Google was able to report it, then they were able to get to it. Which basically means, someone else could. I'm not sure what platform you are using currently using but there are plenty of easy options for a quick fix.
If you wanted a simple option you could just throw up a 302 if you are planning on putting content up SOON. Temporary redirects differ from 301 because they are temporary. Google has to evaluate if they should keep the old page, so in this case, when you are planning on launching the old one (well, technically new one) you might be able to use that to your advantage.
A better idea would be to noindex/follow the pages (before they have content) and let the links to these pages be follow links.
I was unsure exactly what you meant by this. The only difference between follow and no follow links is that you are telling Google not to let your linkjuice carry over to the page you are linking to. You are linking to the page without vetting them. My apologies if you already knew this, but I was slightly confused by your sentence. Google will still go to that page to check it out.
Another option: change your schedule and don't put up pages that far in advance that you aren't planning on publishing. I edit everything offline. Google likes new stuff, especially content. Of course they take longevity into account as well, but as far as making a big splash, putting a website up piecemeal is like having people show up at your birthday saying "Suprise!" when you answer the door when you were the one who invited them. The notion is nice, but it just doesn't have the same effect. Not only in the Google realm, but even more so in Social Media.
My favorite option, and my personal recommendation would be to play the cards you were dealt. Get rid of the 404 and the 200 and embrace that you have a new page! Go on and write a profile piece for the member. Write some sort of Biographical data that can act as a placeholder. At this point it doesn't even need to be stuffed with keywords and amazing seo phrases.
On Second Thought: I'm not sure exactly what your forum is for, and if this issue is specific to a few members or if you are referring to bulk membership, but I have a few ideas on how you might be able to extrapolate some of the signup data into even a simple post to avoid getting a 404! Even if you parsed some of the forms from signup and created a simple little one page that displayed, I think that could help.
At the end of the day, Google loves forums that are strong and authoritative. They also understand that every single person on a forum isn't going to post and isn't going to interact. So depending on what kind of forum you have, and what exactly you are doing, some of the forum issues will just have to be accepted. I think it would be more valuable to clean up negative linking and and analyze internal issues and build form recognition etc. than fix soft 404's coming from a handful of users. Again all of this depends on the size of your pool.
Also, you could just track their usage, if someone is logging in all the time and not posting then fine, I would deal with it because who knows what they are doing in the real world. If someone made an account in 1999, and then threw their computer out of the window in Y2K and never bought another one because they still believe that everything crashed, well then maybe its time to fix that 404.
Hope that helps!
Sorry if my train of thought is a little off this morning.... not enough coffee!
-
Thanks for your quick response, Er.
You are correct about the 404s and I realized that what I wrote in the question was a mistake. We don't have any internal links to these pages (not even nofollow). Until there is content on the page, we make all links to the page into js links. I corrected this in the question now.
Concerning what you said about the pages being useful for SEO even without content: I don't think this is correct. Before we started 404ing the empty profile pages, Webmaster was reporting them as soft 404s. Doesn't this mean that they were hurting us (especially since we have many of them)?
-
Any 404 is bad for SEO. Make a little page for the profile without data, and you can gain a lot of SEO even if your pages doesn't have content.
Even if the links have a nofollow, google follows them to see what is in the other side. For this reason, avoid ALWAYS you can to have a link to a 404 page.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Prioritise a page in Google/why is a well-optimised page not ranking
Hello I'm new to Moz Forums and was wondering if anyone out there could help with a query. My client has an ecommerce site selling a range of pet products, most of which have multiple items in the range for difference size animals i.e. [Product name] for small dog
Intermediate & Advanced SEO | | LauraSorrelle
[Product name] for medium dog
[Product name] for large dog
[Product name] for extra large dog I've got some really great rankings (top 3) for many keyword searches such as
'[product name] for dogs'
'[product name]' But these rankings are for individual product pages, meaning the user is taken to a small dog product page when they might have a large dog or visa versa. I felt it would be better for the users (and for conversions and bounce rates), if there was a group page which showed all products in the range which I could target keywords '[product name]', '[product name] for dogs'. The page would link through the the individual product pages. I created some group pages in autumn last year to trial this and, although they are well-optimised (score of 98 on Moz's optimisation tool), they are not ranking well. They are indexed, but way down the SERPs. The same group page format has been used for the PPC campaign and the difference to the retention/conversion of visitors is significant. Why are my group pages not ranking? Is it because my client's site already has good rankings for the target term and Google does not want to show another page of the site and muddy results?
Is there a way to prioritise the group page in Google's eyes? Or bring it to Google's attention? Any suggestions/advice welcome. Thanks in advance Laura0 -
Hreflang and paginated page
Hi, I can not seem to find good documentation about the use of hreflang and paginated page when using rel=next , rel=prev
Intermediate & Advanced SEO | | TjeerdvZ
Does any know where to find decent documentatio?, I could only find documentation about pagination and hreflang when using canonicals on the paginated page. I have doubts on what is the best option: The way tripadvisor does it:
http://www.tripadvisor.nl/Hotels-g187139-oa390-Corsica-Hotels.html
Each paginated page is referring to it's hreflang paginated page, for example: So should the hreflang refer to the pagined specific page or should it refer to the "1st" page? in this case:
http://www.tripadvisor.nl/Hotels-g187139-Corsica-Hotels.html Looking foward to your suggestions.0 -
We are switching our CMS local pages from a subdomain approach to a subfolder approach. What's the best way to handle this? Should we redirect every local subdomain page to its new subfolder page?
We are looking to create a new subfolder approach within our website versus our current subdomain approach. How should we go about handling this politely as to not lose everything we've worked on up to this point using the subdomain approach? Do we need to redirect every subdomain URL to the new subfolder page? Our current local pages subdomain set up: stores.websitename.com How we plan on adding our new local subfolder set-up: websitename.com/stores/state/city/storelocation Any and all help is appreciated.
Intermediate & Advanced SEO | | SEO.CIC0 -
Product with two common names: A separate page for each name, or both on one page?
This is a real-life problem on my ecommerce store for the drying rack we manufacture: Some people call it a Clothes Drying Rack, while others call it a Laundry Drying Rack, but it's really the same thing. Search volume is higher for the clothes version, so give it the most attention. I currently have 2 separate pages with the On-Page optimization focused on each name (URL, Title, h1, img alts, etc) Here the two drying rack pages: clothes focused page and laundry focused page But the ranking of both pages is terrible. The fairly generic homepage shows up instead of the individual pages in Google searches for the clothes drying rack and for laundry drying rack. But I can get the individual page to appear in a long-tail search like this: round wooden clothes drying rack So my thought is maybe I should just combine both of these pages into one page that will hopefully be more powerful. We would have to set up the On-Page optimization to cover both "clothes & laundry drying rack" but that seems possible. Please share your thoughts. Is this a good idea or a bad idea? Is there another solution? Thanks for your help! Greg
Intermediate & Advanced SEO | | GregB1230 -
Targeting local areas without creating landing pages for each town
I have a large ecommerce website which is structured very much for SEO as it existed a few years ago. With a landing page for every product/town nationwide (its a lot of pages). Then along came Panda... I began shrinking the site in Feb last year in an effort to tackle duplicate content. We had initially used a template only changing product/town name. My first change was to reduce the amount of pages in half by merging the top two categories, as they are semantically similar enough to not need their own pages. This worked a treat, traffic didn't drop at all and the remaining pages are bringing in the desired search terms for both these products. Next I have rewritten the content for every product to ensure they are now as individual as possible. However with 46 products and each of those generating a product/area page we still have a heap of duplicate content. Now i want to reduce the town pages, I have already started writing content for my most important areas, again, to make these pages as individual as possible. The problem i have is that nobody can write enough unique content to target every town in the UK via an individual page (times by 46 products), so i want to reduce these too. QUESTION: If I have a single page for "croydon", will mentioning other local surrounding areas on this page, such as Mitcham, be enough to rank this page for both towns? I have approx 25 Google local place/map listings and grwoing, and am working from these areas outwards. I want to bring the site right down to about 150 main area pages to tackle all the duplicate content, but obviously don't want to lose my traffic for so many areas at once. Any examples of big sites that have reduced in size since Panda would be great. I have a headache... Thanks community.
Intermediate & Advanced SEO | | Silkstream0 -
Best practice to avoid cannibalization of internal pages
Hi everyone, I need help from the best SEO guys regarding a common issue : the cannibalization of internal pages between each other. Here is the case : Let's say I run the website CasualGames.com. This website provides free games, as well as articles and general presentation about given categories of Casual Games. For instance, for the category "Sudoku Games", the structure will be : Home page of the game : http://www.casualgames.com/sudoku/ Free sudoku game listings : (around 100 games listed) http://www.casualgames.com/sudoku/free/ A particular sudoku game : http://www.casualgames.com/sudoku/free/game-1/ A news regarding sudoku games : http://www.casualgames.com/sudoku/news/title The problem is that these pages seem to "cannibalize" each other. Explanation : In the SERPS, for the keyword "Casual Games", the home page doesn't appear well ranked and some specific sudoku games page (one of the 100 games) are better ranked although they are "sub-pages" of the category.. Same for the news pages : a few are better ranked than the category page.. I am kind of lost.. Any idea what would be the best practice in this situation? THANKS a LOT.
Intermediate & Advanced SEO | | laboiteac
Guillaume0 -
Should I block temporary pages
I need some SEO advice on an odd scenario: We are launching a new product line (party supplies) on it's own domain (PartySuperCenter.com). Due to some internal/technical reasons we will not be able to launch the site until the summer. We already have the product in our warehouse so the owners want to created a section on our current site (CostumeSuperCenter.com) for the new products. Once the new site is up the product will be removed from our current site and moved to the new site. I am concerned about the effect this will have on our SEO - having thousands of product pages appear and then disappear after a few months. I was thinking about blocking the pages using the "noindex" tag. Is this how you would handle it? Thanks in advance for your help!
Intermediate & Advanced SEO | | costume0 -
How do I set up a 301 redirect if the default settings for our web servers create multiple URLs for the same page?
How do I set up a 301 redirect if the default settings for our web servers create multiple URLs for the same page but only views it as one page?
Intermediate & Advanced SEO | | ibex0