Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Creating 100,000's of pages, good or bad idea
-
Hi Folks,
Over the last 10 months we have focused on quality pages but have been frustrated with competition websites out ranking us because they have bigger sites. Should we focus on the long tail again?
One option for us is to take every town across the UK and create pages using our activities. e.g.
Stirling
Stirling paintball
Stirling Go Karting
Stirling Clay shootingWe are not going to link to these pages directly from our main menus but from the site map.
These pages would then show activities that were in a 50 mile radius of the towns. At the moment we have have focused our efforts on Regions, e.g. Paintball Scotland, Paintball Yorkshire focusing all the internal link juice to these regional pages, but we don't rank high for towns that the activity sites are close to.
With 45,000 towns and 250 activities we could create over a million pages which seems very excessive! Would creating 500,000 of these types of pages damage our site? This is my main worry, or would it make our site rank even higher for the tougher keywords and also get lots of traffic from the long tail like we used to get.
Is there a limit to how big a site should be?
-
Hi Mark!
Thanks for asking this good question. While there is no limit to how big a website can be, I think you can see from the general response here that most members would encourage you to stick to manually developing quality pages rather than automating hundreds of thousands of pages, solely for ranking purposes. I second this advice.
Now, I would like to clarify your business model. Are you a physical, actual business that customers come to, either to buy paintball equipment or to play paintball in a gallery? Or, is your business virtual, with no in person transactions? I'm not quite understanding this from your description.
If the former, I would certainly encourage you to develop a very strong, unique page for each of your physical locations. If you have 10 locations (with unique street addresses and phone numbers), then that would be 10 pages. If you've got 20 locations, that would be 20 pages, etc. But don't approach these with a 'just switch out the city name in the title tags' mindset. Make these pages as exceptional as possible. Tell stories, show off testimonials, share pictures and videos, entertain, educate, inspire. These city landing pages will be intimately linked into your whole Local SEM campaign, provided they each represent a business location with a unique dedicated street address and unique local area code phone number.
But, if you are considering simply building a page for every city in the UK, I just can't see justification for doing so. Ask yourself - what is the value?
There are business models (such as carpet cleaners, chimney sweeps, general contractors, etc.) that go to their clients' locations to serve and for which I would be advising that they create city landing pages for each of their service cities, but this would be extremely regional...not statewide or national or International. A carpet cleaner might serve 15 different towns and cities in his region, and I would encourage him to start gathering project notes and testimonials, videos and photos to begin developing a body of content important enough for him to start creating strong, interesting and unique pages for each of these cities. But I've also had local business owners tell me they want to cover every city in California, for instance, because they think it will help them to do so, and I discourage this.
Even if the business is virtual and doesn't have any in-person transactions with clients or physical locations, I would still discourage this blanketing-the-whole-nation-with-pages approach. A national retailer needs to build up its brand so that it becomes known and visible organically for its products rather than your theoretical approach of targeting every city in the nation. In short order, the mindset behind that approach just doesn't make good horse sense.
And, as others have stated, adding thousands of thin, potentially duplicate pages to any site could definitely have a very negative effect on rankings.
My advice is to make the time to start developing a content strategy for cities in which you have a legitimate presence. If budget means you can't hire a copywriter to help you with this and to speed up the work, accept that this project deserves all the time you can give it and that a slow development of exceptional pages is better than a fast automation of poor quality pages.
Hope this helps!
-
Hi Mark,
If A,C, and E's page is similar to B,D, and F's page it is still consider dupllicate content. Based on Webmaster's definiton:
"Duplicate content generally refers to substantive blocks of content within or across domains that either completely match other content or are appreciably similar"
Each of your pages should be unique and different from other pages.
I suggest you to continue providing quality content and target the long tail keywords. That alone will help you generate more traffic. Furthermore, out ranking is not a problem. You should focus on getting to the frist page (providing quality content with long tail or regular keywords) and when you are on the first page, try to get searchers to click on your link using Title tag and Meta descriptions.
Out ranking just means they are ranked 4th and you are ranked 5th, 6th but as long as you have a better title tag and meta description. I believe searchers will click on the more attractive results.
-
Cookie cutter pages like these stopped working in Google about ten years ago.
If you toss them up I think that your entire site will tank.
I would go back to focusing on quality pages.
-
If the user experience awesome, and people are staying on your site and looking around, great. If you think the 100,000 pages will make search engines love you, machines can never provide the love users can give you.
-
Can you mix content up from your website e.g. paintball site A, C and E on one page and B,D and F on another if the towns are close together? What I'm not sure about is how different in % terms the content actually has to be.
If we have less written content then the amounts of words we have to actually change would be much less.
The challenge we have is we have build the site this time with filtering in mind, so rather than making customers navigate we allow them to be able to search which is much better in terms of getting the activities they want. The downside is now our site does not show for the long tail as we reduced the pages massively.
-
so we dont have the resources if we did it manually but what would happen is the content would be different on each page as we would only show activity sites within a 50 miles radius. And we would make certain text, h1 etc different and relate to the town.
Below are some examples of sites I see doing well ie number 1 using this method
Our content would be much better than say http://www.justpaintballDOTcoDOTuk/site_guide/Aberfeldy.htm or http://www.goballisticDOTcoDOTuk/paintball_in_/ABERFELDY.asp
But as you say getting this wrong is my worry.
-
Hi Mark,
Creating 100,000 pages is definitely good for Search Engine because you have a lot more contents for them to crawl and have more chances your pages might show up on related keywords. However, the problem is do you have enough unique contents you can post on all those 100,000 pages. If you use similar content, I am afraid it will be duplicate contents. You may think changing up the town names will be enough but it might be risky.
If you can create 100,000 unique contents, Sure go ahead. If not, don't take the risk of duplicate contents.
-
Do you have the resources to create unique content for all those pages? Because adding 500,000 pages of duplicate content will absolutely damage your site.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Are "Powered By..." type footer backlinks good or bad for SEO?
Hi guys, We're running a software company which is also selling WP themes amongst other things. We've heard recently that footer backlinks like "Powered by BigBangThemes" might do more harm than good. Some clients usually forget to change them - so we want to make sure we stop including them in case this is true. Thanks!
Intermediate & Advanced SEO | | andy.bigbangthemes0 -
Geo-Redirect: good idea or not?
Hi Mozzers, The background: I have this very corporate .com domain which is used worldwide. Next to that, we have another .com domain which is specifically created for the US visitors. Within the organic rankings, we notice that our corporate domain is ranking much better in the US. Many visitors are arriving on this domain. As it is a corporate domain being used worldwide, they get lost. My questions: I know there are ways to redirect by location. Would it be smart to automatically redirect US visitors for the corporate domain to the commercial US-specific domain? Is it possible to only redirect US visitors and leave the website as it is for visitors from other countries. Won't this harm the corporate website (organically) worldwide? If this would be a good idea, any recommended plugins or concrete procedures? Thank you so much for helping me out!
Intermediate & Advanced SEO | | WeAreDigital_BE
Sander0 -
Javascript to fetch page title for every webpage, is it good?
We have a zend framework that is complex to program if you ask me, and since we have 20k+ pages that we need to get proper titles to and meta descriptions, i need to ask if we use Javascript to handle page titles (basically the previously programming team had NOT set page titles at all) and i need to get proper page titles from a h1 tag within the page. current course of action which we can easily implement is fetch page title from that h1 tag being used throughout all pages with the help of javascript, But this does makes it difficult for engines to actually read what's the page title? since its being fetched with javascript code that we have put in, though i had doubts, is anyone one of you have simiilar situation before? if yes i need some help! Update: I tried the JavaScript way and here is what it looks like http://islamicencyclopedia.org/public/index/hadith/id/1/book_id/106 i know the fact that google won't read JavaScript like the way we have done with the website, But i need help on "How we can work around this issue" Knowing we don't have other options.
Intermediate & Advanced SEO | | SmartStartMediacom0 -
Using the same content on different TLD's
HI Everyone, We have clients for whom we are going to work with in different countries but sometimes with the same language. For example we might have a client in a competitive niche working in Germany, Austria and Switzerland (Swiss German) ie we're going to potentially rewrite our website three times in German, We're thinking of using Google's href lang tags and use pretty much the same content - is this a safe option, has anyone actually tries this successfully or otherwise? All answers appreciated. Cheers, Mel.
Intermediate & Advanced SEO | | dancape1 -
Multiple 301 redirects for a HTTPS URL. Good or bad?
I'm working on an ecommerce website that has a few snags and issues with it's coding. They're using https, and when you access the website through domain.com, theres a 301 redirect to http://www.domain.com and then this, in turn, redirected to https://www.domain.com. Would this have a deterimental effect or is that considered the best way to do it. Have the website redirect to http and then all http access is redirected to the https URL? Thanks
Intermediate & Advanced SEO | | jasondexter0 -
Remove URLs that 301 Redirect from Google's Index
I'm working with a client who has 301 redirected thousands of URLs from their primary subdomain to a new subdomain (these are unimportant pages with regards to link equity). These URLs are still appearing in Google's results under the primary domain, rather than the new subdomain. This is problematic because it's creating an artificial index bloat issue. These URLs make up over 90% of the URLs indexed. My experience has been that URLs that have been 301 redirected are removed from the index over time and replaced by the new destination URL. But it has been several months, close to a year even, and they're still in the index. Any recommendations on how to speed up the process of removing the 301 redirected URLs from Google's index? Will Google, or any search engine for that matter, process a noindex meta tag if the URL's been redirected?
Intermediate & Advanced SEO | | trung.ngo0 -
Should I use both Google and Bing's Webmaster Tools at the same time?
Hi All, Up till now I've been registered only to Google WMT. Do you recommend using at the same time Bing's WMT? Thanks
Intermediate & Advanced SEO | | BeytzNet0 -
How to check a website's architecture?
Hello everyone, I am an SEO analyst - a good one - but I am weak in technical aspects. I do not know any programming and only a little HTML. I know this is a major weakness for an SEO so my first request to you all is to guide me how to learn HTML and some basic PHP programming. Secondly... about the topic of this particular question - I know that a website should have a flat architecture... but I do not know how to find out if a website's architecture is flat or not, good or bad. Please help me out on this... I would be obliged. Eagerly awaiting your responses, BEst Regards, Talha
Intermediate & Advanced SEO | | MTalhaImtiaz0