Are online tools considered thin content?
-
My website has a number of simple converters.
For example, this one converts spaces to commas
https://convert.town/replace-spaces-with-commasNow, obviously there are loads of different variations I could create of this:
Replace spaces with semicolons
Replace semicolons with tabs
Replace fullstops with commasSimilarly with files:
JSON to XML
XML to PDF
JPG to PNG
JPG to TIF
JPG to PDF
(and thousands more)If somoene types one of those into Google, they will be happy because they can immediately use the tool they were hunting for.
It is obvious what these pages do so I do not want to clutter the page up with unnecessary content.
However, would these be considered doorway pages or thin content or would it be acceptable (from an SEO perspective) to generate 1000s of pages based on all the permutations?
-
Ah - sorry for my misunderstanding. So you are leaning towards combining the pages.
So unit-conversion.info has a combined page: http://www.unit-conversion.info/metric.html
When I search for "convert from micro to deci", they appear as number 8. If I click on their page, it defaults to base and mega, so I then have to change the dropdowns.
The number 1 result for that search is this page https://www.unitconverters.net/prefixes/micro-to-deci.htm - it has Micro and Deci preselected.
unit-conversion.info only has 460 but Unitconverters.net has 50,000 pages indexed by Google. Despite the "thin content", they still appear number 1 (admittedly, this may be due to other factors).
As far as user experience goes, I would prefer to land on unitconverters.net because I have less things to click.
I guess the art is in finding the sweet spot in being able to give a search result with context without spinning out too much thin content.
Thanks again for your detailed response!
-
Hi again,
sorry if I have not expressed myself very well.
In my opinion, you would have only 1 page for each of those tools (with all the conversion options), and along the text of that page (+ title & meta description), there would be optimized the generic keywords like "replace character tool", "replace characters online"... and the conversion specific ones like "replace space with columns", without abusing to avoid keyword stuffing / spam.
The same for the Convert Image Tool, just one page, like this people did:Â unit-conversion.info with the conversion text tool and all the others.
More pages than that would surely create thin content and would divide the authority between all that pages instead of having all that authory in 1 quality page that optimizes along text and metas the most searched of the conversion options of each tool.
In any case, if you create additional pages for the most commonly searched-for variants (just a few), that could be acceptable as you said.
Greetings!
-
Yes I was thinking along the same lines - if I create a page for commonly searched-for variants, then that will be an acceptable "thin page".
OK, so if I understand correctly, you would suggest having one generic "replace text" page. The phrase variants - "replace character tool", "replace characters online", "replace text tool", should appear throughout that same page (not on separate pages).
The following SEPARATE pages would have the find / replace textboxes of the generic converter prefilled (because they are commonly searched for):
- Replace spaces with columns
- Replace spaces with semicolons
- Replace semicolons with spaces
- Replace and with &
...and all other common but relevant search phrases
But you would NOT create a separate page for:
- Replace question mark with space
- Replace the letter t with the letter b
Does that sound right to you?
Then for the Convert Image tool, wouldn't it be best (in a similar fashion) to have one generic tool but then the common searches prefilled on separate pages:
- Convert image to image
- Convert Image to GIF
- Convert PNG to JPG
- Convert PNG to GIF
(and perhaps 100 others)
Each of these tools are different in functionality and will be more helpful to the user if they are prefilled with what they are looking for?
-
So I guess that is actually my argument - that each tool deserves its own page (if it is something commonly searched for). The user experience is not as good if they search for "convert spaces to semicolons", then land on a page where they have to also enter a space and a semicolon before they get what they want. If these are prefilled, surely the user would prefer that. Will Google realise that users prefer that though? That is the big question.
OK - if I don't fill the page with spam, then it won't be considered a gateway page.
Thank you for your response.
-
Hi
It's a difficult question.
By one side, it would be interesting for the searcher to have directly access to the tool with the exact function they are looking for.
By the other, many functions are very similar and they will surely have very similar content that doesn't provide new interesting information (thin content).
I think you should go for the point between this sides. I mean, you can create many different tools, but tools that group all similar functions.
For example:
Replace Character Tool (you can replace with this any character or text by any other). Here you have an example of this tool: http://www.unit-conversion.info/texttools/replace-text/. In this tool you can moderately optimize all the keywords related to the different functions, by mentioning them on the text, h1-h2-h3, or in the Title / Meta Description. Don't try to optimize all different variants because there are too much. Go for the most searched ones (use Google Keyword Planner or a similar tool to identify them). You should also optimize the variants of "replace character tool" like "replace characters online" or "replace text tool", (important to also use "free" if the tools are free)
The same for image conversion with Convert Image Tool ("online picture conversion" + "free convert img tool"... + most popular img format conversion like "png to jpg conversion tool"), all in the same page.
Hope that helps!
-
Hi there,
My personal recommendation here, if possible, would be to compile all of the tools into one easy to use page. So all of the file converting permutations would be under one page and all of the 'replace' tools will be under another page.
Not only would this be better user experience but also you wouldn't clog up your site with thin pages from the multiple permutations of the pages.
You could of course argue that each tool deserves its own page because technically they each do different things.
What would make any one of these pages into a gateway page is if you bulked them out with a large amount of content that was specifically designed for search engines.
I hope this helps to answer your question
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Third part http links on the page source: Social engineering content warning from Google
Hi, We have received "Social engineering content" warning from Google and one of our important page and it's internal pages have been flagged as "Deceptive site ahead". We wonder what's the reason behind this as Google didn't point exactly to the specific part of the page which made us look so to the Google. We don't employ any such content on the page and the content is same for many months. As our site is WP hosted, we used a WordPress plugin for this page's layout which injected 2 http (non-https) links in our page code. We suspect if this is the reason behind this? Any ideas? Thanks
White Hat / Black Hat SEO | | vtmoz1 -
Screaming Frog tool left me stumped
Hi there again, I found a major cloaking hack in our client's website that is really well camouflaged and all the seo tools that I tried to help me check for cloaking couldn't find it. I know that screaming frog is a great tool and I want to use it to help me, however, I can't seem to get my way around their system that I downloaded. Can you help me with the screaming frog program? Do you know where I can make a full site check for cloaking, maybe there are more links that I wasn't notified about? I would really appreciate if you could help me with that. Thanks so much, Ruchy
White Hat / Black Hat SEO | | Ruchy2 -
Apparent Bot Queries and Impressions in Webmaster Tools
I've been noticing some strange stats in Google Webmaster Tools for my forum, which has been getting spam queries with impressions and no clicks. See the queries in the attached images. This might be a motive for the spammers or scrapers. I set the date range to just 22 Aug - 22 Nov and I see very obviously the spike is due to impressions. Questions: What should/can I do? Is Google doing something about this? How to avoid this? o6gKB
White Hat / Black Hat SEO | | SameerBhatia0 -
Pages mirrored on unknown websites (not just content, all the HTML)... blackhat I've never seen before.
Someone more expert than me could help... I am not a pro, just doing research on a website... Google Search Console shows many backlinks in pages under unknown domains... this pages are mirroring the pages of the linked website... clicking on a link on the mirror page leads to a spam page with link spam... The homepage of these unknown domain appear just fine... looks like that the domain is partially hijacked... WTF?! Have you ever seen something likes this? Can it be an outcome of a previous blackhat activity?
White Hat / Black Hat SEO | | 2mlab0 -
Google webmasters tools, Majestic and Ahref in a simple case study (Bad links and Good links)
Hey guys, This case study started from here. A simple summary, I discover that I got +1000 backlinks from Blogspot though Google webmasters tools after making a connection with owners of these blogs which points to my new blog. Before starting I proudly invite Thomas Zickell and Gary Lee in this discussion. I wish you accept my invitation. Let's go to the main point, I've used Google webmaster tools so I will start with. Then Ahref which used by **Thomas **and then Majestic which used by Gary. Take a look at "001" screenshot, you will see that Google webmaster tools discovered 1291 links points to my site. Take another look at "002" screenshot, you will find that there are 22 domains points to my site. Most of them are good links since they are coming from websites such as Google.com, Wikipedia.org, Reddit, Shoutmeload, WordPress.org, ...etc. Beside +1000 backlinks came from Blogspot.com (blogs). Also, there's some bad links such as this one came from tacasino.com Necessary to say that I've got some competitors and they nicely asked me to stop the competition for some keywords and I've ignored their request. So, I'm not surprised to see these bad links. At "002" screenshot, we can see that Google didn't discover the bad links as they discovered the good links. And they discovered a lot of backlinks which not discovered by any other tools. **Let's move to Ahref, ** I will use screenshots provided by Thomas. At "003" screenshot, you can see Ahref report that say 457 links from 10 domains. By the way, social engagements data are wrong. I got more than zero engagements .. really. At "004" screenshot, you can see domains points to my site, links with anchor text. Take a look at the second link you will find that it's a spammy link coming from PR2 home page since it's is over optimized. the third link is also a spammy link since it coming from a not-relevant website. Beside other bad links need to be removed. So, Ahref didn't discover all of my good links. Instead of that it discovered few good links and a lot of bad links. In a case like this a question come needs to be answered since there are some people trying so hard to hurt my site, Do I have to remove all this bad links? Or, just links discovered by Google. Or, Google understand the case? **Let's move to majestic, ** Gray Lee provided data from majestic which say "10 Unique Referring Domains, with 363 links, 2 domains make up a majority." Since Gray didn't take any screenshots I will provide mine. At "005" screenshot, you can see some of the bad links discovered by Majestic. Not all of them discovered by Ahref or Google. In the other hand, Majestic didn't discover all of my Good links. Also, there's a miss understand I would like to explain here. When I published the Discussion about +1000 link. Some people may think that I trying to cheat you by providing fake info and this totally wrong. I said before and I'm saying that again you are so elite and I respect you. Also, I'm preparing for an advanced case study about this thing. If any expert would like to join me this will be great. Thank you for reading and please feel free to share you thoughts, knowledge and experience in this Discussion. EE5bFNc jYg21cf Xyfgp28.png iR4UOwi.png D1pGAFO
White Hat / Black Hat SEO | | Eslam-yosef1 -
Site redesign what to consider to avoid any issues
Hi GUYS I want to avoid getting myself into a bad situation with google, so I'm just wanting to know if there are any steps I would need to take whilst I'm redesigning and developing my site as I'm currently deploying our new designs. One thing I noticed, i have my new designs and content on our development server to run through any checks before deploying it to the live environment, however while our live site is up, I have duplicate content on the live site that exactly matches the dev site for obvious reasons but do I need to tell google that the dev site is for development purposes only so google knows I'm not duplicating content? I have searched around to find some more info about this, if anyone has some insight i would be glad to know your thoughts. Thank you in advance
White Hat / Black Hat SEO | | edward-may0 -
Content within a toggle, Juice or No Juice?
Greetings Mozzers, I recently added a significant amount of information within a single page utilizing toggles to hide the content from a user and for them to see it they must click to reveal. Â Since technically the code is reading "display:none" to start, would that be considered "Black Hat" or "Not There" to crawlers? Â It isn't displayed in any sort of spammy way. Â It is more for the UX of the visitor that toggles were utilized. Thoughts and advice is greatly appreciated!
White Hat / Black Hat SEO | | MonsterWeb280 -
I would like to know if there is a tool to know what keywords
Hi everyone, I am looking for a keywords searcher or a program that can help me to know which keywords my competitors are using. thanks!
White Hat / Black Hat SEO | | lnietob0