Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Is it better to drip feed content?
-
Hi All,
I've assembled a collection of 5 closely related articles each about 700 words for publishing by linking to them from on one of my pages and would appreciate some advice on the role out of these articles.
Backround: My site is a listings based site and a majority of the content is published on my competitors sites too. This is because advertisers are aiming to spread there adverts wide with the hope of generating more responses. The page I'm targeting ranks 11th but I would like to link it to some new articles and guides to beef it up a bit. My main focus is to rank better for the page that links to these articles and as a result I write up an introduction to the article/guide which serves as my unique content.
Question: Is it better to drip feed the new articles onto the site or would it be best to get as much unique content on as quickly as possible to increase the ratio of unique content vs. external duplicate content on the page that links to these articles**?**
Thank you in advance.
-
Good luck. Like I said, this is just me being silly. I pray to my Google shrine twice a day and this is what it tells me.
All at once or drip feed, either way your content gets up there!
-
Thanks guys for your help. Think I'm going to publish it all at once. Was originally in agreement with Bill but after doing a bit of reading it's probably safe to say that the SE's prioritise good content over content age. I've noticed blogs having slightly inflated PR because of the regular content but it's unlikely I'll be able to keep up regular posts and as a result any benefit derived from drip feeding would fall away when I run out of articles. If it doesn't work I'm calling my lawyer on you guys, hehe kidding :)))))
-
I don't think there is any right or wrong answer to this question. More of a preference.
For me, I like to drip my content.
In my own silly mind, it looks more natural to the search engines rather than dumping a bunch of content on your site.
I also think it keeps the search engines coming back to your site as you posting content through the months and years rather than all at one time.
Mind you. I have no scientific basis for this... just my own anal retentivity. LOL
-
When I have new content I can't wait to get it indexed. So even if I am not promoting it yet on the homepage I will put up links to it on relevant pages just to get spiders into it.
Five articles is no worry.
-
There is no advantage to holding back from a search engine perspective. The only reason I can think of to hold back relates to promotion opportunities for the articles. You could publish one article each week, tweet it and otherwise generate interest around the weekly article. If that is not of interest to you, then I would publish all five articles.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Google ranking content for phrases that don't exist on-page
I am experiencing an issue with negative keywords, but the “negative” keyword in question isn’t truly negative and is required within the content – the problem is that Google is ranking pages for inaccurate phrases that don’t exist on the page. To explain, this product page (as one of many examples) - https://www.scamblermusic.com/albums/royalty-free-rock-music/ - is optimised for “Royalty free rock music” and it gets a Moz grade of 100. “Royalty free” is the most accurate description of the music (I optimised for “royalty free” instead of “royalty-free” (including a hyphen) because of improved search volume), and there is just one reference to the term “copyrighted” towards the foot of the page – this term is relevant because I need to make the point that the music is licensed, not sold, and the licensee pays for the right to use the music but does not own it (as it remains copyrighted). It turns out however that I appear to need to treat “copyrighted” almost as a negative term because Google isn’t accurately ranking the content. Despite excellent optimisation for “Royalty free rock music” and only one single reference of “copyrighted” within the copy, I am seeing this page (and other album genres) wrongly rank for the following search terms: “free rock music”
On-Page Optimization | | JCN-SBWD
“Copyright free rock music"
“Uncopyrighted rock music”
“Non copyrighted rock music” I understand that pages might rank for “free rock music” because it is part of the “Royalty free rock music” optimisation, what I can’t get my head around is why the page (and similar product pages) are ranking for “Copyright free”, “Uncopyrighted music” and “Non copyrighted music”. “Uncopyrighted” and “Non copyrighted” don’t exist anywhere within the copy or source code – why would Google consider it helpful to rank a page for a search term that doesn’t exist as a complete phrase within the content? By the same logic the page should also wrongly rank for “Skylark rock music” or “Pretzel rock music” as the words “Skylark” and “Pretzel” also feature just once within the content and therefore should generate completely inaccurate results too. To me this demonstrates just how poor Google is when it comes to understanding relevant content and optimization - it's taking part of an optimized term and combining it with just one other single-use word and then inappropriately ranking the page for that completely made up phrase. It’s one thing to misinterpret one reference of the term “copyrighted” and something else entirely to rank a page for completely made up terms such as “Uncopyrighted” and “Non copyrighted”. It almost makes me think that I’ve got a better chance of accurately ranking content if I buy a goat, shove a cigar up its backside, and sacrifice it in the name of the great god Google! Any advice (about wrongly attributed negative keywords, not goat sacrifice ) would be most welcome.0 -
How to fix duplicate content for homepage and index.html
Hello, I know this probably gets asked quite a lot but I haven't found a recent post about this in 2018 on Moz Q&A, so I thought I would check in and see what the best route/solution for this issue might be. I'm always really worried about making any (potentially bad/wrong) changes to the site, as it's my livelihood, so I'm hoping someone can point me in the right direction. Moz, SEMRush and several other SEO tools are all reporting that I have duplicate content for my homepage and index.html (same identical page). According to Moz, my homepage (without index.html) has PA 29 and index.html has PA 15. They are both showing Status 200. I read that you can either do a 301 redirect or add rel=canonical I currently have a 301 setup for my http to https page and don't have any rel=canonical added to the site/page. What is the best and safest way to get rid of duplicate content and merge the my non index and index.html homepages together these days? I read that both 301 and canonical pass on link juice but I don't know what the best route for me is given what I said above. Thank you for reading, any input is greatly appreciated!
On-Page Optimization | | dreservices0 -
How long should my website content be (max and min number of words)?
I saw a web site which has been number 1 on Google for a long time, and the home page has 5700 words, but the results show it is not spam, so what would be the recommended word number for a home page?
On-Page Optimization | | Majapopa0 -
Thoughts on archiving content on an event site?
I have a few sites that are used exclusively to promote live events (ex. tradeshows, conference, etc). In most cases these sites content fewer than 100 pages and include information for the upcoming event with links to register. Some time after the event has ended, we would redesign the site and start promoting next years event...essentially starting over with a new site (same domain). We understand the value that many of these past event pages have for users who are looking for info from the past event and we're looking for advice on how best to archive this content to preserve for SEO. We tend to use concise urls for pages on these sites. Ex. www.event.com/agenda or www.event.com/speakers. What are your thoughts on archiving the content from these pages so we can reuse the url with content for the new event? My first thought is to put these pages into an archive, like www.event.com/2015/speakers. Is there a better way to do this to preserve the SEO value of this content?
On-Page Optimization | | accessintel0 -
Duplicate Content - Bulk analysis tool?
Hi I wondered if there's a tool to analyse duplicate content - within your own site or on external sites, but that you can upload the URL's you want to check in bulk? I used Copyscape a while ago, but don't remember this having a bulk feature? Thank you!
On-Page Optimization | | BeckyKey0 -
Stolen Content reposted on other sites. How does this affect ranking?
Visitors often copy and paste my content and post it elsewhere... on Facebook, on Tumblr, on forums and sometimes on competing websites... but they don't link to me. How does Google treat this duplicated content? What is the best way to handle it? File DCMA claims or ask them for a link?
On-Page Optimization | | brianflannery0 -
Add content as blog post or to product pages?
Hi, We have around 40 products which we can produce plenty of in-depth and detailed "how to"-type pieces of content for. Our current plan is to produce a "How to make" style post for each as a long blog post, then link that to the product page. There's probably half a dozen or more of these kind of blog posts that we could do for each product. The reason why we planned on doing it like this is that it would give us plenty of extra pages (blog posts) on their own URL which can be indexed and rank for long tail keywords, but also that we can mention these posts in our newsletter. It'd give people a new page full of specific content that they can read instead of us having to say "Hey! We've updated our product page for X!", which seems a little pointless. Most of the products we sell don't get very many searches themselves; Most get a couple dozen and the odd few get 100-300 each, while one gets more than 2,000 per month. The products don't get many searches as it's a relatively unknown niche when it comes to details, but searches for the "categories" these products are in are very well known (Some broad terms that cover the niche get more than 30,000+ searches a month in the UK and 100,000+ world wide) [Exact].
On-Page Optimization | | azu25
Regarding the one product with more than 2,000 searches; This keyword is both the name of the product and also a name for the category page. Many of our competitors have just one of these products, whereas we're one of the first to have more than 6 variations of this product, thus the category page is acting like our other product pages and the information you would usually find on our product pages, is on the category page for just this product. I'm still leaning towards creating each piece of content as it's own blog post which links to the product pages, while the product pages link to the relevant blog posts, but i'm starting to think that it may be be better to put all the content on the product pages themselves). The only problem with this is that it cuts out on more than 200 very indepth and long blog posts (which due to the amount of content, videos and potentially dozens of high resolution images may slow down the loading of the product pages). From what I can see, here are the pros and cons: Pro (For blog posts):
1. More than 200 blog posts (potentially 1000+ words each with dozens of photos and potentially a video)..
2. More pages to crawl, index and rank..
3. More pages to post on social media..
4. Able to comment about the posts in the newsletter - Sounds more unique than "We've just updated this product page"..
5. Commenting is available on blog posts, whereas it is not on product pages..
6. So much information could slow down the loading of product pages significantly..
7. Some products are very similar (ie, the same product but "better quality" - Difficult to explain without giving the niche away, which i'd prefer not to do ATM) and this would mean the same content isn't on multiple pages.
8. By my understanding, this would be better for Google Authorship/Publishership.. Con (Against blog posts. For extended product pages):
1. Customers have all information in one place and don't have to click on a "Related Blog posts" tab..
2. More content means better ability to rank for product related keywords (All but a few receive very few searches per month, but the niche is exploding at an amazing rate at the moment)..
3. Very little chance of a blog post out-ranking the related product page for keywords.. I've run out of ideas for the 'Con' side of things, but that's why I'd like opinions from someone here if possible. I'd really appreciate any and all input, Thanks! [EDIT]:
I should add that there will be a small "How to make" style section on product pages anyway, which covers the most common step by step instructions. In the content we planned for blog posts, we'd explore the regular method in greater detail and several other methods in good detail. Our products can be "made" in several different ways which each result in a unique end result (some people may prefer it one way than another, so we want to cover every possible method), effectively meaning that there's an almost unlimited amount of content we could write.
In fact, you could probably think of the blog posts as more of "an ultimate guide to X" instead of simply "How to X"...0 -
Does schema.org assist with duplicate content concerns
The issue of duplicate content has been well documented and there are lots of articles suggesting to noindex archive pages in WordPress powered sites. Schema.org allows us to mark-up our content, including marking a components URL. So my question simply, is no-indexing archive (category/tag) pages still relevant when considering duplicate content? These pages are in essence a list of articles, which can be marked as an article or blog posting, with the url of the main article and all the other cool stuff the scheme gives us. Surely Google et al are smart enough to recognise these article listings as gateways to the main content, therefore removing duplicate content concerns. Of course, whether or not doing this is a good idea will be subjective and based on individual circumstances - I'm just interested in whether or not the search engines can handle this appropriately.
On-Page Optimization | | MarkCA0