Duplicate content or not? If you're using abstracts from external sources you link to
-
I was wondering if a page (a blog post, for example) that offers links to external web pages along with abstracts from these pages would be considered duplicate content page and therefore penalized by Google.
For example, I have a page that has very little original content (just two or three sentences that summarize or sometimes frame the topic) followed by five references to different external sources. Each reference contains a title, which is a link, and a short abstract, which basically is the first few sentences copied from the page it links to.
So, except from a few sentences in the beginning everything is copied from other pages.
Such a page would be very helpful for people interested in the topic as the sources it links to had been analyzed before, handpicked and were placed there to enhance user experience.
But will this format be considered duplicate or near-duplicate content?
-
Are you going to get some sort of penalty for it? No. Duplicate content doesn't work that way unless you're just a low-quality or scraper site. Are you going to rank for a lot of keywords in the quoted text? No, probably not.
If there's value in your curation, you could in theory rank for the theme or topic that you're covering with the external quotations. This is especially true if you're pulling together hard-to-find or obscure quotations together, or combining them in an interesting/unique way.
Providing unique content is generally a good way to go in organic search, but there are plenty of aggregation sites succeeding. This was all MetaCritic had before it filled up with user reviews, but it was insanely useful. Don't let anyone tell you that content will get you penalized or something just because it can be found elsewhere. Do cite your sources and think about user comments. If you provide something uniquely valuable to the user, there are ways to make even pure duplicate content work in search.
-
Romanbond,
This is thin content/Panda kind of stuff. If your users find it valuable and outside sources link to your abstract pages, it could pass muster. It's likely though, that those pages will not build up the authority that they need to either rank well themselves or pass along link equity to those pages they link to.
-
Hmmm I would say borderline. If this was the mainstay of posts to a site, then I would be worried. However if you have lots of other content published on a regular basis that is content-rich and engaging, then I would be less worried.
If the main goal here really is for users, rather than SERPS, why not noindex, dofollow the page?
Couldn't you twist this a little though, have a unique intro at the start of the article, then a paragraph of your own thoughts on each topic - adding value and provoking thought, then a link to the topic after that? It's what I do on some of my sites, and it works well!
-
It would probably be duplicate content. The page would be useful for people who stumble upon your site, but why would Google want to rank that page over the actual sources themselves? So your best bet is to add plenty of your own content to that page, or rank the rest of your site and link to this useful resource (not expecting it to rank on its own).
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sudden influx of 404's affecting SERP's?
Hi Mozzers, We've recently updated a site of ours that really should be doing much better than it currently is. It's got a good backlink profile (and some spammy links recently removed), has age on it's side and has been SEO'ed a tremendous amount. (think deep-level, schema.org, site-speed and much, much more). Because of this, we assumed thin, spammy content was the issue and removed these pages, creating new, content-rich pages in the meantime. IE: We removed a link-wheel page; <a>https://www.google.co.uk/search?q=site%3Asuperted.com%2Fpopular-searches</a>, which as you can see had a **lot **of results (circa 138,000). And added relevant pages for each of our entertainment 'categories'.
White Hat / Black Hat SEO | | ChimplyWebGroup
<a>http://www.superted.com/category.php/bands-musicians</a> - this page has some historical value, so the Mozbar shows some Page Authority here.
<a>http://www.superted.com/profiles.php/wedding-bands</a> - this is an example of a page linking from the above page. These are brand new URLs and are designed to provide relevant content. The old link-wheel pages contained pure links (usually 50+ on every page), no textual content, yet were still driving small amounts of traffic to our site.
The new pages contain quality and relevant content (ie - our list of Wedding Bands, what else would a searcher be looking for??) but some haven't been indexed/ranked yet. So with this in mind I have a few questions: How do we drive traffic to these new pages? We've started to create industry relevant links through our own members to the top-level pages. (http://www.superted.com/category.php/bands-musicians) The link-profile here _should _flow to some degree to the lower-level pages, right? We've got almost 500 'sub-categories', getting quality links to these is just unrealistic in the short term. How long until we should be indexed? We've seen an 800% drop in Organic Search traffic since removing our spammy link-wheel page. This is to be expected to a degree as these were the only real pages driving traffic. However, we saw this drop (and got rid of the pages) almost exactly a month ago, surely we should be re-indexed and re-algo'ed by now?! **Are we still being algor****hythmically penalised? **The old spammy pages are still indexed in Google (138,000 of them!) despite returning 404's for a month. When will these drop out of the rankings? If Google believes they still exist and we were indeed being punished for them, then it makes sense as to why we're still not ranking, but how do we get rid of them? I've tried submitting a manual removal of URL via WMT, but to no avail. Should I 410 the page? Have I been too hasty? I removed the spammy pages in case they were affecting us via a penalty. There would also have been some potential of duplicate content with the old and the new pages.
_popular-searches.php/event-services/videographer _may have clashed with _profiles.php/videographer, _for example.
Should I have kept these pages whilst we waited for the new pages to re-index? Any help would be extremely appreciated, I'm pulling my hair out that after following 'guidelines', we seem to have been punished in some way for it. I assumed we just needed to give Google time to re-index, but a month should surely be enough for a site with historical SEO value such as ours?
If anyone has any clues about what might be happening here, I'd be more than happy to pay for a genuine expert to take a look. If anyone has any potential ideas, I'd love to reward you with a 'good answer'. Many, many thanks in advance. Ryan.0 -
Is it ok to ask for a non reciprocal link
hey guys, got mini discussion question. With rapgenius.com getting penalized today, it raises some questions about linking. What they did is definitely not ok. A link scheme involving their own affiliate network is against Google guidelines for sure. So is it ok to ask for a non reciprocal link if there is no incentive involved and no money changes hands? ie. Someone writes an article related to your article topic, or they reference you without a link. Then uou email the webmaster requesting a link... They add it. Is this against the guidelines?
White Hat / Black Hat SEO | | Anti-Alex0 -
11 000 links from 2 blogs + Many bad links = Penguin 2.0\. What is the real cause?
Hello, A website has : 1/ 8000 inbound links from 1 blog and 3000 from another one. They are clean and good blogs, all links are NOT marked as no-follow. 2/ Many bad links from directories that have been unindexed or penalized by Google On the 22nd of May, the website got hurt by Penguin 2.0. The link profile contains many directories and articles. The priority we had so far was unindexing the bad links, however shall we no-follow the blog links as well? Thanks!
White Hat / Black Hat SEO | | antoine.brunel0 -
Black linking exploitation
Hi all After watching our ranking for some primary keywords drop on Google from page 1 to 20 and then totally off the charts in relatively short period I've recently discovered through moz tools that our website along with other competitor sites are victims to black linking (may have the terminology wrong). Two primary words are anchor linked to our domain (www.solargain.com.au) being sex & b$tch through over 4000 compromised sites - mostly Wordpress - many which are high profile sites. Searching through the source code through half a dozen compromised sites I noticed that competitors are also linked using other derogatory terms, but the patterns indicate batch or clustered processing. The hacker has left some evidence as to whom they are representing as I can see some credible discussion forums which contain negative feedback on one particular supplier also among the links. Although this is pretty good evidence to why our ranking has dropped there are some interesting questions: A) is there any way to rectify the 4000 or so black links, mass removal or other. (Doesn't sound feasible)
White Hat / Black Hat SEO | | mannydog
B) some competitors who dominate organic ranking through better optimization don't seem to be affected or apparently affected as much as our site at least. Which questions how much we are affected as a direct result from this hack.
C) is there action or support for industrial espionage?
D) can you request from google to ignore the inbound links and would they not have a duty of care to do so? I'm fairly new to this ugly side of the Internet and would like to know how to approach recovery and moving forward. Thoughts ideas very welcome. Thanks in advance.0 -
Using Programmatic Content
My company has been approached a number of times by computer generated content providers (like Narrative Science and Comtex). They are providing computer generated content to a number of big name websites. Does anyone have any experience working with companies like this? We were burned by the first panda update because we were busing boilerplate forms for content
White Hat / Black Hat SEO | | SuperMikeLewis0 -
How do I place the product link on my blog?
I have a shop and also a blog where I explain better the products on the site, such as: how to use, tips, recipes and more. How do I place the product link on my blog? Should I put a link with nofollow? Should not I put link? To put the link anchor text or just put the page URL? Don’t I need to worry about it?
White Hat / Black Hat SEO | | soulmktpro0 -
Link Quality and Anchor Text
ok I was wondering how to determine the quality of a link and if there is a way to tell that the site linking to you could be passing on penalized link juice to your site. Also i would like to know some of yalls opinion on using anchor text links in articles and blogs. Now that google seems to have taken some of its "importance" away
White Hat / Black Hat SEO | | daugherty0 -
My attempt to reduce duplicate content got me slapped with a doorway page penalty. Halp!
On Friday, 4/29, we noticed that we suddenly lost all rankings for all of our keywords, including searches like "bbq guys". This indicated to us that we are being penalized for something. We immediately went through the list of things that changed, and the most obvious is that we were migrating domains. On Thursday, we turned off one of our older sites, http://www.thegrillstoreandmore.com/, and 301 redirected each page on it to the same page on bbqguys.com. Our intent was to eliminate duplicate content issues. When we realized that something bad was happening, we immediately turned off the redirects and put thegrillstoreandmore.com back online. This did not unpenalize bbqguys. We've been looking for things for two days, and have not been able to find what we did wrong, at least not until tonight. I just logged back in to webmaster tools to do some more digging, and I saw that I had a new message. "Google Webmaster Tools notice of detected doorway pages on http://www.bbqguys.com/" It is my understanding that doorway pages are pages jammed with keywords and links and devoid of any real content. We don't do those pages. The message does link me to Google's definition of doorway pages, but it does not give me a list of pages on my site that it does not like. If I could even see one or two pages, I could probably figure out what I am doing wrong. I find this most shocking since we go out of our way to try not to do anything spammy or sneaky. Since we try hard not to do anything that is even grey hat, I have no idea what could possibly have triggered this message and the penalty. Does anyone know how to go about figuring out what pages specifically are causing the problem so I can change them or take them down? We are slowly canonical-izing urls and changing the way different parts of the sites build links to make them all the same, and I am aware that these things need work. We were in the process of discontinuing some sites and 301 redirecting pages to a more centralized location to try to stop duplicate content. The day after we instituted the 301 redirects, the site we were redirecting all of the traffic to (the main site) got blacklisted. Because of this, we immediately took down the 301 redirects. Since the webmaster tools notifications are different (ie: too many urls is a notice level message and doorway pages is a separate alert level message), and the too many urls has been triggering for a while now, I am guessing that the doorway pages problem has nothing to do with url structure. According to the help files, doorway pages is a content problem with a specific page. The architecture suggestions are helpful and they reassure us they we should be working on them, but they don't help me solve my immediate problem. I would really be thankful for any help we could get identifying the pages that Google thinks are "doorway pages", since this is what I am getting immediately and severely penalized for. I want to stop doing whatever it is I am doing wrong, I just don't know what it is! Thanks for any help identifying the problem! It feels like we got penalized for trying to do what we think Google wants. If we could figure out what a "doorway page" is, and how our 301 redirects triggered Googlebot into saying we have them, we could more appropriately reduce duplicate content. As it stands now, we are not sure what we did wrong. We know we have duplicate content issues, but we also thought we were following webmaster guidelines on how to reduce the problem and we got nailed almost immediately when we instituted the 301 redirects.
White Hat / Black Hat SEO | | CoreyTisdale0