How do I identify what is causing my Duplicate Page Content problem?
-
Hello,
I'm trying to put my finger on what exactly is causing my duplicate page content problem... For example, SEOMoz is picking up these four pages as having the same content:
http://www.penncare.net/ambulancedivision/braunambulances/express.aspx
http://www.penncare.net/ambulancedivision/recentdeliveries/millcreekparamedicservice.aspx
http://www.penncare.net/ambulancedivision/recentdeliveries/monongaliaems.aspx
http://www.penncare.net/softwaredivision/emschartssoftware/emschartsvideos.aspx
As you can tell, they really aren't serving the same content in the body of the page. Anybody have an idea what might be causing these pages to show up as Duplicate Page Content? At first I thought it was the photo gallery module that might be causing it, but that only exists on two of the pages...
Thanks in advance!
-
Ah right - OK then.
With regards to data coming back from SEOmoz's crawler, I might be tempted to ask them what it is seeing. I should really have a look at this myself because I haven't yet.
-
I'm currently getting that information from Moz's own web crawler wherein it tells me the pages of that have Duplicate Page Content and the other URLs that that duplicate content exists on.
With regard to the 301's - I have rewrite rules setup to 1.) set all requests to lowercase 2.) trim off home.aspx 3.) append www. to the beginning of the request, etc. When processed these should function as a single redirect / rewrite.
-
Before looking at the duplicate content (what did you use to find there is duplicate content?)... a quick question - you have a lot of 301's. Just want to check, are these just a single redirect or a redirect of a redirect etc?
-
I would add some content to these pages to help differentiate. None of them are text heavy so it may be hard for spiders to see a difference. Add a summary, maybe a text translation of what is in the vids, etc
-
Thanks for your reply... I guess more specifically I was wondering what it was about these particular page elements that makes search engines incapable of deciphering them from one another.
-
- Search engines don't know which version(s) to include/exclude from their indices
- Search engines don't know whether to direct the link metrics (trust, authority, anchor text, link juice, etc.) to one page, or keep it separated between multiple versions
- Search engines don't know which version(s) to rank for query results
When duplicate content is present, site owners suffer rankings and traffic losses and search engines provide less relevant results.
Hope this helps!
Resources, http://www.seomoz.org/learn-seo/duplicate-content
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What’s the best tool to visualize internal link structure and relationships between pages on a single site?
I‘d like to review the internal linking structure on my site. Is there a tool that can visualize the relationships between all of the pages within my site?
Web Design | | QBSEO0 -
Managing Removed Content
I am a Real Estate Developer. Once a home goes off market (is sold), I had been using a 404 for that page. The problem: When the home goes up on market again, google will not re-index the new page (same URL) I have also tried to manage it a different way. Instead of removing the page, I left it as-is. At some later point time, the house goes back up on the market. The page is refreshed with new content. However, google decides to use cached version. Please note in either case, the property appears on the main page for a period of indexing. I have been doing this for 10 years, the problem is increasing with time.
Web Design | | Buckey0 -
Any alternative techniques to display tabbed content without using Javascript / JSON and be SEO Friendly?
John Mueller's input in the EGWMH hangout suggests that Google MAY ignore expandable content served by Javascript. Are there any alternative techniques to display tabbed content without using Javascript / JSON and be SEO Friendly? I do however view these as good for website interactivity and UX - and see many examples of websites performing well and ranking highly whilst using these techniques - are there any Google friendly ways to serve content on a page so that search bots can recognise and choose to crawl / consume the content as legitimate fodder?
Web Design | | Fergclaw0 -
My 404 page is showing a 4xx error. How can that be fixed?
My actual 404 page is giving a 4xx error.
Web Design | | sbetzen
The page address is http://www.ecowindchimes.com/v/404.asp It loads fine... it is the page all 404's are directed to. Why is it showing a 404 error. The page works. How can this be fixed? Stephen0 -
Flag page elements to not be loaded by Instapaper and co.
Does anybody know if there is a way to mark certain elements (especially navigation menus) so that instapaper and co don't pull them? I'm looking for a quick solution (best would be if it was CSS based) nothing fancy like parsing the user-agent. That would be plan B. I've added role="navigation" id="navigation" and class="navigation" to the nav elements in hope that it would work. Seems like it does not; sometimes the elements are present in the page generated by instapaper, sometimes not. Thank you for any replies and have a great day! Jan
Web Design | | jmueller0 -
Auto-Generated META tag on 404 Page
I'm currently creating a 404 error page for my site and I noticed that a similar site uses some sort of code to automatically generate a meta title. Is this useful? For instance type in electrolux.com/john This page does not exist but in the title you'll see John | Electrolux How can i do this on my site?
Web Design | | evacuumstoreSEO0 -
Best way of conserving link juice from non important pages
If I have a bunch of non important pages on my website which are of little use in the SE's index - IE contact us pages, pages which are near duplicate and conflict with KW's targetting other pages etc, what is the best way of retaining the link juice that would normally be passed to these pages? Most recent discussion I have read has said that with nofollow you effectively just loose link juice, as opposed to conserving it, so that doesn't seem a great option. If I do "noindex" on these pages, would that conserve the link juice in the site, or again would it be just lost? It seems quite a tricky situation as many pages are legitimate for customer usability, but are not worth having in the SE's index and you better off consolidating link juice - so it seems you are getting penilised for making something "for users". Thanks
Web Design | | James770 -
Could Website redesign be a cause of drop in rankings?
We had a complete redesign of our website and moved it over to wordpress several months ago. As url's changed, we had appropriate 301 redirects done. Rankings for our top keywords dropped, but others remained intact. Our SEO company told us rankings drop when a redesign is done, but I thought if we did all redirects properly (which they approved), it wouldn't be much of a problem. Additionally, we've been steadily adding good new content. Any advice?
Web Design | | rdreich490