How do I identify what is causing my Duplicate Page Content problem?
-
Hello,
I'm trying to put my finger on what exactly is causing my duplicate page content problem... For example, SEOMoz is picking up these four pages as having the same content:
http://www.penncare.net/ambulancedivision/braunambulances/express.aspx
http://www.penncare.net/ambulancedivision/recentdeliveries/millcreekparamedicservice.aspx
http://www.penncare.net/ambulancedivision/recentdeliveries/monongaliaems.aspx
http://www.penncare.net/softwaredivision/emschartssoftware/emschartsvideos.aspx
As you can tell, they really aren't serving the same content in the body of the page. Anybody have an idea what might be causing these pages to show up as Duplicate Page Content? At first I thought it was the photo gallery module that might be causing it, but that only exists on two of the pages...
Thanks in advance!
-
Ah right - OK then.
With regards to data coming back from SEOmoz's crawler, I might be tempted to ask them what it is seeing. I should really have a look at this myself because I haven't yet.
-
I'm currently getting that information from Moz's own web crawler wherein it tells me the pages of that have Duplicate Page Content and the other URLs that that duplicate content exists on.
With regard to the 301's - I have rewrite rules setup to 1.) set all requests to lowercase 2.) trim off home.aspx 3.) append www. to the beginning of the request, etc. When processed these should function as a single redirect / rewrite.
-
Before looking at the duplicate content (what did you use to find there is duplicate content?)... a quick question - you have a lot of 301's. Just want to check, are these just a single redirect or a redirect of a redirect etc?
-
I would add some content to these pages to help differentiate. None of them are text heavy so it may be hard for spiders to see a difference. Add a summary, maybe a text translation of what is in the vids, etc
-
Thanks for your reply... I guess more specifically I was wondering what it was about these particular page elements that makes search engines incapable of deciphering them from one another.
-
- Search engines don't know which version(s) to include/exclude from their indices
- Search engines don't know whether to direct the link metrics (trust, authority, anchor text, link juice, etc.) to one page, or keep it separated between multiple versions
- Search engines don't know which version(s) to rank for query results
When duplicate content is present, site owners suffer rankings and traffic losses and search engines provide less relevant results.
Hope this helps!
Resources, http://www.seomoz.org/learn-seo/duplicate-content
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Dead end pages are really an issue?
Hi all, We have many pages which are help guides to our features. These pages do not have anymore outgoing links (internal / external). We haven't linked as these are already 4th level pages and specific about particular topic. So these are technically dead end pages. Do these pages really hurt us? We need to link to some other pages? Thanks
Web Design | | vtmoz0 -
Https pages indexed but all web pages are http - please can you offer some help?
Dear Moz Community, Please could you see what you think and offer some definite steps or advice.. I contacted the host provider and his initial thought was that WordPress was causing the https problem ?: eg when an https version of a page is called, things like videos and media don't always show up. A SSL certificate that is attached to a website, can allow pages to load over https. The host said that there is no active configured SSL it's just waiting as part of the hosting package just in case, but I found that the SSL certificate is still showing up during a crawl.It's important to eliminate the https problem before external backlinks link to any of the unwanted https pages that are currently indexed. Luckily I haven't started any intense backlinking work yet, and any links I have posted in search land have all been http version.I checked a few more url's to see if it’s necessary to create a permanent redirect from https to http. For example, I tried requesting domain.co.uk using the https:// and the https:// page loaded instead of redirecting automatically to http prefix version. I know that if I am automatically redirected to the http:// version of the page, then that is the way it should be. Search engines and visitors will stay on the http version of the site and not get lost anywhere in https. This also helps to eliminate duplicate content and to preserve link juice. What are your thoughts regarding that?As I understand it, most server configurations should redirect by default when https isn’t configured, and from my experience I’ve seen cases where pages requested via https return the default server page, a 404 error, or duplicate content. So I'm confused as to where to take this.One suggestion would be to disable all https since there is no need to have any traces to SSL when the site is even crawled ?. I don't want to enable https in the htaccess only to then create a https to http rewrite rule; https shouldn't even be a crawlable function of the site at all.RewriteEngine OnRewriteCond %{HTTPS} offor to disable the SSL completely for now until it becomes a necessity for the website.I would really welcome your thoughts as I'm really stuck as to what to do for the best, short term and long term.Kind Regards
Web Design | | SEOguy10 -
Are pages not included in navigation given less "weight"
Hi, we recently updated our website and our main navigation was dramatically slimmed down to just three pages and no drop down under those. Yet we have many more important pages, which are linked to once on one of those main three pages. However, will this hurt those other pages because they are not included in navigation (some of which were starting to get good traction in rankings)?
Web Design | | LuaMarketing2
Thanks!0 -
Certain PHP Pages Not Showing In SERPs
Hi all, You've all been so helpful so far, I'm hoping you can help me with our trickiest SEO question yet. Last year we migrated 7 sites into one, going from flat html to Joomla. This went fine and although we saw a slight drop in traffic, it wasn't too bad. Now however traffic has started to drop and we've been able to hone it down into a certain area of our website. Each of the 7 sites had their own page with some php code that was fed products. These products were updated everyday and were are second most popular page on the sites aprt from the home page. These pages were found in Google no problem and were an essential resource for our site. What we have found is these pages cannot be found at all, unless you type the full business name and product. If you just type the product and the location our customer is based in, we're no where, using the Moz tools it says we're not in the top 50 results. This is a bit of a shock since we used to be at least on the first page, usually quite high up. I'm a little stumped as SEO wise these pages are technically better. They offer the same functionality but in a much more SEO friendly way. I've asked our developer to check: Nothing is being blocked in robots.txt
Web Design | | HB17
The pages are being indexed
There's no strange code errors Essentially the pages can't be found even if we type the full title, for example 'customer's products in their town' to be found we have to type 'customer's products in their town and their full business name'. The top third of the page is just html text, the bottom like I mentioned is PHP and is fed data from a database which is refreshed each morning. I know our developer did some rel conical work but has assured me that's all working fine. While I know it's a new website, we've owned the domain for a while so our domain authority isn't brand new and 0, we also have pages with worse page authority that show up on page 1 no problem, so I'm leaning towards something else might not be right. Can anyone help me figure out why these pages are being indexed but not even found? Thanks!0 -
Funnel tracking with one page check-out?
Hi Guys, I'm creating a new website with a one page checkout that follows the following steps:
Web Design | | Jerune
1. Check availability
2. Select product
2. Select additional product & Add features
3. Provide personal information
4. Order & Pay I'm researching if it is possible to track all these steps (and even steps within the steps) with Google Analytics in order to analyse checkout abandonment. The problem is only that my one-page checkout has only one URL (I want to keep it that way) and therefore can not be differentiated on URL in the Analytics funnel. To continue to the next step also the same button (in a floating cart) in used to advance. The buttons to select/choose something within one step are all different. Do you guys know how I can set this up and how detailed I can make this? For example, is it also possible to test at which field visitors leave when for example filling in their personal information? Would be great if you can help me out!0 -
Duplicate H1 tag IF it holds SAME text?
Hello people, I know that majority of SEO gurus (?) claim that H1 tag should only be used once per page. In the landing page design I'm working with, we actually need to repeat our core message stated in H1 & H2 - at the bottom of the page. Now the question is: Can that in any way cause any ranking penalty from big G? In my eyes that is not attempt to over optimize page as it contains SAME info as the H1 & H2 at the top of the page. Confusing, so I'm hope that some SEO gurus here will share some light on this. Thanks in advance!
Web Design | | RetroOnline0 -
What is the best information architecture for developing local seo pages?
I think I have a good handle on the external local seo factors such as citations but I'd like to determine the best IA layout for starting a new site or adding new content to a local site. I see lots of small sites with duplicate content pages for each town/area which I know is poor practice. I also see sites that have unique content for each of those pages but it seems like bad design practice, from a user perspective, to create so many pages just for the search engines. To the example... My remodeling company needs to have some top level pages on its site to help the customers learn about my product, call these "Kitchen Remodeling" and "Bathroom Remodeling" for our purposes. Should I build these pages to be helpful to the customer without worrying too much about the SEO for now and focus on subfolders for my immediate area which would target keywords like "Kitchen Remodeling Mytown"? Aside from my future site, which is not a priority, I would like to be equipped to advise on best practices for the website development in situations where I am involved at the beginning of the process rather than just making the local SEO fit after the fact. Thanks in advance!
Web Design | | EthanB0 -
SEOMoz crawl report shows a duplicate content and duplicate title for these two url's http://freightmonster.com/ and http://freightmonster.com/index.html. How do I fix this?
What page is attached to http://freightmonster.com/ if it is not the index.html ? Should I do a redirect from the index page to something more descriptive?
Web Design | | FreightBoy1