Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How get rid of duplicate content, titles, etc on php cartweaver site?
-
my website http://www.bartramgallery.com was created using php and cartweaver 2.0 about five years ago by a web developer. I was really happy with the results of the design was inspired to get into web development and have been studying ever since. My biggest problem at this time is that I am not knowledgable with php and the cartweaver product but am learning as I read more. The issue is that seomoz tools are reporting tons of duplicate content and duplicate title pages etc. This is likely from the dynamic urls and same pages with secondary results etc. I just made a new sitemap with auditmypc I think it was called in an attempt to get rid of all the duplicate page titles but is that going to solve anything or do I need to find another way to configure the site? There are many pages with the same content competing for page rank and it is a bit frustrating to say the least. If anyone has any advice it would be greatly appreciated even pointing me in the right direction.
Thank you,
Jesse
-
I am still researching a bunch of sites trying to figure out a way to get the product ID name at the end which would be great as that is the page title. I just thought I would mention that I am working on it and see if you thought that it was not possible as you mentioned due to Cartweaver's limitations. It's funny that I have spent so much time trying to get my urls to show up how they should... seems this could have been configured into the original product. Beggars can't be choosers.
-
Yes I am going to take a look at that when I get home perhaps I have to change how a few things are referenced as well as create the change of address right? because if you type in the normal dynamic nasty url it still goes to the nasty url but if I select the url and paste it.. it brings up the page as I mentioned above. Basically stripped of images and styling.
I am wondering if it is possible to include that number at the end as it is the actual image and could potentially populate the title of the image at the end which would be sweet. Of course then I would have a new problem of too long of an url as I have the titles pretty keyword rich on a lot of them to make a proper title for the page.
If this all works out I have to create a link to your site at cartweaver and from a couple of my sites as you have been a great help and from what I can tell have been able to properly diagnose a fairly complex issue with php and cartweaver and even if some web page that I have not seen something similar enough you have been a great help. Thank you
-
I'm guessing the paths used to reference the images & css files are relative to the the results.php file.. now that there are "/"s the best thing to do is to change the template to either hard code an absolute path or use forward slash at the start to always start at the root.. eg
Old code:
New code
or
-
I tried the test example you did above and it was pretty cool. With the web address http://www.bartramgallery.com/photographer/charles-cramer/10.php it rendered a page with I believe everything except for design and styles as well as any imagery. Not sure what causes that to occur perhaps it is missing something but that was a pretty quick stab at fixing my url issue. I am too tired now and need to go to bed haha. Thanks
-
No worries Look forward to seeing the site with the new URLs in place - a lot of great photos on that site that need to be shared with everyone
-
Yes it appears that this is a pretty good task to clean up this url issue but well worth it. I was surprised by the system moderators of Cartweaver discounting the url as if it were not important because they are very good developers however I think that the url is much more important than some realize as they are both keyword rich and more interesting to the customer. I am even less likely to click on some random url that has no meaning then if I saw one that clearly spelled out what the page was about.. Thanks Woj I am humbled and realize I have some studying to do.
-
There are 2 issues here:
-
Need to fix the URLs for better user experience & search engines and can do so by using rewrite rules in htaccess
The one suggested by the support forum (I've modified to better match your site but it's untested):
RewriteEngine on
RewriteRule ^photographer/([a-zA-Z0-9_-]+)/([0-9]+).php$ results.php?category=$2 The URLs would then be:
http://www.bartramgallery.com/photographer/charles-cramer/10.php (not ideal with "/10.php" at the end but may be best given the limitations of the cart)
rewrites to: http://www.bartramgallery.com/results.php?category=10 -
Clean up the Google index (remove old URLs & add new ones)
Since both URLs will render the same content we can fix by adding a
tag - attributing 1 source to the duplicate content - check if you can do this dynamically in the templates but be very careful not to canonical everything to the homepage or all your pages will be wiped out the index except the home page!)
-
-
when I read it it seems that the .htaccess was the way to go in that you can have the links appear to google as the old links but in presentation to the customer and keywords the new url would be used. The only thing I was confused about was that it seemed that it would not be good to do redirects but rewrites rather... or is it saying to do both?
-
Thanks
-
Great answer Woj!
-
My pleasure If you set up redirects, you shouldn't loose any traffic
This can also be controlled via htaccess
In google, search for this "site:bartramgallery.com" (without the double quotes) & you will see all the pages you need to redirect
I see the Charles Cramer page as the first photographers page that comes up & the redirect would be something as simple as:
Redirect 301 /results.php?category=10 http://www.bartramgallery.com/charles-cramer
-
Thank you Woj for taking the time to look at my site and I like that organization method. I was not aware of the possibility of being able to reorganize my site like that. I will definately have to research and study a bit to be able to approach this and for awhile I will probably lose traffic but in the end after the changes it should be a much better foot going forward.
-
I'm not familiar with Cartweaver but these are just guides..
First define an organised URL structure - on bartramgallery.com, at a quick glance, a good one could be:
-
bartramgallery.com/photographer (e.g. bartramgallery.com/gordon-michael)
-
bartramgallery.com/photographer/photo (e.g. bartramgallery.com/gordon-michael/juniper-study-joshua-tree)
OR
bartramgallery.com/landscape-photography/photo (e.g. bartramgallery.com/landscape-photography/juniper-study-joshua-tree)
Keep in mind that the shorter URLs the better (could even have bartramgallery.com/photography/juniper-study-joshua-tree)
Second, rewrite the URLs using Rewrite Rules in the htaccess file (see this post: http://www.seomoz.org/blog/rewriterule-split-personality-explained)
I did a search on the Cartweaver support forums and found this:
http://forums.cartweaver.com/topic/google-analytics-identifying-products-and-categoriesOli, from the Cartweaver Support Team, seems to suggest the same "untested" approach as above
Let me know if you need any further help
-
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
SEM Rush & Duplicate content
Hi SEMRush is flagging these pages as having duplicate content, but we have rel = next etc implemented: https://www.key.co.uk/en/key/brand/bott https://www.key.co.uk/en/key/brand/bott?page=2 Or is it being flagged as they're just really similar pages?
Intermediate & Advanced SEO | | BeckyKey0 -
Directory with Duplicate content? what to do?
Moz keeps finding loads of pages with duplicate content on my website. The problem is its a directory page to different locations. E.g if we were a clothes shop we would be listing our locations: www.sitename.com/locations/london www.sitename.com/locations/rome www.sitename.com/locations/germany The content on these pages is all the same, except for an embedded google map that shows the location of the place. The problem is that google thinks all these pages are duplicated content. Should i set a canonical link on every single page saying that www.sitename.com/locations/london is the main page? I don't know if i can use canonical links because the page content isn't identical because of the embedded map. Help would be appreciated. Thanks.
Intermediate & Advanced SEO | | nchlondon0 -
Duplicate Titles caused by multiple variations of same URL
Hi. Can you please advise how I can overcome this issue. Moz.com crawle is indicating I have 100's of Duplicate Title tag errors. However this is caused because many URL's have been indexed multiple times in Google. For example. www.abc.com
Intermediate & Advanced SEO | | adhunna
www.abc.com/?b=123 www.abc.com/ www.abc.com/?b=654 www.abc.com/?b=875 www.abc.com/index.html What can I do to stop this issue being reported as duplictae Titles, as well as content? I was thinking maybe I can use Robots.txt to block various query string parameters. I'm Open to ideas and examples.0 -
International SEO - cannibalisation and duplicate content
Hello all, I look after (in house) 3 domains for one niche travel business across three TLDs: .com .com.au and co.uk and a fourth domain on a co.nz TLD which was recently removed from Googles index. Symptoms: For the past 12 months we have been experiencing canibalisation in the SERPs (namely .com.au being rendered in .com) and Panda related ranking devaluations between our .com site and com.au site. Around 12 months ago the .com TLD was hit hard (80% drop in target KWs) by Panda (probably) and we began to action the below changes. Around 6 weeks ago our .com TLD saw big overnight increases in rankings (to date a 70% averaged increase). However, almost to the same percentage we saw in the .com TLD we suffered significant drops in our .com.au rankings. Basically Google seemed to switch its attention from .com TLD to the .com.au TLD. Note: Each TLD is over 6 years old, we've never proactively gone after links (Penguin) and have always aimed for quality in an often spammy industry. **Have done: ** Adding HREF LANG markup to all pages on all domain Each TLD uses local vernacular e.g for the .com site is American Each TLD has pricing in the regional currency Each TLD has details of the respective local offices, the copy references the lacation, we have significant press coverage in each country like The Guardian for our .co.uk site and Sydney Morning Herlad for our Australia site Targeting each site to its respective market in WMT Each TLDs core-pages (within 3 clicks of the primary nav) are 100% unique We're continuing to re-write and publish unique content to each TLD on a weekly basis As the .co.nz site drove such little traffic re-wrting we added no-idex and the TLD has almost compelte dissapread (16% of pages remain) from the SERPs. XML sitemaps Google + profile for each TLD **Have not done: ** Hosted each TLD on a local server Around 600 pages per TLD are duplicated across all TLDs (roughly 50% of all content). These are way down the IA but still duplicated. Images/video sources from local servers Added address and contact details using SCHEMA markup Any help, advice or just validation on this subject would be appreciated! Kian
Intermediate & Advanced SEO | | team_tic1 -
Artist Bios on Multiple Pages: Duplicate Content or not?
I am currently working on an eComm site for a company that sells art prints. On each print's page, there is a bio about the artist followed by a couple of paragraphs about the print. My concern is that some artists have hundreds of prints on this site, and the bio is reprinted on every page,which makes sense from a usability standpoint, but I am concerned that it will trigger a duplicate content penalty from Google. Some people are trying to convince me that Google won't penalize for this content, since the intent is not to game the SERPs. However, I'm not confident that this isn't being penalized already, or that it won't be in the near future. Because it is just a section of text that is duplicated, but the rest of the text on each page is original, I can't use the rel=canonical tag. I've thought about putting each artist bio into a graphic, but that is a huge undertaking, and not the most elegant solution. Could I put the bio on a separate page with only the artist's info and then place that data on each print page using an <iframe>and then put a noindex,nofollow in the robots.txt file?</p> <p>Is there a better solution? Is this effort even necessary?</p> <p>Thoughts?</p></iframe>
Intermediate & Advanced SEO | | sbaylor0 -
PDF for link building - avoiding duplicate content
Hello, We've got an article that we're turning into a PDF. Both the article and the PDF will be on our site. This PDF is a good, thorough piece of content on how to choose a product. We're going to strip out all of the links to our in the article and create this PDF so that it will be good for people to reference and even print. Then we're going to do link building through outreach since people will find the article and PDF useful. My question is, how do I use rel="canonical" to make sure that the article and PDF aren't duplicate content? Thanks.
Intermediate & Advanced SEO | | BobGW0 -
Duplicate content on subdomains.
Hi Mozer's, I have a site www.xyz.com and also geo targeted sub domains www.uk.xyz.com, www.india.xyz.com and so on. All the sub domains have the content which is same as the content on the main domain that is www.xyz.com. So, I want to know how can i avoid content duplication. Many Thanks!
Intermediate & Advanced SEO | | HiteshBharucha0 -
Can PDF be seen as duplicate content? If so, how to prevent it?
I see no reason why PDF couldn't be considered duplicate content but I haven't seen any threads about it. We publish loads of product documentation provided by manufacturers as well as White Papers and Case Studies. These give our customers and prospects a better idea off our solutions and help them along their buying process. However, I'm not sure if it would be better to make them non-indexable to prevent duplicate content issues. Clearly we would prefer a solutions where we benefit from to keywords in the documents. Any one has insight on how to deal with PDF provided by third parties? Thanks in advance.
Intermediate & Advanced SEO | | Gestisoft-Qc1