Getting Rid of Duplicate Page Titles After URL Structure Change
-
I've had all sorts of issues with google when they just dropped us on our head a few weeks ago. Google is crawling again after I made some changes, but they're still not ranking our content like they were so I have a few questions.
I changed our url structure from /year/month/date/post-title to just /post-title and 301 redirected the old link structure to the new. When I look I see over 3000 duplicate title errors listing both versions of the url.
1. How do I get google to crawl the old url structure and recognize the 301 redirect and update the index?
2. Google is crawling the site again, but they're not ranking us like they were before. We're in a highly competitive category and I'm aware of that, but we've always been an authority in our niche. We have plenty of quality backlinks and often we're originators of the content which is then rewritten by a trillion websites everywhere. We're not the best at writing and titles, but we're working on it and this did not matter much to google previously as it was ranking us pretty highly on the front page and certainly ranking us over many sites that are ranking above us today.
Some backlinks http://www.alexa.com/site/linksin/dajaz1.com
A few examples - if you google twista gucci louis prada you'll see many of the sites who trackbacked to us since we premiered the song rank much higher than us. 3 weeks ago we were ranking above them.
http://dajaz1.com/twista-gucci-louis-prada/
google search jadakiss consignment mixtape 3 weeks ago we were ranking higher than all 4 sites ranking above us. The sites ranking above us even link to us or mention us, yet they rank above us now.
original content here
http://dajaz1.com/watch-jadakiss-confirms-cosignment-mixtape-2012-schedule/
I could throw out a ton of examples like this. How do we get google to rank us again.
It should be noted that I'm not using any SEO plugin's on the site. I hand coded what's in there, and I know I can probably do it better so any tips or ideas is welcome. I'm pretty sure that our issues were caused by the Yoast SEO Plugin as when I search site:dajaz1.com the pages and topics that display were all indexed while the plugin was active. I've since removed it and all calls to it in the database, but I'm pretty nervous about plugins right now. Which brings me to my third and final question
How do I get rid of the page category and topic pages that were indexed and seem to be ranking higher than the rest of our content?
I lied one more. For category url I've set it to remove the category base so the url is dajaz1.com/news or dajaz1.com/music
is that preferable or is this causing me issues?
Any feedback is appreciated.
Also google is crawling again (see attached image) but the Kilobytes downloaded per day hasn't. Should I be concerned about this?
-
Thanks so much for your response.
Honestly I'm still not sure why or what happened.
Essentially the domain has a bit of history. It's been up since 2008 and is very well known authority figure in the niche. The site was taken down by the Govt in their domain seizure crap and held for 13 months. They had to give it back because they had no probable cause, and we were innocent. It's been a big issue in the media, especially when all the SOPA madness was going on but not sure if that 13 month disruption caused any issues for us.
When we came back up we were doing really well, ranking very high and doing about 5k visitors a day from google. We had our highest google traffic day on March 4th and then for some reason we did a massive free fall. Google just left. When I searched site:dajaz1.com it showed paginated pages, category, tags, search pages ranking, it stopped ranking our post pages.
I started going through the site to try and figure out what was wrong. I set everything that's not the home page or the actual post as noindex, follow, we started working on writing more, better titles, better seo urls, removed the pagination plugins etc. We're often the originators of content and sourced as such so it's not necessarily an original content issue.
Google bot came back, but it's still not ranking us like it was and now ranking people above us who link to us, source us, or copy our content. When I check site:dajaz1.com now it's at least pulling posts, but it's pulling posts with numbers in the url, or very old urls. Our newer posts are not ranking. We were getting 5k-8k unique visitors a day organically, now we're getting less than 1000. I'm still at a loss as to why and how to correct.
-
Hi Melissa,
Let me try to answer your question, with the caveat that I'm missing part of the back story. Specifically, it sounds like Google wasn't crawling your site for awhile - any idea why?
If Google is indexing duplicate pages for much of your content, then in all likelihood you won't reach your ranking potential for that content. It sounds like this is because you changed your URL structure, and Google has crawled the new URLs without crawling/de-indexing the old URLs. Does this sound about right?
One solution, and probably your best bet, it to maintain a sitemap of the old URLs. This way, Google is more likely to crawl those URLs, discover the 301 redirect, and de-index those pages much faster. Ideally this is setup before the change in URL structure. At this point, it may be a major pain... if you catch my drift.
(there are a couple ways you could accomplish this, both require a lot of work. You could rewrite the URLs using Excel magic, or write a script to generate an old sitemap. Or you could very briefly switch back to your old url structure, generate a new sitemap to save, and then switch back)
Another method would be to use the URL removal tools in Google Webmaster tools. If your old URLs all used a unique directory structure (that differed from your new directory structure) you can remove entire directories from Google's index with the click of a button.
Regardless, whatever way you choose to remove those URLs, doing so should help improve your rankings.
Hope this helps! Best of luck with your SEO.
-
This is what I have
The reason I have it set to noindex, follow for category is because it's ranking my category pages over my actual blog posts. I wouldn't have an issue with having the categories indexed, but I do have an issue with posts going no where and my category pages showing up in search results.
I keep getting conflicting information on this so I'm not sure what I should with that.
What concerns me is that it seems to be crawling everything, but it's not dumping the old url structure for the new. I realize that it loses a small amount of link juice on the old urls, but on the newer urls it shouldn't and those are ranking lower than were. I can't figure out for the life of me why.
-
It's known that when you 301 a page, you loose some of the power it once had. That being said, if you're basically 301'ing your whole site, Google will have to reevaluate everything.
For your category pages, add the robots tag and Google will not index them.
if you're using wordpress, here's a quick snippet.
if (is_category()) {
//Allow indexation of category URLs
echo ‘’. “\n”; } elseif (is_404() || is_search() || is_author() || is_archive()) {
//Do not allow indexation and crawling of 404, search, author and archived pages
echo ‘’. “\n”; } else {
//For other URLs, not mentioned above such as post URLs, Page URLs - allow indexation and crawling.
echo ‘’. “\n”; } ?>
-
We are also doing a URL rewrite soon....I am looking forward to seeing responses. It is probably not relevant but I read on here somewhere about making sure you old site map and new site map or on the page for a while until google starts making the connections.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do I prevent duplicate page title errors from being generated by my multiple shop pages?
Our e-commerce shop has numerous pages within the main shop page. Users navigate through the shop via typical pagination. So while there may be 6 pages of products it's all still under the main shop page. Moz keeps flagging my shop pages as having duplicate titles (ie shop page 2). But they're all the same page. Users aren't loading unique pages each time they go to the next page of products and they aren't pages I can edit. I'm not sure how to prevent this issue from popping up on my reports.
Technical SEO | | NiteSkirm0 -
SEO URLs: 1\. URLs in my language (Greek, Greeklish or English)? 2\. Αt the end it is good to put -> .html? What is the best way to get great ranking?
Hello all, I must put URLs in my language Greek, Greeklish or in English? And at the end of url it is good to put -> .html? For exampe www.test.com/test/test-test.html ? What is the best way to get great ranking? I am a new digital marketing manager and its my first time who works with a programmer who doesn't know. I need to know as soon as possible, because they want to be "on air" tomorrow! Thank you very much for your help! Regards, Marios
Technical SEO | | marioskal0 -
Multi page with title with same keywords
Greetings Team, I have a webiste home page optimize for the keyword ''example apartments'' and i have included in the title but that i have another page called site.com/exampleapartments and is optimize for the keyword example apartments How can i make them different ?
Technical SEO | | WorldEscape0 -
Duplicate pages in Google index despite canonical tag and URL Parameter in GWMT
Good morning Moz... This is a weird one. It seems to be a "bug" with Google, honest... We migrated our site www.three-clearance.co.uk to a Drupal platform over the new year. The old site used URL-based tracking for heat map purposes, so for instance www.three-clearance.co.uk/apple-phones.html ..could be reached via www.three-clearance.co.uk/apple-phones.html?ref=menu or www.three-clearance.co.uk/apple-phones.html?ref=sidebar and so on. GWMT was told of the ref parameter and the canonical meta tag used to indicate our preference. As expected we encountered no duplicate content issues and everything was good. This is the chain of events: Site migrated to new platform following best practice, as far as I can attest to. Only known issue was that the verification for both google analytics (meta tag) and GWMT (HTML file) didn't transfer as expected so between relaunch on the 22nd Dec and the fix on 2nd Jan we have no GA data, and presumably there was a period where GWMT became unverified. URL structure and URIs were maintained 100% (which may be a problem, now) Yesterday I discovered 200-ish 'duplicate meta titles' and 'duplicate meta descriptions' in GWMT. Uh oh, thought I. Expand the report out and the duplicates are in fact ?ref= versions of the same root URL. Double uh oh, thought I. Run, not walk, to google and do some Fu: http://is.gd/yJ3U24 (9 versions of the same page, in the index, the only variation being the ?ref= URI) Checked BING and it has indexed each root URL once, as it should. Situation now: Site no longer uses ?ref= parameter, although of course there still exists some external backlinks that use it. This was intentional and happened when we migrated. I 'reset' the URL parameter in GWMT yesterday, given that there's no "delete" option. The "URLs monitored" count went from 900 to 0, but today is at over 1,000 (another wtf moment) I also resubmitted the XML sitemap and fetched 5 'hub' pages as Google, including the homepage and HTML site-map page. The ?ref= URls in the index have the disadvantage of actually working, given that we transferred the URL structure and of course the webserver just ignores the nonsense arguments and serves the page. So I assume Google assumes the pages still exist, and won't drop them from the index but will instead apply a dupe content penalty. Or maybe call us a spam farm. Who knows. Options that occurred to me (other than maybe making our canonical tags bold or locating a Google bug submission form 😄 ) include A) robots.txt-ing .?ref=. but to me this says "you can't see these pages", not "these pages don't exist", so isn't correct B) Hand-removing the URLs from the index through a page removal request per indexed URL C) Apply 301 to each indexed URL (hello BING dirty sitemap penalty) D) Post on SEOMoz because I genuinely can't understand this. Even if the gap in verification caused GWMT to forget that we had set ?ref= as a URL parameter, the parameter was no longer in use because the verification only went missing when we relaunched the site without this tracking. Google is seemingly 100% ignoring our canonical tags as well as the GWMT URL setting - I have no idea why and can't think of the best way to correct the situation. Do you? 🙂 Edited To Add: As of this morning the "edit/reset" buttons have disappeared from GWMT URL Parameters page, along with the option to add a new one. There's no messages explaining why and of course the Google help page doesn't mention disappearing buttons (it doesn't even explain what 'reset' does, or why there's no 'remove' option).
Technical SEO | | Tinhat0 -
Is Duplicate title made Sanbox?
I use Seomoz tool and discover that my webpage have 1000 duplicate title, my keyword with domain key 's position is 25, but i dont have any keyword ontop 100 of google. Is the Duplicate title effect in the key word position or SEO ? right?
Technical SEO | | magician0 -
According to 1 of my PRO campaigns - I have 250+ pages with Duplicate Content - Could my empty 'tag' pages be to blame?
Like I said, my one of my moz reports is showing 250+ pages with duplicate content. should I just delete the tag pages? Is that worth my time? how do I alert SEOmoz that the changes have been made, so that they show up in my next report?
Technical SEO | | TylerAbernethy0 -
Blogger Blog URL Structure Questions
I'm starting to use my blog more and wanted to ask about an issue I've read about on SEOmoz in the past. I use blogger instead of wordpress. It's quick and simple - I have no interest in switching to wordpress for this particular blog. My blog is currently setup as blog.site.com. Is it still important (for seo reasons) to switch from blog.site.com to site.com/blog? If so, is there a way to do this in blogger? And if I do this, will my past posts lose their authority if their redirected to the new url structure? Rand mentions in this article: http://www.seomoz.org/blog/11-best-practices-for-urls "never use multiple subdomains" - This is an old article, but I've seen this mentioned several times. Does this still hold true? Am I losing out on links to my blog? Thanks in advance.
Technical SEO | | ChaseH0 -
Getting Google to index new pages
I have a site, called SiteB that has 200 pages of new, unique content. I made a table of contents (TOC) page on SiteB that points to about 50 pages of SiteB content. I would like to get SiteB's TOC page crawled and indexed by Google, as well as all the pages it points to. I submitted the TOC to Pingler 24 hours ago and from the logs I see the Googlebot visited the TOC page but it did not crawl any of the 50 pages that are linked to from the TOC. I do not have a robots.txt file on SiteB. There are no robot meta tags (nofollow, noindex). There are no 'rel=nofollow' attributes on the links. Why would Google crawl the TOC (when I Pinglered it) but not crawl any of the links on that page? One other fact, and I don't know if this matters, but SiteB lives on a subdomain and the URLs contain numbers, like this: http://subdomain.domain.com/category/34404 Yes, I know that the number part is suboptimal from an SEO point of view. I'm working on that, too. But first wanted to figure out why Google isn't crawling the TOC. The site is new and so hasn't been penalized by Google. Thanks for any ideas...
Technical SEO | | scanlin0