How long does it take before URL's are removed from Google?
-
Hello,
I recently changed our websites url structures removing the .html at the end. I had about 55 301's setup from the old url to the new. Within a day all the new URL's were listed in Google, but the old .html ones still have not been removed a week later.
Is there something I am missing? Or will it just take time for them to get de-indexed?
As well, so far the Page Authority hasn't transfered from the old pages to the new, is this typical?
Thanks!
-
I assume the new url format / structure is also the new inner link structure (all links in the site are updated with the new format). if this is the case the indexation is based on this not based on old urls following the 301s.
As far as testing them - did you test to see what response code do you get when accessing the old urls ?
Though I don't understand why you'd submit a sitemap with the old urls?
** To send google's bot to crawl those and see the 301s and delist them from the index.
How do you ping it?
** There are a lot of free services available that you can use - just run a search and you will have several to choose from.
-
Thanks for your response! I would assume the 301 are setup correctly if Google is indexing them and of course they work when I test them.
Though I don't understand why you'd submit a sitemap with the old urls? How do you ping it?
Any thoughts on the Page Authority?
-
Hi Sean,
For this small number of urls you can help Google's bot to dexindex those by having a separate xml sitemap in your web master tools with the old urls only - submit the sitemap and ping it. This will help speep up the process a little bit.
However since the number of urls is small - 55 - google will delist those old urls based on the 301 redirect (if the 301 setup is correct) at the next crawl cycle for your website.
You can also check Web master tool crawl rates to be able to make some prediction on how fast and often google is "visiting" your site.
There is no fix term for this. I have a large account with several million pages in the index, with 410 set on several thousands pages that were removed and those are still in the index after 4 months - it 's related with the size of the website, the crawl rate, freshness of those pages ...
Hope it helps.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Potential downside of removing past years' calendar events
hi there. my website is for a school. We have several calendars for athletic events, school events, etc. there are thousands of events each year that link to pages for each event. the URLs/pages are generated dynamically, and the urls and pages are very similar from year to year, so we're being penalized for duplicate content. I can delete past years events in bulk, but there's no way to redirect them in bulk. am i taking a big chance by deleting events that occurred prior to 1/1/2019?
Intermediate & Advanced SEO | | BGR0 -
Best way to permanently remove URLs from the Google index?
We have several subdomains we use for testing applications. Even if we block with robots.txt, these subdomains still appear to get indexed (though they show as blocked by robots.txt. I've claimed these subdomains and requested permanent removal, but it appears that after a certain time period (6 months)? Google will re-index (and mark them as blocked by robots.txt). What is the best way to permanently remove these from the index? We can't use login to block because our clients want to be able to view these applications without needing to login. What is the next best solution?
Intermediate & Advanced SEO | | nicole.healthline0 -
What's the best SEO practice for having dynamic content on the same URL?
Let's use this example... www.miniclip.com and there's a function to log in... If you're logged in and a cookie checks that you're logged in and you're on page, let's say, www.miniclip.com/racing-games however the banners being displayed would have more call to action and offers on the page when a user is not logged in to entice them to sign up but the URL would still be www.miniclip.com/racing-games if and if not logged in, what would be the best URL practice for this? just do it?
Intermediate & Advanced SEO | | AdiRste0 -
Multiple URL's exist for the same page, canonicaliazation issue?
All of the following URL's take me to the same page on my site: 1. www.mysite.com/category1/subcategory.aspx 2. www.mysite.com/subcategory.aspx 3. www.mysite.com/category1/category1/category1/subcategory.aspx All of those pages are canonicalized to #1, so is that okay? I was told the following my a company trying to make our sitemap: "the site's platform dynamically creates URLs that resolve as 200 and should be 404. This is a huge spider trap for any search engine and will make them wary of crawling the site." What would I need to do to fix this? Thanks!
Intermediate & Advanced SEO | | pbhatt0 -
How are pages ranked when using Google's "site:" operator?
Hi, If you perform a Google search like site:seomoz.org, how are the pages displayed sorted/ranked? Thanks!
Intermediate & Advanced SEO | | anthematic0 -
Charity project for local women's shelter - need help: will Google notice if you alter the document title with Javascript after the page loads?
I am doing some pro-bono work with a local shelter for female victims of domestic abuse. I am trying to help visitors to the site cover their tracks by employing a document.title change when the page loads using JavaScript. This shelter receives a lot of traffic from Google. I worry that the Google bots will see this javascript change and somehow penalize this site or modify the title in the SERPs. Has anyone had any experience with this kind of javascript maneuver? All help would be greatly appreciated!
Intermediate & Advanced SEO | | jkonowitch0 -
Best solution to get mass URl's out the SE's index
Hi, I've got an issue where our web developers have made a mistake on our website by messing up some URL's . Because our site works dynamically IE the URL's generated on a page are relevant to the current URL it ment the problem URL linked out to more problem URL's - effectively replicating an entire website directory under problem URL's - this has caused tens of thousands of URL's in SE's indexes which shouldn't be there. So say for example the problem URL's are like www.mysite.com/incorrect-directory/folder1/page1/ It seems I can correct this by doing the following: 1/. Use Robots.txt to disallow access to /incorrect-directory/* 2/. 301 the urls like this:
Intermediate & Advanced SEO | | James77
www.mysite.com/incorrect-directory/folder1/page1/
301 to:
www.mysite.com/correct-directory/folder1/page1/ 3/. 301 URL's to the root correct directory like this:
www.mysite.com/incorrect-directory/folder1/page1/
www.mysite.com/incorrect-directory/folder1/page2/
www.mysite.com/incorrect-directory/folder2/ 301 to:
www.mysite.com/correct-directory/ Which method do you think is the best solution? - I doubt there is any link juice benifit from 301'ing URL's as there shouldn't be any external links pointing to the wrong URL's.0 -
What's your best hidden SEO secret?
Don't take that question too serious but all answers are welcome 😉 Answer to all:
Intermediate & Advanced SEO | | petrakraft
"Gentlemen, I see you did you best - at least I hope so! But after all I suppose I am stuck here to go on reading the SEOmoz blog if I can't sqeeze more secrets from you!9