Error reports showing pages that don't exist on website
-
I have a website that is showing lots of errors (pages that cannot be found) in google webmaster tools. I went through the errors and re-directed the pages I could. There are a bunch of remaining pages that are not really pages this is why they are showing errors. What's strange is some of the URL's are showing feeds which these were never created. I went into Google webmaster tools and looked at the remove URL tool. I am using this but I am confused if I need to be selecting "remove page from search results and cache" option or should I be selecting this other option "remove directory" I am confused on the directory. I don't want to accidentally delete core pages of the site from the search engines.
Can anybody shed some light on this or recommend which I should be selecting?
Thank you
Wendy
-
I would avoid using the "remove URL" option in GWT. The 301s are more ideal in my opinion because let's say I have that old URL posted on my website somewhere, and now it's going to a 404 page. When you redirect it, people will be taken to a different page, and you don't have to worry about having me update the old URL on my website. The link will work, it will take you to an active page and can get you some traffic. However, the "remove URL" option won't give you this same benefit.
Here's a helpful link straight from the source on when NOT to use the Remove URL option: https://support.google.com/webmasters/answer/1269119?hl=en
-
Ok this sounds good because it's not exactly duplicate content so it would be better I agree to do the redirect. I downloaded a redirect plugin yesterday that worked pretty good. I noticed that some pages have already redirected people (the tool has an area to see this)
Final question for you, what are your thoughts on that "remove URL" option in Google Webmaster tools? Wasn't sure if that would be better than a 301 redirect for these remaining senseless errors.
Just curious on your thoughts.
Thank you
-
Rel=canonical is used more when you have duplicate content. If you have the same post or page in two areas, you can use the rel=canonical tag to tell Google where the original of the duplicate is. It sounds like you don't need rel=canonical in this situation.
It sounds like you have 80-something 404 Page Not Found errors. I would use the "Redirection" plugin with Wordpress. Take each URL that is giving you the 404 error in your report, and redirect each one to the most relevant page associated with what was supposed to be on the page that is giving the 404 error. If there really is no relevant page at all, I would just redirect it to the homepage. In my opinion, it's better to have it redirect to the homepage than to have the user land on a 404 page. I would do that for every 404 error you are getting. Doing this, I don't think you'll need rel=canonical at all.
-
Thank you for your reply. You are correct the site is in wordpress. The long story of this whole situation is this....I had initially built the client a wordpress site, things were fine with, traffic and business was good for him. Then one day one of his employees suggested that her father build him a new site that was more graphically pleasing (rather than saying can you please update the graphics on the current site) so the father built an entirely new site on Joomla (I didn't find this out until he was launching this new Joomla site) The guy also went and changed the domain to a www. the original site I had built had no www on it. Fast Forward.....I have re-built the site in wordpress, went back to the non www version. The errors I am getting I have 301 re-directed where I could. I have also in webmaster tools changed the site settings to the preferred domain.(the new site) I fetched all the new pages in Google. I have submitted new site maps. I'm down to 82 errors. The errors are showing pages that do not exist and to re-direct those pages I don't have pages that would really make sense to re-direct them to. I'm wondering now if I'm to a point where I need to "remove the URL's" as offered in Google webmaster tools...??? What do you think on this?
As for the rel-canonical....I understand why I would use these. I see in the Yoast plug in where I can insert the rel-canonical. My question is this:Do I insert the rel-canonical on the page that is correct? So I go to my correct website (the non www one) and go to that page that is correct and I prefer the engines index that page and insert the rel-canonical on that page that is the preferred one? Or am I to go to the non preferred page and insert the rel-canonical on that page so when search engines see that wrong page they see the rel-canonical tag showing them the correct page to index? I looked at a video by Matt Cutts and I wasn't clear on which page I put the rel-canonical on (old site or new one?) I don't have access to the old site so this is why I was thinking maybe just "remove the URL's" as offered up in webmaster tools.
Your input? I really appreciate your help. Thank you
-
For me personally, on Wordpress I use the Yoast SEO tool and I went through the tutorial on the Yoast website. He shows you how to eliminate a lot of the duplicate content that automatically gets created with all Wordpress websites. Once you noindex and get rid of all the unnecessary archives and all that, at that point I would recommend going back to the error report and see the difference and see if those pages keep coming up. If they do, just simply 301 redirect them to another page on your website. Then check again after you redirect them and see what you're left with. Sometimes it takes a couple weeks to reflect from what I've seen. Not sure if this is the exact issue you're having, or if you're even using Wordpress at all, but it sounds like if you are this might help you as it helped me get my errors down to zero.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Website can't break into Google Top100 for main keywords, considering 301 Redirect to a new domain
A little background on our case. Our website, ex: http://ourwebsite.com was officially live in December 2015 but it wasn't On-Site optimized and we haven't done any Off-site SEO to it. In April we decided to do a small redesign and we did it an online development server. Unfortunately, the developers didn't disallow crawlers and the website got indexed while we were developing it on the development server. The development version that got indexed in Google was http://dev.web.com/ourwebsite We learned that it got indexed when we migrated the new redesigned website to the initial domain. When we did the migration we decided to add www and now it looks like: http://www.ourwebsite.com Meanwhile, we deleted the development version from the development server and submitted "Remove outdated content" from the development server's Search Console. This was back in early May. It took about 15-20 days for the development version to get de-indexed and around 30 days for the original website (http://www.ourwebsite.com) to get indexed. Since then we have started our SEO campaign with Press Releases, Outreach to bloggers for Guest and Sponsored Posts etc. The website currently has 55 Backlinks from 44 Referring domains (ahrefs: UR25, DR37) moz DA:6 PA:1 with various anchor text. We are tracking our main keywords and our brand keyword in the SERPs and for our brand keyword we are position #10 in Google, but for the rest of the main (money) keywords we are not in the Top 100 results in Google. It is very frustrating to see no movement in the rankings for the past couple of months and our bosses are demanding rankings and traffic. We are currently exploring the option of using another similar domain of ours and doing a complete 301 Redirect from the original http://www.ourwebsite.com to http://www.ournewebsite.com Does this sound like a good option to you? If we do the 301 Redirect, will the link-juice be passed from the backlinks that we already have from the referring domains to the new domain? Or because the site seems "stuck," would it not pass any power to the new domain? Also, please share any other suggestions that we might use to at least break into the Top 100 results in Google? Thanks.
Intermediate & Advanced SEO | | DanielGorsky0 -
Why isn't the rel=canonical tag working?
My client and I have a problem: An ecommerce store with around 20 000 products has nearly 1 000 000 pages indexed (according to Search Console). I frequently get notified by messages saying “High number of URLs found” in search console. It lists a lot of sample urls with filter and parameters that are indexed by google, for example: https://www.gsport.no/barn-junior/tilbehor/hansker-votter/junior?stoerrelse-324=10-11-aar+10-aar+6-aar+12-aar+4-5-aar+8-9-aar&egenskaper-368=vindtett+vanntett&type-365=hansker&bruksomraade-367=fritid+alpint&dir=asc&order=name If you check the source code, there’s a canonical tag telling the crawler to ignore (..or technically commanding it to regard this exact page as another version of the page without all the parameters) everything after the “?” Does this url showing up in the Search Console message mean that this canonical isn’t working properly? If so: what’s wrong with it? Regards,
Intermediate & Advanced SEO | | Inevo
Sigurd0 -
When I type link:mydomainname.com in Google I don't see any result, why?
My other website is 4 years old and Page Rank 3. We are into business of design and development for 5 years and still we don't have any result from Google Searches. When I type link:mydomainname.com I don't get any result. What's the reason?
Intermediate & Advanced SEO | | vikaspooja1 -
Responsive design (Showing diffrent pages(icons) for Mobile/Tablet users)
I'm writing this question just to insure that we are implementing the responsive design correctly.Example of pages: http://www.yamsafer.me/en/united-arab-emirates/abu-dhabi/hotel/beach-rotana-abu-dhabiAnother : http://www.yamsafer.me/enCan we show different pages(Enhanced for mobile users) to mobile/Tablets visitors (sure same content) but with new icons that enhance the User experience for mobile/tablet users , while hiding these items to PC, laptop users?.Thanks
Intermediate & Advanced SEO | | Yamsafer.com0 -
Indexing a several millions pages new website
Hello everyone, I am currently working for a huge classified website who will be released in France in September 2013. The website will have up to 10 millions pages. I know the indexing of a website of such size should be done step by step and not in only one time to avoid a long sandbox risk and to have more control about it. Do you guys have any recommandations or good practices for such a task ? Maybe some personal experience you might have had ? The website will cover about 300 jobs : In all region (= 300 * 22 pages) In all departments (= 300 * 101 pages) In all cities (= 300 * 37 000 pages) Do you think it would be wiser to index couple of jobs by couple of jobs (for instance 10 jobs every week) or to index with levels of pages (for exemple, 1st step with jobs in region, 2nd step with jobs in departements, etc.) ? More generally speaking, how would you do in order to avoid penalties from Google and to index the whole site as fast as possible ? One more specification : we'll rely on a (big ?) press followup and on a linking job that still has to be determined yet. Thanks for your help ! Best Regards, Raphael
Intermediate & Advanced SEO | | Pureshore0 -
How to get around Google Removal tool not removing redirected and 404 pages? Or if you don't know the anchor text?
Hello! I can’t get squat for an answer in GWT forums. Should have brought this problem here first… The Google Removal Tool doesn't work when the original page you're trying to get recached redirects to another site. Google still reads the site as being okay, so there is no way for me to get the cache reset since I don't what text was previously on the page. For example: This: | http://0creditbalancetransfer.com/article375451_influencial_search_results_for_.htm | Redirects to this: http://abacusmortgageloans.com/GuaranteedPersonaLoanCKBK.htm?hop=duc01996 I don't even know what was on the first page. And when it redirects, I have no way of telling Google to recache the page. It's almost as if the site got deindexed, and they put in a redirect. Then there is crap like this: http://aniga.x90x.net/index.php?q=Recuperacion+Discos+Fujitsu+www.articulo.org/articulo/182/recuperacion_de_disco_duro_recuperar_datos_discos_duros_ii.html No links to my site are on there, yet Google's indexed links say that the page is linking to me. It isn't, but because I don't know HOW the page changed text-wise, I can't get the page recached. The tool also doesn't work when a page 404s. Google still reads the page as being active, but it isn't. What are my options? I literally have hundreds of such URLs. Thanks!
Intermediate & Advanced SEO | | SeanGodier0 -
Do Outbound NoFollow Links Reduce the Page's Ability to Pass PageRank?
I get the recent change where adding a nofollow to one link wont increase the juice passed to other links. I'm wondering if nofollow still passes link-juice into the void. i.e. if a page has $10 of link-juice and has one link then regardless of whether this link is follow or nofollow will the page will leak the same juice? Specifically, Is this site benefitting from having a nofollow on the links in it's car buyer's checklist: http://www.trademe.co.nz/motors/used-cars/mitsubishi/diamante/auction-480341592.htm
Intermediate & Advanced SEO | | seomoz8steer0 -
Most Painless way of getting Duff Pages out of SE's Index
Hi, I've had a few issues that have been caused by our developers on our website. Basically we have a pretty complex method of automatically generating URL's and web pages on our website, and they have stuffed up the URL's at some point and managed to get 10's of thousands of duff URL's and pages indexed by the search engines. I've now got to get these pages out of the SE's indexes as painlessly as possible as I think they are causing a Panda penalty. All these URL's have an addition directory level in them called "home" which should not be there, so I have: www.mysite.com/home/page123 instead of the correct URL www.mysite.com/page123 All these are totally duff URL's with no links going to them, so I'm gaining nothing by 301 redirects, so I was wondering if there was a more painless less risky way of getting them all out the indexes (IE after the stuff up by our developers in the first place I'm wary of letting them loose on 301 redirects incase they cause another issue!) Thanks
Intermediate & Advanced SEO | | James770