Dealing with Not Found errors
-
Hi,
I have a problem with Google Webmaster, it reported that I have many errors 404 Not Found.
I have detected and found the links are coming from my site, I have fixed them and submit a new sitemap to google again. Then, I wait a few days, google still report errors 404 Not found same as before.I don't know why? Please help!
-
That is why I 301 redirect all of my 404 banks simply because Google has already indexed and archived them as I mentioned. So you need to 301 redirect the links I hope this helps sincerely, Thomas
-
I have marked as fix many times, and then the errors still occur. When I check my sitemap and did not see the not found links which are listed in the tab linked from sitempap.
-
considering that your website has both index and archiving on and Google obviously has indexed and archived your website I would strongly suggest you eliminate this is a obstacle in the future and simply 301 redirected the correct path
http://www.vietnamvisacorp.com/tips/6.html
If using Nginx like I think you are simply implement this in your htaccess
nginx configuration location /tips/6.html { rewrite ^(.*)$ http://www.vietnamvisacorp.com/tips.html redirect; }
I used this tool to accomplish this goal http://winginx.com/htaccess you can do the exact same thing using this and it will help you if you already understand how 301 redirect for Apache
http://www.anilcetin.com/convert-apache-htaccess-to-nginx/
if you want to be certain to get the code right which I would I would use an Apache generator then put it into the Nginx generator and properly 301 redirect your 404
simply use the tool below to create the and copy and paste that redirect into one of the 2 tools I have given you above.
-
Hello John & Dennis,
if it is completely gone and there's no 404 you can check this by going to the Google Webmaster tools panel and instead of selecting ignore or anything like that hit fetch as a Google bot this will tell you if the link is still a real link or not. You do not have to wait a set amount of time to find out if one link is in fact okay.
I would be hard-pressed to imagine that it is worth anything meaning it has back links pointing to it however if it does you will want to use open site Explorer to check the link prior to ignoring it or marking it as fixed.
The rule I use is if it shows up as a 404 I 301 redirect it that way you never lose back links. It's a little more complicated using Nginx but in my opinion it's worth the complication as the improved speed and ability to take incredible amount of traffic make it really worth it.
What Dennis has stated is true if the link is worthless And of course does not exist Anywhere on a webpage.
I hope this has been of help.
Thomas
-
here is the information I retrieved regarding how your website is made website
https://builtwith.com/?http%3A%2F%2Fwww.vietnamvisacorp.com%2Ftips.html
http://www.vietnamvisacorp.com/tips.html
http://wiki.nginx.org/LikeApache-htaccess
so let's say you wanted to make it back into the original URL I have above
to fix it and Apache
Permanent URL redirect - generated by www.rapidtables.com
Redirect 301 /tips/6.html http://www.vietnamvisacorp.com/tips.html
now to fix it in Nginx
Use
or
http://www.anilcetin.com/convert-apache-htaccess-to-nginx/
nginx configuration location /tips/6.html { rewrite ^(.*)$ http://www.vietnamvisacorp.com/tips.html redirect; }
http://www.vbseo.com/f77/converting-htaccess-301-redirect-nginx-49834/
-
Hi John
Mark it as fixed in webmaster tools
After that, just leave it alone for a week or so. Once your site is recrawled, youll see that it's not there anymore (as long as it's really fixed)
-
this is not a problem you simply have to take these links that you're finding is 404's and 301 redirect them to relevant pages on your website. What I mean is do not simply take them and put them all to the home page. Spread them around if they have no direct use. I've noticed you are using Nginx along with Apache where you might be just using Nginx in order to make the 301 redirects correctly with engine acts I'm going to need to grab a couple links for you. I will be right back.
sincerely,
Thomas
-
For an example about this problem:
Google webmmaster report Not found link: www.vietnamvisa...com/tips/6.html (this is link before update).
and in the tab Linked from: http://www.vietnamvisacorp.com/tips.html, but in this page did not include /tips/6.html
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Many of my pages are showing up as 4xx errors because they require a login/subscription to view. How can I fix this?
Should these pages be redirecting to a login page? Or does there need to be something other than an Access Denied message? I feel like these are bringing the overall site score down.
On-Page Optimization | | CaraMc0 -
Suggestions on dealing with duplicate content?
What are the best ways to protect / deal with duplicate content? I've added an example scenario, Nike Trainer model 1 – has an overview page that also links to a sub-page about cushioning, one about Gore-Tex and one about breathability. Nike Trainer model 2,3,4,5 – have an overview page that also links to sub-pages page about cushioning , Gore-Tex and breathability. In each of the sub-pages the URL is a child of the parent so a distinct page from each other e.g. /nike-trainer/model-1/gore-tex /nike-trainer/model-2/gore-tex. There is some differences in material composition, some different images and of course the product name is referred multiple times. This makes the page in the region of 80% unique. Any suggestions welcome about the above example or any other ways you guys know of dealing with duplicate content.
On-Page Optimization | | punchseo0 -
Dynamic URL Parameters + Woocommerce create 404 errors
Hi Guys,
On-Page Optimization | | jeeyer
Our latest Moz crawl shows a lot of 404-errors for pages that create dynamical links for users? (I guess it are dynamic links, not sure). Situation: On a page which shows products from brand X users can use the pagination icons on the bottom, or click on: View: / 24/48/All.
When a user clicks 48 the end of the link will be /?show_products=48
I think there were some pages that could show 48 products but do not exist anymore (because products are sold out for example), and that's why they show 404's and Moz reports them. How do I deal with these 404-errors? I can't set a 301-redirect because it depends on how many products are shown (it changes every time).
Should I just ignore these kind of 404-errors? Or what is the best way to handle this situation?0 -
Strange error on my website
Hi Wondering if anyone could help out here please. I tried to crawl my website using screaming frog and got a 500 error warning (odd considering it worked yesterday). I tried google speed test and got the following message - PageSpeed Insights received a 500 response from the server. To analyze pages behind firewalls or that require authentication. However, it also crawled the site and gave me a score of 70/100. It showed the site thumbnail, as normal. This was updated this morning and it showed the latest one so must have crawled. I can access the site, as can my 'people', the site is getting hits from real users in Google Analytics so must be working. Am very puzzled how the site can be working and not be working at the same time! If possible i would rather not share the url. The site is a test of concept and is very messy looking. it is more a test to see what happens if i do x... Many thanks Carl
On-Page Optimization | | daedriccarl0 -
Video Sitemap Error
Hi I am using a Video snippet and this error was generated about the duration tag. Can someone tell me how to fix it? and will this error prevent snippet to work on the sear results? Thanks Mozzers! 1rC5q7g.png
On-Page Optimization | | Ideas-Money-Art0 -
Dealing with thin content/95% duplicate content - canonical vs 301 vs noindex
My client's got 14 physical locations around the country but has a webpage for each "service area" they operate in. They have a Croydon location. But a separate page for London, Croydon, Essex, Luton, Stevenage and many other places (areas near Croydon) that the Croydon location serves. Each of these pages is a near duplicate of the Croydon page with the word Croydon swapped for the area. I'm told this was a SEO tactic circa 2001. Obviously this is an issue. So the question - should I 301 redirect each of the links to the Croydon page? Or (what I believe to be the best answer) set a rel=canonical tag on the duplicate pages). Creating "real and meaningful content" on each page isn't quite an option, sorry!
On-Page Optimization | | JamesFx0 -
SEOmoz crawl error
Hi, I'm getting a crawl error and it complains about there being missing meta description... But, the errors are all for non existent index files in directories that only contain pdf files and some thumbs of the front page... Just started trying to learn this stuff...! Cheers Rod
On-Page Optimization | | DrWho0