Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Diagnosing Canonical Errors Is Screaming frog reliable?
-
Morning from suny & warm wetherby UK
On this page http://www.goldsboroughestates.co.uk/how-we-care-for-you/right-to-manage/ screaming frog is citing a canonical error but I'm confused as this piece of code is in place:
http://www.goldsboroughestates.co.uk/About/right-to-manage" />
So my question is please - "Does this page http://www.goldsboroughestates.co.uk/how-we-care-for-you/right-to-manage/ have a caninical error or is screaming frog useless?
Other examples where screaming frog is picking up canonical errors include:
http://www.goldsboroughestates.co.uk/what-our-customers-say/right-to-manage/
http://www.goldsboroughestates.co.uk/buying-a-home/right-to-manage/Oh forgot to say the preffered version is http://www.goldsboroughestates.co.uk/About/right-to-manage/
Any insights welcvome
-
Hey,
Long time since the Question, I was just wondering if you worked it out or not.
Gr.,
Istvan
-
I think Screaming Frog is just warning you that the canonical version doesn't seem to match the display URL. They can't really tell (we have the same problem in SEOmoz tools) what the "right" canonical is - they can just warn of a mismatch.
I'm a bit confused as to the purpose of the dual URLs here. The best canonical implementation is to use one URL. The canonical tag can act as a band-aid, but consistency is still the best defense. Having multiple paths to the same page is rarely beneficial.
-
Having spoke to oiur internal helpdesk (Who I trust & do know what theyre talking about) theyve taken a look at:
http://www.goldsboroughestates.co.uk/footer-links/left/right-to-manage/
http://www.goldsboroughestates.co.uk/how-we-care-for-you/right-to-manage/
http://www.goldsboroughestates.co.uk/buying-a-home/right-to-manage/
http://www.goldsboroughestates.co.uk/what-our-customers-say/right-to-manage/
and I'm afraid they have a different perspective which is they see no canonical problem
Hey ho think I'll just set my head on fire then maybe things will be more clearer
-
Hi Istvan - your advice is good but ive just discovered its not been implemented! Time to kick some ass, I'll update you
-
Hey,
Any news on how it went? I am curious if that was the problem or not.
Gr.,
Istvan
-
Hey,
Maybe this helps you a littlebit: http://www.seomoz.org/blog/an-seos-guide-to-http-status-codes
Dr. Pete's article explains well how the status codes work.
Gr.,
Istvan
-
Wow great anser, I'm on to this now & will updat you with how things went
-
Hey there!
I think I have found what your problem is with you canonical link
In your code you have:
And probably you are somewhere forcing the URls to have a / at the end.
So basically you are confusing browsers and search engine bots, because they now cannot tell which is the real version:
SE enters the page. Then it sees that the right version should be the one WITHOUT a "/" at the end, then that pages has a 301 redirect to the version which HAS a "/" at the end of the URL (but that has a canonical which points out that the preffered version should be ). So it is a non-ending circle.
So if you add a / to the end of your URl, your problem should be solved.
Final thought: Screaming Frog is working well.
I hope this was a solution.
Cheers,
Istvan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Errors In Search Console
Hi All, I am hoping someone might be able to help with this. Last week one of my sites dropped from mid first day to bottom of page 1. We had not been link building as such and it only seems to of affected a single search term and the ranking page (which happens to be the home page). When I was going through everything I went to search console and in crawl errors there are 2 errors that showed up as detected 3 days before the drop. These are: wp-admin/admin-ajax.php showing as response code 400 and also xmlrpc.php showing as response code 405 robots.txt is as follows: user-agent: * disallow: /wp-admin/ allow: /wp-admin/admin-ajax.php Any help with what is wrong here and how to fix it would be greatly appreciated. Many Thanks
Technical SEO | | DaleZon0 -
How long does it take for canonical tags to work
How long on average does it take for a canonical tag to work? Understand that canonicals are just a suggestion, but after adding a canonical tag and submitting the page via Google fetch, assuming Google follows the canonical, would you expect it to work after a day or two or does it take longer? We added canonicals to old PPC landing pages that are ranking organically, though our new landing pages (which we want to rank organically) are not identical and have a bit more content/features. They are similar though. Canonicals were added to the old pages (pointing to new pages) and requested indexing via search console. Old pages are still ranking and new pages not so much. FYI we are unable to 301 old PPC pages due to other non negotiable reasons unfortunately. Thanks.
Technical SEO | | SoulSurfer80 -
Can you use Screaming Frog to find all instances of relative or absolute linking?
My client wants to pull every instance of an absolute URL on their site so that they can update them for an upcoming migration to HTTPS (the majority of the site uses relative linking). Is there a way to use the extraction tool in Screaming Frog to crawl one page at a time and extract every occurrence of _href="http://" _? I have gone back and forth between using an x-path extractor as well as a regex and have had no luck with either. Ex. X-path: //*[starts-with(@href, “http://”)][1] Ex. Regex: href=\”//
Technical SEO | | Merkle-Impaqt0 -
Set Canonical for Paginated Content
Hi Guys, This is a follow up on this thread: http://moz.com/community/q/dynamic-url-parameters-woocommerce-create-404-errors# I would like to know how I can set a canonical link in Wordpress/Woocommerce which points to "View All" on category pages on our webshop.
Technical SEO | | jeeyer
The categories on my website can be viewed as 24/48 or All products but because the quanity constantly changes viewing 24 or 48 products isn't always possible. To point Google in the right direction I want to let them know that "View All" is the best way to go.
I've read that Google's crawler tries to do this automatically but not sure if this is the case on on my website. Here is some more info on the issue: https://support.google.com/webmasters/answer/1663744?hl=en
Thanks for the help! Joost0 -
Screaming Frog showing 503 status code. Why?
Screaming Frog is showing a 503 code for images. If I go and use a header checker like SEOBook it shows 200. Why would that be? Here is an example link- http://germanhausbarn.com/wp-content/uploads/2014/07/36-UPC-5145536-John-Deere-Stoneware-Logo-Mug-pair-25.00-Heavy-4-mugs-470x483.jpg
Technical SEO | | EcommerceSite0 -
Exclude status codes in Screaming Frog
I have a very large ecommerce site I'm trying to spider using screaming frog. Problem is I keep hanging even though I have turned off the high memory safeguard under configuration. The site has approximately 190,000 pages according to the results of a Google site: command. The site architecture is almost completely flat. Limiting the search by depth is a possiblity, but it will take quite a bit of manual labor as there are literally hundreds of directories one level below the root. There are many, many duplicate pages. I've been able to exclude some of them from being crawled using the exclude configuration parameters. There are thousands of redirects. I haven't been able to exclude those from the spider b/c they don't have a distinguishing character string in their URLs. Does anyone know how to exclude files using status codes? I know that would help. If it helps, the site is kodylighting.com. Thanks in advance for any guidance you can provide.
Technical SEO | | DonnaDuncan0 -
Best 404 Error Checker?
I have a client with a lot of 404 errors from Web Master Tools, and i have to go through and check each of the links because Some redirect to the correct page Some redirect to another url but its a 404 error Some are just 404 errors Does anyone know of a tool where i can dump all of the urls and it will tell me If the url is redirected, and to where if the page is a 404 or other error Any tips or suggestions will be really appreciated! Thanks SEO Moz'rs
Technical SEO | | anchorwave0 -
Robots.txt and canonical tag
In the SEOmoz post - http://www.seomoz.org/blog/robot-access-indexation-restriction-techniques-avoiding-conflicts, it's being said - If you have a robots.txt disallow in place for a page, the canonical tag will never be seen. Does it so happen that if a page is disallowed by robots.txt, spiders DO NOT read the html code ?
Technical SEO | | seoug_20050