How google bot see's two the same rel canonicals?
-
Hi,
I have a website where all the original URL's have a rel canonical back to themselves. This is kinda like a fail safe mode. It is because if a parameter occurs, then the URL with the parameter will have a canonical back to the original URL.
For example this url: https://www.example.com/something/page/1/ has this canonical: https://www.example.com/something/page/1/ which is the same since it's an original URL
This url https://www.example.com/something/page/1/?parameter has this canonical https://www.example.com/something/page/1/ like i said before, parameters have a rel canonical back to their original url's.
SO: https://www.example.com/something/page/1/?parameter and this https://www.example.com/something/page/1/ both have the same canonical which is this https://www.example.com/something/page/1/
Im telling you all that because when roger bot tried to crawl my website, it gave back duplicates. This happened because it was reading the canonical (https://www.example.com/something/page/1/) of the original url (https://www.example.com/something/page/1/) and the canonical (https://www.example.com/something/page/1/) of the url with the parameter (https://www.example.com/something/page/1/?parameter) and saw that both were point to the same canonical (https://www.example.com/something/page/1/)...
So, i would like to know if google bot treats canonicals the same way. Because if it does then im full of duplicates
thanks.
-
Its not about the canonical, its about the crawl optimization. I know that canonical URL saves the situation here, i am working under a fail safe mode in matter of duplicates and i want to believe that the canonical URL implementation is better than good in my website.
I just don't want bot's spending time on pages that have nothing actual to say and are canonicalized to pages that have the important content. That is why i configured the bot to not crawl those parameters in the URL parameters tab in GWT and eventually some time to even drop those results.
-
I would think that you're going a little over the top with what essentially is the job of a canonical tag. you don't need to block robots going to the pages as the canonical tag will be telling robots that its a duplicate version. if the urls have already been indexed it will take time for them to drop off.
-
All the parameters are configured to NO URL's in google webmaster tools URL parameters tab. Check the image http://prntscr.com/e9fs91
Its a better setting to do it straight from webmaster tools than disallowing the parameters in robots.txt
Tho, i have a problem with that because google is indexing these parameters even if its configured to NO URL's check my post here: https://moz.com/community/q/web-master-tools-url-parameters
-
Hello,
Rogerbot struggles a bit with canonical last I checked. You've the right set up you want to stop parameters it's especially helpful for stopping people rankings pages on your site like /?this-site-sucks! Always remember Rogerbot of any other services are a guide only to help you not a 100% true resource that will help you rank so use them like a tool not an authority.
TL:DR - your set up is all ok!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Sub domain? Micro site? What's the best solution?
My client currently has two websites to promote their art galleries in different parts of the country. They have bought a new domain (let's call it buyart.com) which they would eventually like to use as an e-commerce platform. They are wondering whether they keep their existing two gallery websites (non e-commerce) separate as they always have been, or somehow combine these into the new domain and have one overarching brand (buyart.com). I've read a bit on subdomains and microsites but am unsure at this stage what the best option would be, and what the pros and cons are. My feeling is to bring it all together under buyart.com so everything is in one place and creates a better user journey for anyone who would like to visit. Thoughts?
Technical SEO | | WhitewallGlasgow0 -
How to get into Google's Tops Stories?
Hi All, I have been doing research for a few weeks and I cannot for the life of me figure out why I cannot get my website (Racenet) into the top stories in Google. We are in Google News, have "news article" schema, have AMP pages. Our news articles also perform quite well organically and we typically dominate the Google News section. We have two main competitors (Punters and Just Horse Racing) who are both in top stories and I cannot find anything that we are doing that they aren't. Apparently the AMP "news article" schema is incorrect and that could be the reason why we aren't showing up in Google Top Stories, but I can't find anything wrong with the schema and it looks the same as our competitors. For example: https://search.google.com/structured-data/testing-tool/u/0/#url=https%3A%2F%2Fwww.racenet.com.au%2Fnews%2Fblake-shinn-booked-to-ride-doncaster-handicap-favourite-alizee-20190331%3FisAmp%3D1 Does anyone have any ideas of why I cannot get my site into Google Top Stories? Any and all help would be greatly appreciated. Thanks! 🙂
Technical SEO | | Saba.Elahi.M.0 -
Drupal, http/https, canonicals and Google Search Console
I’m fairly new in an in-house role and am currently rooting around our Drupal website to improve it as a whole. Right now on my radar is our use of http / https, canonicals, and our use of Google Search Console. Initial issues noticed: We serve http and https versions of all our pages Our canonical tags just refer back to the URL it sits on (apparently a default Drupal thing, which is not much use) We don’t actually have https properties added in Search Console/GA I’ve spoken with our IT agency who migrated our old site to the current site, who have recommended forcing all pages to https and setting canonicals to all https pages, which is fine in theory, but I don’t think it’s as simple as this, right? An old Moz post I found talked about running into issues with images/CSS/javascript referencing http – is there anything else to consider, especially from an SEO perspective? I’m assuming that the appropriate certificates are in place, as the secure version of the site works perfectly well. And on the last point – am I safe to assume we have just never tracked any traffic for the secure version of the site? 😞 Thanks John
Technical SEO | | joberts0 -
Redirecting .edu subdomains to our site or taking the link, what's more valuable?
We have a relationship built through a service we offer to universities to be issued a .edu subdomain that we could redirect to our landing page relevant to that school. The other option is having a link from their website to that same page. My first question is, what would be more valuable? Can you pass domain authority by redirecting a subdomain to a subdirectory in my root domain? Or would simply passing the link equity from a page in their root domain to our page pass enough value? My second question is, if creating a subdomain with a redirect is much more valuable, what is the best process for this? Would we simply have their webmaster create the subdomain for us an have them put a 301 redirect to our page? Is this getting in the greyer hat area? Thanks guys!
Technical SEO | | Dom4410 -
How Google sees my page
When looking for crawlability issues, what is the difference between using webmaster tools Fetch as google, looking at the cached pages in google index site:mypage.com, or using spider simulator tools.
Technical SEO | | shashivzw0 -
Google see 302 but it is 200 OK - OsCommerce
Hi! I´ve a problem with a webshop that run OsCOmmerce. Google report thousands of urls "Not followed" in Webmaster Tools, and it´s increasing for every day. When you fetch as Google you get a 302 to ../cookie_usage.php. But if you fetch the page in your browser you get 200 OK. Why does Google get 302 but users 200? We´ve checked for malware and restored old backups, nothing helps. Thoughts about this?
Technical SEO | | Vivamedia0 -
Querystring params, rel canonical and SEO
I know ideally you should have as clean as possible url structures for optimal SEO. Our current site contains clean urls with very minimal use of query string params. There is a strong push, for business purposes to include click tracking on our site which will append a query string param to a large percentage of our internal links. Currently: http://www.oursite.com/section/content/ Will change to: http://www.oursite.com/section/content/?tg=zzzzwww We currently use rel canonical on all pages to properly define the true url in order to remove any possible duplicate content issues. Given we are already using rel canonical, if we implement the query string click tracking, will this negatively impact our SEO? If so, by how much? Could we run into duplicate content issues? We get crawled by Google a lot (very big site) and very large percent of our traffic is from Google, but there is a strong business need for this information so trying to weigh pros/cons.
Technical SEO | | NicB10 -
Switching Site to a Domain Name that's in Use
I'm comfortable with the steps of moving a site to a new domain name as recommended by Google. However, in this case, the domain name I'm asked to move to is not really "new" ... meaning it's currently hosting a website and has been for a long time. So my question is, do I do this in steps and take the old website down first in order to "free up" the domain name in they eyes of search engines to avoid large numbers of 404s and then (in step 2) switch to the "new" domain in a few months? Thanks.
Technical SEO | | R2iSEO0