Thank you, David.
- Home
- TalkInThePark
Latest posts made by TalkInThePark
-
A/B Tests: How To Verify Difference In Average Order Value?
Hi there! When the data from an A/B test shows a difference in AOV between the variants, how do you determine if this difference is statistically probable or not? Best Regards, Martin
-
RE: Google Algorithm change? - Brand name now overwriting title tag?
I am seeing this too, Google flips the order of our brand name and our USP in our homepage title.
"USP - Brand Name"
now becomes
"Brand Name: USP"
-
RE: Googlebot does not obey robots.txt disallow
We do not currently have any sanitation rules in order to maintain the nocrawl param. But that is a good point. 301:ing will be difficult for us but I will definitely add the nocrawl param to the rel canonical of those internal SERPs.
-
RE: Googlebot does not obey robots.txt disallow
Thank you, Igol. I will definitely look into your first suggestion.
-
RE: Googlebot does not obey robots.txt disallow
Thank you, Cyrus.
This is what it looks like:
User-agent: *
Disallow: /nocrawl=1The weird thing is that when testing one of the sample URLs (given by Google as "problematic" in the GWMT message and that contains the nocrawl param) on the GWMT "Blocked URLs" page by entering the contents of our robots.txt and the sample URL, Google says crawling of the URL is disallowed for Googlebot.
On the top of the same page, it says "Never" under the heading "Fetched when" (translated from Swedish..). But when i "Fetch as Google" our robots.txt, Googlebot has no problems fetching it. So i guess the "Never" information is due to a GWMT bug?
I also tested our robots.txt against your recommended service http://www.frobee.com/robots-txt-check. It says all robots has access to the sample URL above, but I gather the tool is not wildcard-savvy.
I will not disclose our domain in this context, please tell me if it is ok to send you a PW.
About the noindex stuff. Basically, the nocrawl param is added to internal links pointing to internal search result pages filtered by more than two params. Although we allow crawling of less complicated internal serps, we disallow indexing of most of them by "meta noindex".
-
RE: Googlebot does not obey robots.txt disallow
Igal, thank your for replying.
But robots.txt disallowing URLs by matching patterns has been supported by Googlebot for a long time now.
-
Googlebot does not obey robots.txt disallow
Hi Mozzers!
We are trying to get Googlebot to steer away from our internal search results pages by adding a parameter "nocrawl=1" to facet/filter links and then robots.txt disallow all URLs containing that parameter.
We implemented this late august and since that, the GWMT message "Googlebot found an extremely high number of URLs on your site", stopped coming.
But today we received yet another. The weird thing is that Google gives many of our nowadays robots.txt disallowed URLs as examples of URLs that may cause us problems.
What could be the reason?
Best regards,
Martin
-
RE: Erroneous "Weekly Keyword Ranking & On-page Optimization Report" For Campaign
It looks good now.
Thanks!
Best posts made by TalkInThePark
-
HTML Forms Dilute Pagerank?
Today, we have way too many links on our homepage. About 30 of them are add-to-basket links (regular html links) pointing to a separate application. This application 302 redirects the client back to the referring page.
I have two questions:
1. Does the current implementation of our buttons dilute pagerank? Bear in mind the 302 redirect.
2. If the answer to the first question is yes, would transforming the buttons into form buttons change anything to the better? We would still 302 back to the referring page. I know Gbot follows GET forms and even POST forms, but does GBot pass on pagerank to the form URL?
-
RE: Erroneous "Weekly Keyword Ranking & On-page Optimization Report" For Campaign
It looks good now.
Thanks!
Looks like your connection to Moz was lost, please wait while we try to reconnect.