Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Direct traffic is up 2100% (due to a bot/crawler I believe)
-
Hi,
The direct traffic to website www.webgain.dk has increased by over 2100% recently. I can see that most of it is from US (my target audience is in Denmark and the website is in danish).
What can I do about this? All this traffic gives my website a bounce rate of 99.91% for direct traffic. I believe it is some sort of bot/crawler. -
Already done. They also included the tip in their newsletter for beta-testers.
-
You might want to let them know about this, so they can add in documentation so future users know what is up before panicking.
-
Follow up: I have fixed this now. It was a monitoring tool by Digicure, where I have signed up to be a beta tester. Their platform checks the website like a normal visitor from servers around the world (in my test case it is Denmark and California) and thereby it looked like normal direct traffic in my data. I excluded their stated server IP-addresses in my Google Analytics filters and that helped. Thanks again guys for the help.
-
Thank you for all your great advice. I will follow them and see how it works.
-
If you are running Wordpress also check what page / pages are being accessed. I have had bots nail my wp-login like that before. If that is the case harden your installation, one thing I have found that stopped it was setting a deny in the htaccess on wp-login / wp-admin.
-
I was having the same problem ( for me it seemed to be Bings ads bot) . I used this guide below and it seems to filter out most of the bot visits.
-
I would check the service providers first just to know for sure they're all coming from the same provider. You can check this by visiting your Audience > Technology > Network report on the left side of your Google Analytics. If you see the same network and browsers being used I would use a filter (only if you're really determined/ 100% sure that it's bot traffic) to get them completely out of your Google Analytics view.
-
It's weird that the bot is accepting cookies, but with a bounce rate that high, I agree it's probably something automated (though it could be people who were looking for something else or were directed there by an email or an app accidentally). You can look through your logs to see IP addresses and then do as Atakala says and block the traffic if you're worried about bandwidth. You can also just filter it out in GA by excluding US traffic (if your'e worried about analytics being messed up).
-
It's probably AWS, which is amazon buyable crawl service.
If ıt costs you too much then try to ban it.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
How do you optimise a website for European traffic?
I have a design portfolio website here https://www.nicholsoncreative.com/ which uses a .com but is currently configured through the Search Console to appear in results for Google.co.uk. I am going to be restructuring the website and optimisation and I want to bring in more traffic/enquiries/business from around Europe. As there's no Google.eu, and as Google also serves results based on the searchers geographic location it would seem difficult to structure and optimise content so that results can be found across all of Europe. I assume simply switching to a .eu domain extension for my own website wouldn't solve the problem? I also assume that creating content in different languages would be a logical (if time consuming) option? Are there any other tried and trusted techniques that can be used to target traffic throughout Europe? I'd appreciate any advice.
International SEO | | JCN-SBWD0 -
Using a top level domain name and directing it to a subfolder
Hi, we have a large international network. Our main website sits on .com domain and is used by the UK market. We have an international site in a subdirectory .com/dk/ for Denmark for example. We have also purchased the domain name www.ourcompany.dk/. Should we be forwarding the domain name (www.ourcompany.dk/) to point to the subdirectory www.ourcomany.com.dk/ so in the browser it shows up as www.ourcompany.dk or should we be displaying it as www.ourcompany.com/dk/? Are there any pros and cons to this method? Which one is best and are there any benefits in SEO. Ideally we want the .com domain name to have the best domain authority so would this impact it in any way? Any tips would be great.
International SEO | | Easigrass0 -
Hreflang for bilingual website in the same region/location
Hi everyone, got a quick question concerning the hreflang tag. I have a website with 2 different language versions targeting to the same region(Reason: The area is bilingual however not everyone speaks the other language fluently) Question:
International SEO | | ennovators
Can I use hreflang in that case like: Many thanks in advance0 -
International SEO Subfolders / user journey etc
Hi According to all the resources i can find on Moz and elsewhere re int seo, say in the context of having duplicate versions of US & UK site, its best to have subfolders i.e. domain.com/en-gb/ & domain.com/en-us/ however when it comes to the user journey and promoting web address seems a bit weird to say visit us at: domain.com/en-us/ !? And what happens if someone just enters in domain.com from the US or UK ? My client wants to use an IP sniffer but i've read thats bad practice and should employ above style country/language code instead, but i'm confused about both the user journey and experience in the case of multiple sub folders. Any advice much appreciated ? Cheers Dan
International SEO | | Dan-Lawrence0 -
What are the best practices for translation of city/state names for international SEO? (ie. New York in English vs. Nueva York in Spanish)
I'm working on international SEO / translation of a global travel site. While we have a global keyword research and translation strategy in process for each market they serve, I've run into a unique question. Overall, we are translating (and localizing) content for each market but aren't sure what to do with location names. Each country/state has cities and locations that have their own dedicated pages. I see three options for these location names (when titling a page and writing content): keep them in English, translate the names in the market languages, or use a combination of the two. The challenge with altering the location names to the market languages is that they are truly not known by those names. Though there are some instances where it may make sense…for instance **New York **in Spanish would be "Nueva York" with **‘**Nueva' being the Spanish translation of ‘new’. There are other instances, where no translation exists. If you’ve had a similar experience I'd love to hear your approach/recommendation.
International SEO | | JonClark150 -
Best URL structure for Multinational/Multilingual websites
Hi I am wondering what the best URL format to use is when a website targets several countries, in several languages. (without owning the local domains, only a .com, and ideally to use sub-folders rather than sub-domains.) As an example, to target a hotel in Sweden (Google.se) are there any MUST-HAVE indicators in the URL to target the relevant countries? Such as hotelsite.com**/se/**hotel-name. Would this represent the language? Or is it the location of the product? To clarify a bit, I would like to target around 10 countries, with the product pages each having 2 languages (the local language + english). I'm considering using the following format: hotelsite.com/en/hotel-name (for english) and hotelsite.com/se/hotel-name (for swedish content of that same product) and then using rel=”alternate” hreflang=”se-SV” markup to target the /se/ page for Sweden (Google.se) and rel=”alternate” hreflang=”en” for UK? And to also geotarget those in Webmaster tools using those /se/ folders etc. Would this be sufficient? Or does there need to be an indicator of both the location, AND the language in the URLs? I mean would the URL's need to be hotelsite.com/se/hotel-name/se-SV (for swedish) or can it just be hotelsite.com/se/hotel-name? Any thoughts on best practice would be greatly appreciated.
International SEO | | pikka0 -
SEO for Subdomains for different languages .com/fr, .com/es
Hi All, I was wondering how best to to approach optimisation of a site that exists on a single .com domain, but has different subfolders for different languages. The site is a .com and it has subfolders for French, Spanish, Russian and English. The business is situated in France and the vast majority of clients are French and English speakers. I've read that it's possible to geo target these subfolders using webmaster tools however I believe this is an inferior method of optimisation than having tld's. Just wondered if anyone had experience of htis and could provide any advice ? As they won't be rebuilding the site for another year or so I wondered if there were any quick wins? My second question is to do with how best to set these campaigns up within SEO Moz. would it be better to track at a subdomain or subfolder leverl (for different languages)? If someone could advise I would greatly appreciate it! Thanks, vantresca
International SEO | | vanvallejo0 -
IP Redirection vs. cloaking: no clear directives from Google
Hi there, Here is our situation:we need to force an IP Redirection for our US users to www.domain.com and at the same time we have different country-specific subfolders with thei own language such as www.domain.com/fr. Our fear is that by forcing an IP redirection for US IP, we will prevent googlebot (which has an US IP) from crawling our country-specific subfolders. I didn't find any clear directives from Google representatives on that matter. In this video Matt Cutts says it's always better to show Googlebot the same content as your users http://www.youtube.com/watch?v=GFf1gwr6HJw&noredirect=1, but on the other hand in that other video he says "Google basically crawls from one IP address range worldwide because (they) have one index worldwide. (They) don't build different indices, one for each country". This seems a contradiction to me... Thank you for your help !! Matteo
International SEO | | H-FARM0