What would cause these ⠃︲蝞韤諫䴴SPপ� emblems in my urls?
-
In Search Console I am getting errors under other. It is showing urls that have this format-
https://www.site.com/Item/654321~SURE⠃︲蝞韤諫䴴SPপ�.htm
When clicked it shows 蝞韤諫䴴SPপ� instead of the % stuff.
As you can see this is an item page and the normal item page pulls up fine with no issues. This doesn't show it is linked from anywhere. Why would google pull this url? It doesn't exist on the site anywhere. It is a custom asp.net site. This started happening in mid May but we didn't make any changes then.
-
They are encoded URLs. For example, Google will turn all " into %22 and all spaces in the URL to %20. You can learn more about them here http://www.w3schools.com/tags/ref_urlencode.asp . And here is a useful tool for encoding and decoding URLs: http://meyerweb.com/eric/tools/dencoder/ .
What you need to do is have the developer "escape out" or "rewrite" all non-alphanumeric characters in the URL. You'll also have to 301 redirect the old URLs to these new, search-engine friendly ones without the characters that get automatically encoded, like parentheses, commas, tildas and plus signs.
-
It could be due to many reasons. One would need to know the domain in order to do further analysis on the issue.
-
What I want to know is why Google is finding these pages at all. I can normally look at linked from and find any problems. But if these are not linked from anywhere why is Google finding them?
-
Hello,
I believe that in URLs % signs plus letters/number can be translated into different characters. For instance %20 is a space. %21 is a !. WC3 have a guide here http://www.w3schools.com/tags/ref_urlencode.asp.
I don't know why it would translate into oriental characters, but that may give you a place to start your investigation.
Hope this helps.Cheers,
Luke
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Changing URLS: from a short well optimised URL to a longer one – What's the traffic risk
I'm working with a client who has a website that is relatively well optimised, thought it has a pretty flat structure and a lot of top level pages. They've invested in their content over the years and managed to rank well for key search terms. They're currently in the process of changing CMS and as a result of new folder structuring in the CMS the URLs for some pages look to have significantly changed. E.g Existing URL is: website.com/grampians-luxury-accommodation which ranked quite well for luxury accommodation grampians New URL when site is launched on new CMS would be website.com/destinations/victoria/grampians My feeling is that the client is going to lose out on a bit of traffic as a result of this. I'm looking for information or ways or case studies to demonstrate the degree of risk, and to help make a recommendation to mitigate risk.
Intermediate & Advanced SEO | | moge0 -
URL Rewriting Best Practices
Hey Moz! I’m getting ready to implement URL rewrites on my website to improve site structure/URL readability. More specifically I want to: Improve our website structure by removing redundant directories. Replace underscores with dashes and remove file extensions for our URLs. Please see my example below: Old structure: http://www.widgets.com/widgets/commercial-widgets/small_blue_widget.htm New structure: https://www.widgets.com/commercial-widgets/small-blue-widget I've read several URL rewriting guides online, all of which seem to provide similar but overall different methods to do this. I'm looking for what's considered best practices to implement these rewrites. From what I understand, the most common method is to implement rewrites in our .htaccess file using mod_rewrite (which will find the old URLs and rewrite them according to the rewrites I implement). One question I can't seem to find a definitive answer to is when I implement the rewrite to remove file extensions/replace underscores with dashes in our URLs, do the webpage file names need to be edited to the new format? From what I understand the webpage file names must remain the same for the rewrites in the .htaccess to work. However, our internal links (including canonical links) must be changed to the new URL format. Can anyone shed light on this? Also, I'm aware that implementing URL rewriting improperly could negatively affect our SERP rankings. If I redirect our old website directory structure to our new structure using this rewrite, are my bases covered in regards to having the proper 301 redirects in place to not affect our rankings negatively? Please offer any advice/reliable guides to handle this properly. Thanks in advance!
Intermediate & Advanced SEO | | TheDude0 -
URL Injection Hack - What to do with spammy URLs that keep appearing in Google's index?
A website was hacked (URL injection) but the malicious code has been cleaned up and removed from all pages. However, whenever we run a site:domain.com in Google, we keep finding more spammy URLs from the hack. They all lead to a 404 error page since the hack was cleaned up in the code. We have been using the Google WMT Remove URLs tool to have these spammy URLs removed from Google's index but new URLs keep appearing every day. We looked at the cache dates on these URLs and they are vary in dates but none are recent and most are from a month ago when the initial hack occurred. My question is...should we continue to check the index every day and keep submitting these URLs to be removed manually? Or since they all lead to a 404 page will Google eventually remove these spammy URLs from the index automatically? Thanks in advance Moz community for your feedback.
Intermediate & Advanced SEO | | peteboyd0 -
Pagination causing duplicate content problems
Hi The pagination on our website www.offonhols.com is causing duplicate content problems. Is the best solution adding add rel=”prev” / “next# to the hrefs As now the pagination links at the bottom of the page are just http://offonhols.com/default.aspx?dp=1
Intermediate & Advanced SEO | | offonhols
http://offonhols.com/default.aspx?dp=2
http://offonhols.com/default.aspx?dp=3
etc0 -
Will SEO cause a drop in the number of impressions?
Hello, I have been a member of the Moz community for a long time. I very seldom ask questions here but this time I really need your help to make sure I will not make mistakes that will negatively affect my site. My site monetizes according to the number of impressions visualized by the users who visit it. I now want to try to optimize it by using all those nice SEO techniques I have learned through Moz. My goal is to make sure that if I use the various SEO strategies, I will still be able to obtain the high number of impressions I get now. If not, I prefer to leave the situation untouched and I will not start optimizing the site for SEO. Please kindly read the situation below and give me a little support to make sure I am doing the right thing. I would like to ask for your professional advice to solve and issue related to duplicate content. Please visit my site: www.chhedonna.it. The sitemap has indexed 21.890 articles, but if I digit the command 'site:www.chedonna.it', I obtain 158.000 results.I believe that the duplicated content has emerged due to three errors I would like to indicate in order for you to tell me whether my interpretation is correct or not: The article writers made a mistake in that although the content of the articles is different, they have employed the same title for all of them. Therefore, you can see 5 articles with similar/duplicated Tag Titles but the content of said articles is different. For example, http://www.chedonna.it/attualita/cronaca-rosagossip/2012/12/03/heidi-klum-senza-trucco-e-parrucco-foto and http://www.chedonna.it/attualita/cronaca-rosagossip/2012/12/03/heidi-klum-senza-trucco-e-parrucco-foto-2/ are different articles (i.e., the content of the articles is different from each other) that have been published using the same titles. If I inserted the 'follow-noindex' tag to 3 of the 4 duplicates, as shown in the example above, would that be a solution? I fear that if I did insert the follow-noindex, I would cause a drop in the number of impressions visualized by my site's users. It is important to point out, in fact, that I monetize the site via the number of impressions I generate. Therefore, it is fundamental that I do not compromise the number of impressions that the site gets, if I try to optimize it for SEO reasons. On the other hand, I believe that the idea to operate via a 'rel canonical' would not be right, considering that the content of every post having the same title is different, that is, the articles are different even if they focus on the same topic. Also, I would not find it beneficial to use a '301 redirect', since the number of duplicated Titles Tags is very consistent. 2. The second error concerns the duplicate content due to the images that have been included in the articles. For instance, http://www.chedonna.it/che-donna-di-mondo/fare-la-valigia/2012/08/06/campeggio-vacanza-in-liberta/attachment/tenda/ http://www.chedonna.it/che-donna-di-mondo/fare-la-valigia/2012/08/06/campeggio-vacanza-in-liberta/attachment/tenda-2/ http://www.chedonna.it/che-miss/2012/12/04/tatuaggi-fiore-di-loto-significato-e-foto/attachment/tatuaggio-fiore-di-loto-2/ http://www.chedonna.it/che-miss/2013/03/15/tatuaggi-fiore-di-loto-significato-e-foto-2/attachment/tatuaggio-fiore-di-loto-2-2/. I could solve this problem by preventing the media indexing. But I fear that this would produce a very high number of '404 error' messages. If such a thing did happen, my site would stop monetizing overall and I cannot allow this to occur, as you can understand. My very important question is the following: if I prevent the indexing of the photos, will I get in return a drop in the number of impressions that my site would normally generate? 3. Duplicate content generated by the indexing of archive subpages. For example: http://www.chedonna.it/tag/angelica-e-ferdinando/ http://www.chedonna.it/tag/angelica-e-ferdinando/page2/ http://wwwchedonna.it/tag/angelica-e-ferdinando/page3/ If I prevented the media indexing, will I get in return a drop in the number of impressions and many 404 errors? Thank you very much for taking the time to help me sort out this very important issue. Cheers, Sal
Intermediate & Advanced SEO | | salvyy0 -
Switching Url
I started working with a Roofer/Contractor about a year ago. His website is http://www.lancasterparoofing.com/. The name of his business is Spicher Home Improvements. He used to have spicherhomeimprovements.com, well he still does. He was focusing on Roofing and Siding but now would like to branch to other areas like Interior remodeling. So adding interior work under LancasterPaRoofing.com is not applicable. I do not think starting another domain and having two is the best option. I think he should go back to using SpicherHomeImprovements.com and I assume he would take a small hit but in time he should be better off. Plus the url is more applicable to the real name of his business. Thanks for any feedback I receive. Chad
Intermediate & Advanced SEO | | ChadEisenhart0 -
Should /node/ URLs be 301 redirect to Clean URLs
Hi All! We are in the process of migrating to Drupal and I know that I want to block any instance of /node/ URLs with my robots.txt file to prevent search engines from indexing them. My question is, should we set 301 redirects on the /node/ versions of the URLs to redirect to their corresponding "clean" URL, or should the robots.txt blocking and canonical link element be enough? My gut tells me to ask for the 301 redirects, but I just want to hear additional opinions. Thank you! MS
Intermediate & Advanced SEO | | MargaritaS0 -
Non-www URL showing in Blog
Thanks to Sanket Patel in an earlier query I've now got non-www pages showing as www. pages on my www.nile-cruises-4u.co.uk website. But the Blog which is part of the site posts and pages still show as non-www pages. For example: http://nile-cruises-4u.co.uk/blog/makadi-palace-hotel-makadi-bay/ I wonder if anyone has come upon the same problem and what the solution might be? Thanks, Colin
Intermediate & Advanced SEO | | NileCruises1