Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Should me URLs be uppercase or lowercase
-
I'm in the middle of doing a bunch of 301 redirects for me site. Should I make them Lowercase, uppercase, or does it matter? Also, do I want to be using hyphens (-), or underscores (_)?
Any other tips?
EX:
http://www.stupid.com/golf-slippers.html
OR
-
Are you serving the same page for both /MBA and /mba? You should set up a 301 redirect from one to the other.
In Analytics, you can set a custom filter to make your URLs case insensitive, but I don't believe that'll fix the data currently in your account, it'll only fix them going forward. That process is outlined here: http://support.google.com/googleanalytics/bin/answer.py?hl=en&answer=90397.
My URLs are all lowercased so I can't actually find an example in my account to test, but when I do an advenced filter and select Include Page with the match type of "Matching RegExp" and try URLs with uppercase characters, Analytics appears to be making the query case insensitive. So you can try that as well.
If the prior paragraph didn't work for you, you can do this on a URL by URL basis, by doing an advanced search by regular expression and substituting in "[M|m][B|b][A|a]" for "mba".
-
John and Team,
Suppose I have a site with 8000 pages and a small percentage of them in google analytics are showing up twice because the pages are linked through upper and lowercase links. So in Google analytics (GA) there would be two rows:
http:///www.blah.com/MBA
Is there a type of query I can run in GA to allow me to find all pages where the url is the same except for the upper and lower casing? I want to go into GA at Standard Reporting -> Content -> all pages and do an advanced search.
-
Is the Short Title a redirect or the actual URL? If it's a redirect then it shouldn't be a problem. You can also use canonical if you want to use the lowercase and host the other page. But that's is a sloppy solution.
-
I agree with this completely. Lowercase with hyphens is my preference. Some E-commerce companies (let's use Volusion as an example), give you the choice to use hyphens or underscores in a product's URL. This is nice and they even offer you the ability to write what the URL will be under the "Short TItle" option. However, this "Short Title" option is what will show in your URL AND the link text for a "Related Product" listing on the site. So if I want the link text to look normal when someone sees a Related Product link, I have to capitalize the text I put for the "Short Title" which will then cause the URL to be in Caps as well. This is a pretty significant flaw in their system and I have alerted them to the problem numerous times. Hopefully they will get it fixed soon.
-
I agree. Lowercase in the example above. For ease of direct typing.
But when advertising just a website URL in print offline, caps. www.BulwarkPestControl.com is easier to read than www.bulwarkpestcontrol.com.
-
The fact a Caps create a 404 error on LAMP site is a pet peeve of mine - so is the fact Google thinks mix cases on IIS are separate (thus duplicate) URLs.
Too arbitrary to be picky about and cause user frustration.
Thanks goodness at lease DoMaInS can be what ever.
-
I like lowercase because when I type URLs by hand, I don't think to capitalize things. If you capitalize things, you have to get the casing right to make the URL valid (unless you're setting up all sorts of fancy redirects), otherwise you a 404 and are left scratching your head. Also, I agree with Dan that it looks better.
Hyphens vs. underscores is a classic question; Matt Cutts says to go with hyphens: http://www.youtube.com/watch?v=Q3SFVfDIS5k. I like that better too.
-
Justin
I personally prefer lowercase because to me it looks better. And I prefer hyphens for the same reason, and it seems like these days everyone from WordPress right to the SEOmoz site does it that way.
-Dan
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Trailing slash URLs and canonical links
Hi, I've seen a fair amount of topics speaking about the difference between domain names ending with or without trailing slashes, the impact on crawlers and how it behaves with canonical links.
Technical SEO | | GhillC
However, it sticks to domain names only.
What about subfolders and pages then? How does it behaves with those? Say I've a site structured like this:
https://www.domain.com
https://www.domain.com/page1 And for each of my pages, I've an automatic canonical link ending with a slash.
Eg. rel="canonical" href="https://www.domain.com/page1/" /> for the above page. SEM Rush flags this as a canonical error. But is it exactly?
Are all my canonical links wrong because of that slash? And as subsidiary question, both domain.com/page1 and domain.com/page1/ are accessible. Is it this a mistake or it doesn't make any difference (I've read that those are considered different pages)? Thanks!
G0 -
Robots.txt Syntax for Dynamic URLs
I want to Disallow certain dynamic pages in robots.txt and am unsure of the proper syntax. The pages I want to disallow all include the string ?Page= Which is the proper syntax?
Technical SEO | | btreloar
Disallow: ?Page=
Disallow: ?Page=*
Disallow: ?Page=
Or something else?0 -
Tool to Generate All the URLs on a Domain
Hi all, I've been using xml-sitemaps.com for a while to generate a list of all the URLs that exist on a domain. However, this tool only works for websites with under 500 URLs on a domain. The paid tool doesn't offer what we are looking for either. I'm hoping someone can help with a recommendation. We're looking for a tool that can: Crawl, and list, all the indexed URLs on a domain, including .pdf and .doc files (ideally in a .xls or .txt file) Crawl multiple domains with unlimited URLs (we have 5 websites with 500+ URLs on them) Seems pretty simple, but we haven't been able to find something that isn't tailored toward management of a single domain or that can crawl a huge volume of content.
Technical SEO | | timfrick0 -
Url folder structure
I work for a travel site and we have pages for properties in destinations and am trying to decide how best to organize the URLs basically we have our main domain, resort pages and we'll also have articles about each resort so the URL structure will actually get longer:
Technical SEO | | Vacatia_SEO
A. domain.com/main-keyword/state/city-region/resort-name
_ domain.com/family-condo-for-rent/orlando-florida/liki-tiki-village_ _ domain.com/main-keyword-in-state-city/resort-name-feature _
_ domain.com/family-condo-for-rent/orlando-florida/liki-tiki-village/kid-friend-pool_ B. Another way to structure would be to remove the location and keyword folders and combine. Note that some of the resort names are long and spaces are being replaced dynamically with dashes.
ex. domain.com/main-keyword-in-state-city/resort-name
_ domain.com/family-condo-for-rent-in-orlando-florida/liki-tiki-village_ _ domain.com/main-keyword-in-state-city/resort-name-feature_
_ domain.com/family-condo-for-rent-in-orlando-florida/liki-tiki-village-kid-friend-pool_ Question: is that too many folders or should i combine or break up? What would you do with this? Trying to avoid too many dashes.0 -
Wordpress BackupBuddy adding ?doing_wp_cron= in URLS
Hi Has anyone found WordPress Backup Buddy causing a problem with SEO. I understand why it does it, but wondered if anyone experienced issues with this? Only sometimes it adds /?doing_wp_cron=****** on to the end of a URL Thanks Tom
Technical SEO | | TomPryor831 -
Best URL format for pagination
We're currently changing the URL format of our website search, we have been discussing a lot and cannot decide the past way to pass the pagination parameter for SEO. We narrowed down to the options. www.website.com/apples/p2 - www.website.com/apples?page=2 - www.website.com/apples/page/2 What would give us best ranking returns? What do you think?
Technical SEO | | HelpSaude0 -
404 errors on non-existent URLs
Hey guys and gals, First Moz Q&A for me and really looking forward to being part of the community. I hope as my first question this isn't a stupid one but I was just struggling to find any resource that dealt with the issue and am just looking for some general advice. Basically a client has raised a problem with 404 error pages - or the lack thereof- on non-existent URLs on their site; let's say for example: 'greatbeachtowels.com/beach-towels/asdfas' Obviously content never existed on this page so its not like you're saying 'hey, sorry this isn't here anymore'; its more like- 'there was never anything here in the first place'. Currently in this fictitious example typing in 'greatbeachtowels.com/beach-towels/asdfas**'** returns the same content as the 'greatbeachtowels.com/beach-towels' page which I appreciate isn't ideal. What I was wondering is how far do you take this issue- I've seen examples here on the seomoz site where you can edit the URI in a similar manner and it returns the same content as the parent page but with the alternate address. Should 404's be added across all folders on a site in a similar way? How often would this scenario be and issue particularly for internal pages two or three clicks down? I suppose unless someone linked to a page with a misspelled URL... Also would it be worth placing 301 redirects on a small number of common mis-spellings or typos e.g. 'greatbeachtowels.com/beach-towles' to the correct URLs as opposed to just 404s? Many thanks in advance.
Technical SEO | | AJ2340 -
What tool do you use to check for URLs not indexed?
What is your favorite tool for getting a report of URLs that are not cached/indexed in Google & Bing for an entire site? Basically I want a list of URLs not cached in Google and a seperate list for Bing. Thanks, Mark
Technical SEO | | elephantseo3