Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
How do you implement dynamic SEO-friendly URLs using Ajax without using hashbangs?
-
We're building a new website platform and are using Ajax as the method for allowing users to select from filters. We want to dynamically insert elements into the URL as the filters are selected so that search engines will index multiple combinations of filters. We're struggling to see how this is possible using symfony framework. We've used www.gizmodo.com as an example of how to achieve SEO and user-friendly URLs but this is only an example of achieving this for static content.
We would prefer to go down a route that didn't involve hashbangs if possible.
Does anyone have any experience using hashbangs and how it affected their site?
Any advice on the above would be gratefully received.
-
Hey,
so I guess you're wanting to achieve this (http://extensions.ecommerce-team.com/apparel?___store=shopbypro&reset_filter=1) - aor something similar using ajax so the page updates without re-loading?
I guess you need to use token's in the url path rather than a dynamic url. Its a big job as you'll have to define some rules to say colour must come first, brand second, range last or something similar so you end up with domain.com/brown/levis/jackets/ - is that right?
Essentially its working as a site search with some defined rules, so you'd have to think about the following:
404's - if your pages are built on queries, a user could potentially create a link to a page domain.com/brown/jackets/f**k/, obviously this would need to return a 404 or this could be indexed.
Canonical - the previosu point brings about dudplicate content issues, if you don't define the rules then you could have /brown/jackets/levis/ vs. /brown/levis/jackets/ showing the same content, so there needs to be some canonical management.
So you can use ajax to call a particular URL, you would need to use the filters as links ( <a>tags) so the pages are crawled and pull that content (search results) back, the url wouldn't changes though unfortunately, but it would still be linked to from the</a> <a>tag.</a>
<a>Is there a specific reason why you don't want to use a hashbang?</a>
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
AJAX requests and implication for SEO
Hi, I got a question in regard to webpages being served via AJAX request as I couldn't find a definitive answer in regard to an issue we currently face: When visitors on our site select a facet on a Listing Page, the site doesn't fully reload. As a consequence only certain tags of the content (H1, description,..) are updated, while other tags like canonical URLs, meta noindex,nofollow tag, or the title tag are not updating as long as you don't refresh the page. We have no information about how this will be crawled and indexed yet but I was wondering if anyone of you knows, how this will impact SEO?
Intermediate & Advanced SEO | | FashionLux0 -
Using a US CDN (Cloudflare) for a UK Site. Should I use a UK Based CDN as it says my server is based in USA
Hi All, We are a UK Company with Uk customers only and use CloudFlare CND. Our Site is hosted by a UK company with servers here but from looking online and checking where my site is hosted etc etc , some sites are telling me the name of our UK Hosted company and other sites are telling me my site is hosted in San Fran (USA) , where I presume the Cloudflare is based. I know Cloudflare has a couple of servers in the UK it uses but given all my customers are UK based ,I don't want this is affect rankings etc , as I thought it was a ranking benefit to be hosted in the country you are based. Is there any issue with this and should I change or is google clever enough to know so i shouldn't worry. thanks Pet
Intermediate & Advanced SEO | | PeteC120 -
Weird 404 URL Problem - domain name being placed at end of urls
Hey there. For some reason when doing crawl tests I'm finding pages with the domain name being tacked on the end and causing 404 errors.
Intermediate & Advanced SEO | | Jay328
For example: http://domainname.com/page-name/http://domainname.com This is happening to all pages, posts and even category type 1. Site is in Wordpress
2. Using Yoast SEO plugin Any suggestions? Thanks!0 -
Is Using a Question, Answer Format Appropriate for a Blog? Is a 300 Word Micro Blog An SEO Plus?
My PR agency has suggested a question answer format be incorporated in my blog. They suggest a microblog with a single sentence question and an answer of about 300 words. My blog currently has about 35 posts. I would like to ramp up blog entries to about one or two per week of these "mini blog" posts. The format of the new blog begins as a question with the responses being paragraphs that do not use headings. My concerns are as follows: 1. No headings in an answer of 300 words will fail to provide Google with context regarding the content's meaning. Everything I have read about SEO suggests text be broken up in short sections and that it be divided by headings (preferably H2s). I very much like my agency's concept for a question answer format blog. It provides very practical info for visitors. How can I use it in a manner that supports SEO best practices? 2. According to a reputable SEO firm that has been assisting me, Google does not consider a blog post of less than 600 words to be superior quality. They told me that blog posts of 300 words, from an SEO purpose will not be a great helpful, that the content will not be rich enough to generate incoming links. Is this really the case? What if this abbreviated content is very well written and engaging? If so, is 300 words sufficient? From the visitor's perspective I am not sure they would have the patience to read 600 words when 300 words is more than than enough to answer these basic questions. From a PR perspective I think the shorter content in a question answer format is superior at least for my line of business (commercial real estate brokerage). 3. If 500-600 words is the minimum word count, and headings are necessary, what is the best way to execute a question and answer blog format? The purpose of this blog is to provide very useful info to my visitors while generating incoming links to that will boast my rankings. Thanks in advance for your feedback!!! Alan
Intermediate & Advanced SEO | | Kingalan10 -
Do UTM URL parameters hurt SEO backlink value?
Does www.example.com and www.example.com/?utm_source=Google&utm_medium=Press+Release&utm_campaign=Google have the same SEO backlink value? I would assume that Google knows the difference.
Intermediate & Advanced SEO | | mkhGT0 -
Overly-Dynamic URL
Hi, We have over 5000 pages showing under Overly-Dynamic URL error Our ecommerce site uses Ajax and we have several different filters like, Size, Color, Brand and we therefor have many different urls like, http://www.dellamoda.com/Designer-Pumps.html?sort=price&sort_direction=1&use_selected_filter=Y http://www.dellamoda.com/Designer-Accessories.html?sort=title&use_selected_filter=Y&view=all http://www.dellamoda.com/designer-handbags.html?use_selected_filter=Y&option=manufacturer%3A&page3 Could we use the robots.txt file to disallow these from showing as duplicate content? and do we need to put the whole url in there? like: Disallow: /*?sort=price&sort_direction=1&use_selected_filter=Y if not how far into the url should be disallowed? So far we have added the following to our robots,txt Disallow: /?sort=title Disallow: /?use_selected_filter=Y Disallow: /?sort=price Disallow: /?clearall=Y Just not sure if they are correct. Any help would be greatly appreciated. Thank you,Kami
Intermediate & Advanced SEO | | dellamoda2 -
Removing Dynamic "noindex" URL's from Index
6 months ago my clients site was overhauled and the user generated searches had an index tag on them. I switched that to noindex but didn't get it fast enough to avoid being 100's of pages indexed in Google. It's been months since switching to the noindex tag and the pages are still indexed. What would you recommend? Google crawls my site daily - but never the pages that I want removed from the index. I am trying to avoid submitting hundreds of these dynamic URL's to the removal tool in webmaster tools. Suggestions?
Intermediate & Advanced SEO | | BeTheBoss0 -
How to deal with old, indexed hashbang URLs?
I inherited a site that used to be in Flash and used hashbang URLs (i.e. www.example.com/#!page-name-here). We're now off of Flash and have a "normal" URL structure that looks something like this: www.example.com/page-name-here Here's the problem: Google still has thousands of the old hashbang (#!) URLs in its index. These URLs still work because the web server doesn't actually read anything that comes after the hash. So, when the web server sees this URL www.example.com/#!page-name-here, it basically renders this page www.example.com/# while keeping the full URL structure intact (www.example.com/#!page-name-here). Hopefully, that makes sense. So, in Google you'll see this URL indexed (www.example.com/#!page-name-here), but if you click it you essentially are taken to our homepage content (even though the URL isn't exactly the canonical homepage URL...which s/b www.example.com/). My big fear here is a duplicate content penalty for our homepage. Essentially, I'm afraid that Google is seeing thousands of versions of our homepage. Even though the hashbang URLs are different, the content (ie. title, meta descrip, page content) is exactly the same for all of them. Obviously, this is a typical SEO no-no. And, I've recently seen the homepage drop like a rock for a search of our brand name which has ranked #1 for months. Now, admittedly we've made a bunch of changes during this whole site migration, but this #! URL problem just bothers me. I think it could be a major cause of our homepage tanking for brand queries. So, why not just 301 redirect all of the #! URLs? Well, the server won't accept traditional 301s for the #! URLs because the # seems to screw everything up (server doesn't acknowledge what comes after the #). I "think" our only option here is to try and add some 301 redirects via Javascript. Yeah, I know that spiders have a love/hate (well, mostly hate) relationship w/ Javascript, but I think that's our only resort.....unless, someone here has a better way? If you've dealt with hashbang URLs before, I'd LOVE to hear your advice on how to deal w/ this issue. Best, -G
Intermediate & Advanced SEO | | Celts180