New folder structure
-
We are in the process of relaunching one of our website's that will use a totally need folder structure.
Previously we used mydomain.com/content/country/region/city/district/hotel_name/
Now we are changing to make the URL shorter, more precise - since we are using a new CMS, to be mydomain.com/gb_Hotel-Name/
My question is currently we've in the region of 10,000 pages indexed in Google. So we are going to have to create 301 permanent redirects from the old URLs to the new URLs.
From your previous experience, is this the correct way of approaching the task.
-
Hi Neil.
The sites we have done are all new but from experience dealing with moving to new urls the best thing to do is create a mapping document in excel. It'squite easy if you know that for example:
www.domain.com/berlin-hotel is moving to www.domain.com/de/berlin-hotel.
Then all you need to do is put in the 301s based on the mapping and monitor WMT for issues - you will always miss something.
From what you are saying however there is no logical structure to your site - which will make this harder. I have had to deal with this in the past too, you might just need to identify all of the more important pages and 301 these first and go via mechanical turk of get an intern in or something to just plough away and find all of your urls.
If you have an XML sitemap you should be able to get them all pretty quickly and map from here.
Hope this helps.
-
Hi ASR, can you advise more on how you moved to the new folder structure ? Did you leave old pages in place, launch new site, then go through all old URLs and 301 them to the new URL ?
The problem I have is the new folder structure is not a standard pattern, so any 301 will need to be done manually.
-
either way, it would not hurt to 301 all.
-
Ooops, should have said link juice.
As you can all see Alan and I have different views on this but at least you have a range of views Tommo!
Good luck - hope it all goes well.
-
"A 301 will pass 80% of the page authority over."
No a 301 will pass 85% of its incomming link juice, you are redireceting the inomming links not the page
A 301, passes 85% of its PR from the page to another page though a link from the page.
But if the page no longer exists it can not. Once you delete that page, it no longer has any page rank, but the pages that may link to the old page are still passing PR, so in that case you can redirect those links to the new page. But you cannot redirect the non existing page or any authority it had to the new page.
Read Google’s algorithm, I assure that is not how it works.
http://www.webworkshop.net/pagerank.html
By the way, all links pass 85% of link juice it has though its links, if the link hits a 301 redirect, it loses 15% again. This is so you don’t get infinite loops.
Imagine if you could move a page and somehow gets its PR by doing a 301 redirect. You could keep moving it and it would keep getting more PR.
If I could use an analogy
If you have a store and Bob has a store, and Bob closes down, but puts a sign on his door telling his customers to go to your store, you will get his customers, but you don’t get his stock. And if he had no customers you get nothing.
Anyhow if you read the link you will see that I am correct -
A 301 will pass 80% of the page authority over. You also don't leave any 404s.
I'm not sure why you are so against 301s? It's tidier, "best practice" and not hard to do. Why risk missing something out that might prove the difference?
-
They will reindex them, you dont need to 301 then to get them reindexed.
But they dont have any page rank if the no longer exist. you can only pass link juice though a link, if the page no longer exists then you can not link from it, and there is no link juce juice to pass.
If that were true, you could keep moving a page and its link juice would keep rising.
-
I would 301 all the pages; it's "best practice," IMO.. The old pages had juice just from the fact that they existed; ie. inbound links aren't the only value you want to preserve. And, as ASR points out, best to avoid the 404s. The redirects are not just for the PageRank, but to allow search engines to easily reindex the content.
-
"The pages will still have some authority even without links"
they would if they were still there, but they no longer are, its just a ref to them in the index,
sure if they have some like or links, I agree
If they are in another lingo, they maybe, I dont know, but if they are in the same lingo, I still say they will be duplicate content
-
The pages will still have some authority even without links, so I would definitely pass any little bit there. Social signals too if there are any that might not have been picked up. And I would also worry that while some tools say there are no links, I don't know one 100% reliable tool to tell you page X has no links.
If the local sites are in DE, FR, ES, or whatever, then they are not duplicate, they are local language. IBM, Apple to name a couple certainly do this route also.
Matt Cutts may say that, but we certainly do not suffer from this problem in the least. Another Cutts "we do this but really don't" comment maybe?!
-
I agree it would of worked, but if they had no incomming links, then it was un-necessary.
-
Why must you 301 it, if they have no in comming links. There is no link juice to keep.
As for theTLD's, if they are duplicate content websites, which I assume they are, then you will have duplicate content problems. How will you get around that?
If yopu listen to the Matt Cutts video again, he says at the beginging, if the are all on the same TLD you will be pinged for DC
-
Disagree with you on that one Alan. We have no issue with duplicate content and it is also what everyone (including those at MOZcon this year) recommend.
In fact I have an email from SEOMoz themselves recommending it.
TLD will get better over time but from experience running 7 sites (6 country sites) I would only ever use folders now.
I also disagree with your comments below about not 301ing all old content. You must do this when re-launching.
-
I have to disagree, using the TLD's you dont need to worry about duplicate content.
See Matt Cutts
-
You are having to link build to multiple sites, they are start from zero in a search engines eyes (no matter how strong your current site is), your marketing materials cannont just mention domain.com etc.
Our use of folders with our strong domain strength has seen them launch and a month later be at the top of local search engines for hard to rank for search terms.
I cannot recommend strongly enough that going down the folder route is much better for SEO.
-
Yes, the other .de site for example will be 100% in the local language. What pitfalls do you feel we will have if we went the route of many TLD sites ?
-
Yes that is what I am saying. Definitely head down the domain.com/country-folder ie domain.com/uk
This has a lot of advantages around marketing, all of the links to this one domain help strengthen the entire site instead of having to having to link building to a number of new sites (which are starting from scratch in Google's eyes).
And you can still target them in GWMT by country too - which will definitely work. I am presuming they will be in local language as well?
Hope this helps.
A
-
Currently our domain is domain.com.
But we were thinking about having domain.de, domain.com.cn, domain.es etc. Are you saying this isn't the way to go ? We were looking to host these sites in the TLD country e.g domain.de hosted in Germany.
Examples where this has worked very well is TripAdvisor.
-
Okay. That makes sense but I would stick to one TLD if you could as you can run with folders and all of the country country sites (in the folder) will inherit all of the your one TLD strenght, makes things like link building easier too.
I would strongly recommend this. We have 6 international sites now and three are TLD (before I got here) and three folders and all of the folder sites are doing so much better than the TLD - the difference is amazing.
Even if you go with the new TLD I would miss out the GD_ bit - you don't need that at all if you have a TLD and makes no sense to me. I would rather have domain.com/city-hotels/hotel-name and optimise around this. You can then have landing pages around city hotels ie berlin hotels and pull in traffic this way.
Just a thought!
-
For the new site we will be having international TLD like mydomain.de. Re the GB this is so we initially know what country the hotel is located in, and in some instances hotels have the same name, but in different countries.
-
I would 301 all of the pages - to relevant new pages otherwise you will end up with a heap of 404s too if the old content just disappears. Blogs etc will have linked to them etc, you want to make the US as good as possible.
Mapping out the 301s will take time but be worth it in the long run.
I have done a website with 500,000 pages and mapped it and it worked well.
A
-
i would only bother 301 redirecting those that had incomming links or you will have a mess on your hands.
If you take a short cut and try to 301 on mass to the one page, bing for one will just ignore them as far as link juice goes.
also up untill recenly Matt Cutts said that _ or - were a matter of choice, but now he has stated to use -, cant remember the reason.
After 301ing all pages with in-links, i would just go to google and and remove site (enter root domain). This way yopu get rid of all the old pages and aviod duplicate content.
Others will proabaly not agree. but this is how i do it. Block the site with robots.txt, then remove site in GWMT by entering the domain name in remove url.
Wait a day till its removed, then remove robots.txt block.dont woory, you need to leave a block in place for 90 days to really remove from index. doing this all the pages that still exist will be back in a few days, the rest that 404 will be removed from index.
-
Hi Tommo.
Sounds like the clean up is a great idea. I wonder if you still need the GB at all? Is this for language or country? If you are doing country/language I would have a folder for each ie
domain/en/us/title-of-hotel
You can then geo-target these to country specific areas in GWMT as well - which would be helpful.
And definitely 301 them all - this is a must.
A
-
Yes this is correct, the 301 redirect will forward all juice to the new structure and Google will find this change pretty fast.
If you have a sitemap you could also submit that map to google through webmaster tools, that would update the index faster. Good luck!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Job Posting Page and Structured Data Issue
We have a website where we do job postings. We manually add the data to our website. The Job Postings are covered by various other websites including the original recruiting organisations. The details of the job posting remain the same, for instance, the eligibility criteria, the exam pattern, syllabus etc. We create pages where we list the jobs and keep the detailed pages which have the duplicate data disallowed in robots.txt. Lately, we have been thinking of indexing these pages as well, as the quantum of these non-indexed pages is very high. Some of our competitors have these pages indexed. But we are not sure whether doing this is gonna be the right move or if there is a safe way to deal with this. Additionally, there is this problem that some job posts have very less data like fees, age limit, salary etc which is thin content so that might contribute to poor quality issue. Secondly, we wanted to use enriched result snippets for our job postings. Google doesn't want snippets to be used on the listing page: "Put structured data on the most detailed leaf page possible. Don't add structured data to pages intended to present a list of jobs (for example, search result pages). Instead, apply structured data to the most specific page describing a single job with its relevant details." Now, how do we handle this situation? Is it safe to allow the detailed pages which have duplicate job data and sometime not so high quality data in robots.txt?
Intermediate & Advanced SEO | | dailynaukri0 -
How do I get more video reviews on a new site?
Hi, I'm working on a site that has video reviews of various places. It's general information/experience that most people possess and production-wise they are selfies. These videos are then transcribed and, voila... searchable content. My problem is this... how do I get large numbers of people to go to the trouble to make a 2 minute selfie? I thought about HARO, since one could work in a plug for something, but they have a site traffic threshold that this new site isn't at. Any and all ideas on how to efficiently generate this content for a new site with very little traffic would be appreciated. Thanks!
Intermediate & Advanced SEO | | 945010 -
Measure impact from new meta descriptions
Hi guys, I'm looking to implement new meta descriptions across a site and i want to measure the impact. So far I'm thinking of extracting the CTR data from GWT for the last 90 days to get the most accurate CTR averages for each URL. Then once the new meta descriptions have been implemented, compare the CTR with the old CTR averages accross URLs. Do you think this would be the most accurate way of measuring the impact? Cheers, Chris
Intermediate & Advanced SEO | | jayoliverwright1 -
Changing URL structure of date-structured blog with 301 redirects
Howdy Moz, We've recently bought a new domain and we're looking to change over to it. We're also wanting to change our permalink structure. Right now, it's a WordPress site that uses the post date in the URL. As an example: http://blog.mydomain.com/2015/01/09/my-blog-post/ We'd like to use mod_rewrite to change this using regular expressions, to: http://newdomain.com/blog/my-blog-post/ Would this be an appropriate solution? RedirectMatch 301 /./././(.) /blog/$1
Intermediate & Advanced SEO | | IanOBrien0 -
E commerce canonical links: include category structure?
I have a client on shopify. All categories have correct canonical links. however, the links from all menus, category pages, etc. follow this structure: /collections/COLLECTION_NAME/products/PRODUCT_NAME but the canonical link on the above product url is: /products/PRODUCT_NAME I have a feeling this is hurting our product detail page's seo. Our collection pages are ranking fine, but for some reason the detail pages aren't. It could be that they are deeper, but I am trying to make sure nothing big is causing it first before I get into the smaller factors. Any best practices on this?
Intermediate & Advanced SEO | | no6thgear0 -
Making sense of MLB.com domain structure
Although the subject of subdomains has been discussed quite often on these boards, I never found a clear answer to something I am pondering. I am about to launch a network of 8 to 10 related sites - all sharing a the same concept, layout, etc. but each site possessing unique content. My concept will be somewhat similar to how MLB.com (Major League Baseball) is set up. Each of the 30 teams in the league has it's unique content as a subdomain. My goal in the initial research was to try to find the answer to this question - **"Do the subdomains of a network contribute any increased value to the Root Domain? ** As I was trying to find the answer to my question and analyzing how MLB.com did it, I began to notice some structure that made very little sense to me and am hoping an expert can explain why they are doing it the way they are. Let me try to illustrate: Root Domain = http://mlb.com (actually redirects to: http://mlb.mlb.com/index.jsp) This root domain serves universal content that appeals to all fans of the league and also as a portal to the other subdomains from the main navigation. SubDomain Example = http://tampabay.rays.mlb.com/index.jsp **Already there are a couple of questions. ** 1. Why does MLB.com redirect to http://mlb.mlb.com/index.jsp ? - why the mlb subdomain? 2. - Why two subdomains for tampabay.rays.mlb.com/index.jsp.? Why not just make the subdomain "tampabayrays", "newyorkmets", "newyorkyankees" etc. **Here is where things get a little more complicated and confusing for me. **
Intermediate & Advanced SEO | | bluelynxmarketing
From the home page, if I click on an article about the San Francisco Giants, I was half expecting to be led to content hosted from the http://sanfrancisco.giants.mlb subdomain but instead the URL was: http://mlb.mlb.com/news/article.jsp?ymd=20121030&content_id=40129938&vkey=news_mlb&c_id=mlb I can understand the breakdown of this URL
YMD = Year, Month, Date
Content ID = Identifying the content
VKey = news_MLB (clicked from the "news section found from the mlb subdomain.
c_id=mlb (?) Now, if I go to the San Francisco Giants page, I see a link to the same exact article but the URL is this: http://sanfrancisco.giants.mlb.com/news/article.jsp?ymd=20121030&content_id=40129938&vkey=news_sf&c_id=sf It get's even stranger...when I went to the Chicago Cubs subdomain, the URL to the same exact article does not even link to the general mlb.mlb.com content, instead the URL looks like this: http://chicago.cubs.mlb.com/news/article.jsp?ymd=20121030&content_id=40129938&vkey=news_mlb&c_id=mlb When I looked at the header from the http://chicago.cubs.mlb.com ULR, I could see the OG:URL as: http://sanfrancisco.giants.mlb.com/news/article.jsp?ymd=20121030&content_id=40129938&vkey=news_sf&c_id=sf but I did not see anything relating to rel=canonical I am sure there is a logical answer to this as the content management for a site like MLB.COM must be a real challenge. However, it seems that they would have some major issues with duplicate content. So aside from MLB's complex structure...I am also still searching for the answer to my initial question which is - **"Do the subdomains of a network contribute any increased value to the Root Domain?" For example, does http://tampabay.rays.mlb.com/index.jsp bring value to http://mlb.com? And what if the subdomain is marketed as http://raysbaseball.com and then redirected to the subdomain? Thanks in advance. **0 -
New Site Launch- Bad Rankings
So I am slightly stumped. I work on a few eCom sites and over the past few months we have been launching new products as well as some site updates/upgrades. The issue that I am having is this. Product that is actually new this season, as in we did not carry it before, is ranking fine...generally speaking we are top 3 in Google with Brand/Product name. However product that is not new, meaning it has a new color or graphic but has been in the catalog for years is not ranking. In Prior years the URL's ended with the product name, now however they have additional information added to the end for this season. Would not directing the old URL to the new URL effect the rankings of these pages? I am open to suggestions beyond redirecting as well. Thank you!
Intermediate & Advanced SEO | | K2_Sports0 -
New Website.. From where to start with a SEO Plan ?
Hi.. We have got two SEO and link builders working here in the office, putting in efforts is not a problem. And we have been doing SEO for our websites since a long time and they all rank well. Since, I have signed up for SEOMOZ, which is the one stop with all the heavy duty SEO guys, so we would like to get some suggestions from the experts. Ok here is the question. We have launched a new website this week and we have got a new SEO client, whose website is 6 months old. 1. What should be a proper SEO plan for a new website ? Got any Plan PDF or Buleprint, which you can send me ? 2. For the client website , he is ranking on 4-7th pages for his 4 keywords. Whats should be the correct way or plan to start doing SEO for them and get them to the first page ? 3. I feel like, doing competitive analysis is lacking with us and we are not good at evaluating that step. Whats the bet way to do competitive analysis ? 4. We are looking to build more and more links to the website . Whats the best way to make links ? Got any SEOMoz Blog post link, which explain this better or any forum thread which has got the tricks for this ?
Intermediate & Advanced SEO | | qubesys0