Blog for a Prestashop: WP at subfolder or blog with dedicated keyword based domain?
-
Hi
Scenario 1: site running with prestashop at domain whatever.com There's a blog at a subfolder of domain, running with Wordpress whatever.com/blog
Scenario 2: site running with prestashop at whatever.com. The blog is at a different server and domain with keywords, e.g. cloudtechnology.com
What's better for the blog posts to have better SERP results? Scenario 1 or 2? From the point of view of a user, I have seen for friends and colleagues that scenario 2 works for the blog, and might work for the shop, since blog posts contain links to shop etc. However, all examples I've seen for scenario 1 are still struggling...it's hard for them to make the blog visible at SERPs...is it a coincidence or is there a reason?
Thanks for your answers!
-
If you go with Scenario 1 you will give your shop domain the following benefits:
-If frequently updated, fresh content (remember: freshness is something Google likes so much)
-Improved internal linking
-All resources in the same domain: products, relevant info, etc.
-Great landing pages
-Posts are more likely to be shared than products, so take advantage of social media to publicize your store through the blog.
One final note: you are planning to build your blog with WordPress. I'm absolutely in love with WordPress, but PrestaShop gives you the chance to use own modules to have this feature. I give you some examples:
http://addons.prestashop.com/en/content-management/1213-blog-module.html
http://www.themes-prestashop.com/es/modulos-gratis/31-prestashop-blog.html
http://www.buhoc.com/prestashop/es/modulos-gratuitos-prestashop/44-modulo-blog-prestashop.html
Hope it helps.
Sergio
-
I would definitely go with scenario 1. Whenever someone links to your blog post it gives you site the authority rather than an external URL. This way you can also build allot of internal links from your posts to your pages or other posts. Also you want people to like your posts from your site/share the posts from your site.
Think about it like this.. If a post was to go viral would you rather thew article be on:
or
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Blog separate from Website
One of my clients has a well established website, and a well established blog - each with its own domain. Is there any way to move the blog to his website domain without losing the SEO and links that he has built up over time?
Technical SEO | | EchelonSEO0 -
Do bad links to a sub-domain which redirects to our primary domain pass link juice and hurt rankings?
Sometime in the distant past there existed a blog.domain.com for domain.com. This was before we started work for domain.com. During the process of optimizing domain.com we decided to 301 blog.domain.com to www.domain.com. Recently, we discovered that blog.domain.com actually has a lot of bad links pointing towards it. By a lot I mean, 5000+. I am curious to hear people's opinions on the following: 1. Are they passing bad link juice? 2. does Google consider links to a sub-domain being passed through a 301 to be bad links to our primary domain? 3. The best approach to having these links removed?
Technical SEO | | Shredward0 -
How to prevent duplicat content issue and indexing sub domain [ CDN sub domain]?
Hello! I wish to use CDN server to optimize my page loading time ( MaxCDN). I have to use a custom CDN sub domain to use these services. If I added a sub domain, then my blog has two URL (http://www.example.com and http://cdn.example.com) for the same content. I have more than 450 blog posts. I think it will cause duplicate content issues. In this situation, what is the best method (rel=canonical or no-indexing) to prevent duplicate content issue and prevent indexing sub domain? And take the optimum service of the CDN. Thanks!
Technical SEO | | Godad0 -
Exact match domain
Will buying an exact match domain and redirecting it to our main site a good idea, if such a domain is available ? What are the pros and cons ? Are exact match domains still powerful for ranking purpose ?
Technical SEO | | seoug_20050 -
URL paths and keywords
I'm recommending some on-page optimization for a home builder building in several new home communities. The site has been through some changes in the past few months and we're almost starting over. The current URL structure is http://homebuilder.com/oakwood/features where homebuilder = builder name Oakwood Estates= name of community features = one of several sub-paths including site plan, elevations, floor plans, etc. The most attainable keyword phrases include the word 'home' and 'townname' I want to change the URL path to: http://homebuilder.com/oakwood-estates-townname-homes/features Is there any problem with doing this? It just seems to make a lot of sense. Any input would be appreciated.
Technical SEO | | mikescotty0 -
Parked Domain blog directory not redirecting
My newly parked domain name, (our main website had to switch primary domains) is not redirecting properly and is causing our blog to be duplicate content. My 301 redirects work for everything else, but our parked domain /blog directory is not redirecting. I can type in both urls and then the blog appears on both sites. Not good. If I delete my blog .htaccess file, then it redirects fine. However, then our blog links are broken. So it has to do something with our .htaccess files. I do have a .htaccess file for our website, saying redirect everything to correct location, so i think this is interfering, but I cannot pinpoint it. this is the .htaccess file for the blog. BEGIN WordPress <ifmodule mod_rewrite.c="">RewriteEngine On
Technical SEO | | hfranz
RewriteBase /blog/
RewriteRule ^index.php$ - [L]
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule . /blog/index.php [L]</ifmodule> END WordPress main sites .htaccess (i am trying to pinpoint the issue here) Options +Includes
AddType text/html .htm .html
AddHandler server-parsed .htm .html
Options +FollowSymLinks RewriteEngine on RewriteBase / RewriteCond %{HTTP_HOST} ^(www.)?parkeddomain.com [NC,OR] RewriteCond %{HTTP_HOST} ^newdomain.com [NC] RewriteRule (.*) http://www.newdomain/$1 [R=301,L] RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /([^?]*)? RewriteRule (.*) /$1? [R=301,L] RewriteCond %{THE_REQUEST} ^.*/index.php RewriteRule ^(.*)index.php$ http://www.newdomain.com/$1 [R=301,L] RewriteCond %{THE_REQUEST} ^.*/index.htm RewriteRule ^(.*)index.htm$ http://www.newdomain/$1 [R=301,L] RedirectMatch 301 /index.php/(.*) /$1 Is there something obvious here, that does not look right?0 -
Multiple Domains, Same IP address, redirecting to preferred domain (301) -site is still indexed under wrong domains
Due to acquisitions over time and the merging of many microsites into one major site, we currently have 20+ TLD's pointing to the same IP address as our "preferred domain:" for our consolidated website http://goo.gl/gH33w. They are all set up as 301 redirects on apache - including both the www and non www versions. When we launched this consolidated website, (April 2010) we accidentally left the settings of our site open to accept any of our domains on the same IP. This was later fixed but unfortunately Google indexed our site under multiple of these URL's (ignoring the redirects) using the same content from our main website but swapping out the domain. We added some additional redirects on apache to redirect these individual pages pages indexed under the wrong domain to the same page under our main domain http://goo.gl/gH33w. This seemed to help resolve the issue and moved hundreds of pages off the index. However, in December of 2010 we made significant changes in our external dns for our ip addresses and now since December, we see pages indexed under these redirecting domains on the rise again. If you do a search query of : site:laboratoryid.com you will see a few hundred examples of pages indexed under the wrong domain. When you click on the link, it does redirect to the same page but under the preferred domain. So the redirect is working and has been confirmed as 301. But for some reason Google continues to crawl our site and index under this incorrect domains. Why is this? Is there a setting we are missing? These domain level and page level redirects should be decreasing the pages being indexed under the wrong domain but it appears it is doing the reverse. All of these old domains currently point to our production IP address where are preferred domain is also pointing. Could this be the issue? None of the pages indexed today are from the old version of these sites. They only seem to be the new content from the new site but not under the preferred domain. Any insight would be much appreciated because we have tried many things without success to get this resolved.
Technical SEO | | sboelter0 -
External Links from own domain
Hi all, I have a very weird question about external links to our site from our own domain. According to GWMT we have 603,404,378 links from our own domain to our domain (see screen 1) We noticed when we drilled down that this is from disabled sub-domains like m.jump.co.za. In the past we used to redirect all traffic from sub-domains to our primary www domain. But it seems that for some time in the past that google had access to crawl some of our sub-domains, but in december 2010 we fixed this so that all sub-domain traffic redirects (301) to our primary domain. Example http://m.jump.co.za/search/ipod/ redirected to http://www.jump.co.za/search/ipod/ The weird part is that the number of external links kept on growing and is now sitting on a massive number. On 8 April 2011 we took a different approach and we created a landing page for m.jump.co.za and all other requests generated 404 errors. We added all the directories to the robots.txt and we also manually removed all the directories from GWMT. Now 3 weeks later, and the number of external links just keeps on growing: Here is some stats: 11-Apr-11 - 543 747 534 12-Apr-11 - 554 066 716 13-Apr-11 - 554 066 716 14-Apr-11 - 554 066 716 15-Apr-11 - 521 528 014 16-Apr-11 - 515 098 895 17-Apr-11 - 515 098 895 18-Apr-11 - 515 098 895 19-Apr-11 - 520 404 181 20-Apr-11 - 520 404 181 21-Apr-11 - 520 404 181 26-Apr-11 - 520 404 181 27-Apr-11 - 520 404 181 28-Apr-11 - 603 404 378 I am now thinking of cleaning the robots.txt and re-including all the excluded directories from GWMT and to see if google will be able to get rid of all these links. What do you think is the best solution to get rid of all these invalid pages. moz1.PNG moz2.PNG moz3.PNG
Technical SEO | | JacoRoux0