Best practices for types of pages not to index
-
Trying to better understand best practices for when and when not use a content="noindex". Are there certain types of pages that we shouldn't want Google to index? Contact form pages, privacy policy pages, internal search pages, archive pages (using wordpress). Any thoughts would be appreciated.
-
Certainly! When it comes to SEO (Search Engine Optimization), there are certain types of pages that you may want to prevent search engines from indexing. This can help ensure that only your most relevant and valuable content is displayed in search engine results. Here are some best practices for types of pages not to index:
Duplicate Content Pages:
Avoid indexing pages with duplicate content, as search engines prefer unique content.
Use canonical tags to indicate the preferred version of a page.
Thin or Low-Quality Content Pages:Pages with little to no valuable content may harm your site's overall SEO.
Consider adding substantial content to these pages or use meta tags to prevent indexing.
Internal Search Results Pages:Exclude internal search results pages from indexing, as they may lead to a poor user experience in search results.
Use the robots.txt file to disallow crawling of these pages.
Thank You and Confirmation Pages:Pages that users see after completing a form submission or transaction may not provide significant value to search engine users.
Use the noindex meta tag to prevent indexing of thank you and confirmation pages.
Login and Account Pages:Secure pages containing login forms or user account information to prevent unauthorized access.
Use the robots.txt file to disallow crawling of these pages.
Tag and Category Pages:Depending on your content management system, tag and category pages may be automatically generated. These can sometimes result in duplicate content issues.
Use the noindex meta tag or canonical tags as appropriate.
Paginated Pages:For large sets of paginated content, consider only indexing the main paginated page and using the rel="next" and rel="prev" tags to indicate the paginated structure.
Privacy Policy and Terms of Service Pages:While it's important to have these pages, they might not need to be indexed.
Use the noindex meta tag if you don't want search engines to index these legal pages.
Media Files and Non-HTML Content:Files like PDFs, images, and other non-HTML content may not need to be indexed.
Use appropriate meta tags or header directives to prevent indexing.
Test and Development Pages:Pages used for testing or development purposes should not be indexed.
Use authentication or the robots.txt file to block search engine bots from accessing these pages.
Always keep in mind that SEO best practices evolve, and it's essential to stay updated with the latest recommendations from search engines. Regularly check your website's performance in search engine results and adjust your indexing strategy accordingly. -
Best practices for determining which types of pages not to index involve strategic decisions to enhance the overall performance and relevance of your website on search engines. Here are some key considerations:
Thin or Low-Quality Content:
Recommendation: Identify and exclude pages with thin or low-quality content that doesn't provide substantial value to users. Focus on creating high-quality, informative content that aligns with user intent.
Duplicate Content:
Recommendation: Avoid indexing pages with duplicate content, as it can lead to confusion for search engines and may result in lower rankings. Use canonical tags to specify the preferred version of the content.
Internal Search Result Pages:
Recommendation: Exclude internal search result pages from indexing, as they often lead to duplicate content issues. Ensure that search engines focus on the primary content pages of your site.
Archive or Staging Pages:
Recommendation: Prevent search engines from indexing archive or staging pages. Use robots.txt or meta tags to disallow indexing of such pages to maintain the integrity of your live content.
Thank You and Confirmation Pages:
Recommendation: Non-essential pages like thank you or confirmation pages for form submissions may not need indexing. Exclude these pages to avoid unnecessary clutter in search engine results.
Login or Session-Specific Pages:
Recommendation: Exclude pages that require user authentication or are session-specific. This prevents search engines from indexing content that's not meant for public access.
Paginated Pages:
Recommendation: For paginated content, consider using rel="next" and rel="prev" tags to signal the relationship between pages. This helps search engines understand the structure without indexing each individual page.
Category or Tag Pages:
Recommendation: Depending on your website structure, category or tag pages may not need indexing. Ensure that these pages don't dilute the overall relevance of your site and use noindex tags if necessary.
Privacy Policy, Terms of Service, and Legal Pages:
Recommendation: While important for compliance, legal and policy pages may not require indexing in search results. Use noindex tags for these pages, allowing them to serve their purpose without being prominent in search listings.
Dynamic URLs with Parameters:
Recommendation: Exclude dynamically generated pages with URL parameters that don't represent unique content. Utilize canonical tags or parameter handling in Google Search Console to manage these pages.
Unnecessary Media or File Attachment Pages:
Recommendation: Media or file attachment pages may not need indexing. Use noindex tags to prevent these pages from appearing in search results while still providing access to the media itself.
Regularly audit and monitor your site's performance in search engine results to ensure that the selected pages for non-indexing align with your SEO strategy and user experience goals. Always consider the specific needs and structure of your website when implementing these best practices.
Read My Recent post here : PBU in Football
-
Here are some best practices for types of pages not to index:
Duplicate content pages. If you have multiple pages with the same or similar content, it's generally a good idea to avoid indexing all of them. This could include printer-friendly versions, alternate language versions, or slight variations of the same content.
Thin or low-quality content pages. Pages with little or no content, or pages with content that is poorly written or irrelevant to your target audience, should not be indexed.
Internal search results pages. These pages are typically not meant for users to land on directly, and they can clutter up search engine results pages (SERPs).
Privacy and policy pages. These pages are typically not relevant to search users, and they can contain sensitive information that should not be indexed.
Thank-you pages. These pages are typically displayed after a user submits a form or makes a purchase, and they are not meant for indexing.
Login and checkout pages. These pages are typically not relevant to search users, and they can contain sensitive information that should not be indexed.
Staging or test pages. These pages are not meant to be seen by the public, and they can clutter up SERPs.
Paginated pages. If your paginated pages contain the same content as your main product or category pages, you may want to consider noindexing them. This will help to avoid duplicate content issues.
You can use a number of methods to prevent pages from being indexed, including:
Robots.txt file. You can use your robots.txt file to block search engines from crawling certain pages on your website.
Noindex meta tag. You can add a noindex meta tag to the header of a page to prevent it from being indexed.
Canonical tags. You can use canonical tags to specify which version of a page is the preferred version for search engines. This can be helpful for preventing duplicate content issues.
It's important to note that noindexing pages is not always necessary. For example, if you have a blog with a lot of high-quality content, you may want to index all of your pages, even if they have similar content. However, if you have a lot of low-quality or irrelevant pages on your website, it's a good idea to noindex them to avoid harming your SEO.
-
Thanks to all of you for sending me valuable information.
-
To prevent specific types of web pages from being indexed by search engines, follow these best practices: Use the robots.txt file to disallow indexing for entire sections of your website or directories. On individual pages, utilize the meta robots tag to specify "noindex" or "nofollow" directives. Employ the X-Robots-Tag HTTP header to communicate indexing preferences, either at the server level or on specific pages. Password-protect pages that should be accessible only to authorized users. Implement canonical tags to indicate the preferred version of a page. Include only desired pages in your XML sitemap. Maintain a clean URL structure, and use "noindex" directives in robots meta headers for dynamic or user-generated content. For pages you want completely removed from search results, return 404 or 410 HTTP status codes. Regularly monitor indexed pages using tools like Google Search Console to ensure compliance with your indexing preferences while considering the potential impact on SEO and user experience.
-
When it comes to search engine optimization (SEO), there are certain types of pages that you may consider excluding from being indexed by search engines. Here are some common examples:
Duplicate content pages: If you have multiple pages with similar or identical content, it's generally a good idea to avoid indexing all of them. This could include printer-friendly versions, alternate language versions, or slight variations of the same content.
Temporary or seasonal pages: Pages that are only relevant for a limited time, such as seasonal promotions or special event pages, may not need to be indexed. Once the event or promotion has passed, you can remove them from being indexed to prevent clutter in search engine results.
Private or internal pages: If you have pages that are intended for internal use only, such as employee login pages, private user profiles, or administrative sections, it's typically best to exclude them from indexing. This ensures that sensitive or irrelevant content doesn't appear in search results.
Thin or low-quality pages: Pages with minimal or insufficient content, such as placeholder pages, thin affiliate pages, or low-quality auto-generated content, might not provide much value to search engine users. It's generally better to improve or remove such pages rather than indexing them.
Pagination and sorting pages: Pages that only differ in sorting, filtering, or pagination functionality, such as category listings or search result pages, may not need individual indexing. In these cases, it's often recommended to use canonical tags or URL parameters to consolidate them into a single indexed page.
Remember, these are general recommendations, and the specific needs of your website may vary. It's always a good idea to consult with an SEO professional to understand what pages should or shouldn't be indexed based on your unique circumstances.
things i cant accomplish ai seems interesting being looking at this blog
https://givevaluefirst.com/artificial-intelligence-for-dummies/
-
To prevent certain types of pages from appearing in search engine results, use methods like robots.txt, meta robots tags, or canonicalization. Common pages to exclude include duplicates, low-quality content, search result pages, login/profile pages, thank you pages, and outdated content. Be cautious when choosing which pages to exclude to avoid affecting your site's SEO and user experience.
-
Best practices for preventing indexing of certain types of pages on your website include:
Avoid indexing duplicate content pages.
Exclude pages with thin or low-quality content.
Do not index internal search results pages.
Privacy and policy pages are typically not meant for indexing.
Consider "noindexing" tag or category archive pages.
Author pages may be "noindexed" if they lack substantial content.
"Thank-you" pages after form submissions or purchases can often be excluded.
Dynamic parameters or session IDs should not be indexed.
Pagination pages can be "noindexed" if they duplicate content.
Login or registration pages often don't need indexing.
Implement these practices using "noindex" meta tags or "robots.txt" directives while being cautious not to inadvertently block essential pages. Regularly monitor indexing status through tools like Google Search Console. -
Best practices for preventing indexing of certain types of pages on your website include:
- Avoid indexing duplicate content pages.
- Exclude pages with thin or low-quality content.
- Do not index internal search results pages.
- Privacy and policy pages are typically not meant for indexing.
- Consider "noindexing" tag or category archive pages.
- Author pages may be "noindexed" if they lack substantial content.
- "Thank-you" pages after form submissions or purchases can often be excluded.
- Dynamic parameters or session IDs should not be indexed.
- Pagination pages can be "noindexed" if they duplicate content.
- Login or registration pages often don't need indexing.
Implement these practices using "noindex" meta tags or "robots.txt" directives while being cautious not to inadvertently block essential pages. Regularly monitor indexing status through tools like Google Search Console.
-
I try not to index terms and conditions pages and privacy policy. Then there's the "Thank you" pages that might have conversion tracking pixels on. I do this for a few sites.
-
I have created a new website. I am new to blogging, so I need help with indexing and indexing. Some experts say we need to index our all pages and a few blog posts are just the opposite of that opinion.
I have a few pages like Affiliate Disclosures, Contact, About, and Policy.
Should I index them or not? -
Duplicate Content: Avoid indexing pages with duplicate or thin content to prevent SEO issues.
Private or Confidential Pages: Secure login, checkout, or admin pages should not be indexed.
Thank You Pages: Exclude pages users see after form submissions or purchases.
Low-Value Pages: Hide pages with low-quality or outdated content from indexing.
Temporary Pages: Prevent indexing of staging, test, or under-construction pages.
Pagination: Use rel="nofollow" for paginated pages to consolidate value.
Canonicalization: Set canonical tags to specify preferred URLs for similar content.
Sitemaps: Exclude non-essential pages from your sitemap to control indexing.
Non-Content Files: Don't index PDFs, images, or other non-HTML files.
Disallowed in Robots.txt: Use robots.txt to block search engines from indexing unwanted pages.
-
Certainly! Best practices for types of pages not to index are essential for optimizing your website's SEO performance. By carefully selecting which pages to exclude from search engine indexing, you can improve crawl budget allocation, enhance user experience, and maintain the quality of your site. These practices typically include:
Duplicate Content: Identifying and addressing duplicate content issues by using canonical tags to consolidate signals to search engines.
Thin Content: Evaluating and improving pages with thin or low-quality content by either costume updating them with relevant information or redirecting them to more pertinent pages.
Private or Internal Pages: Ensuring that private or internal pages, such as login pages or admin sections, are not indexed to prevent them from appearing in search results.
Search Result Pages: Excluding search result pages from indexing to prevent user-generated queries from appearing in SERPs.
Media Files: Preventing indexing of media files like images, videos, and PDFs, as they may not provide valuable information in search results.
-
There are a few types of pages that you may not want to index, for a variety of reasons. Here are some of the best practices for types of pages not to index:
Pages that are not relevant to your website's visitors: If a page is not relevant to the content of your website, or if it is not likely to be of interest to your visitors, then there is no reason to index it. This could include pages such as login pages, error pages, and internal pages that are only used by administrators.
Pages that are duplicate content: If a page is duplicate content of another page on your website, then there is no need to index both pages. This could include pages that are generated dynamically, such as search results pages or product pages.
Pages that are not secure: If a page is not secure, such as a page that uses HTTP instead of HTTPS, then you may not want to index it. This is because search engines may flag these pages as insecure, which could deter visitors from visiting your website.
Pages that are frequently updated: If a page is frequently updated, such as a blog page, then you may not want to index it. This is because the search engines will have to crawl the page more often, which could slow down your website.
Pages that are not mobile-friendly: If a page is not mobile-friendly, then you may not want to index it. This is because more and more people are using mobile devices to access the internet, and search engines are starting to favor mobile-friendly websites.
By following these best practices, you can ensure that your website is indexed by search engines only with the pages that are most relevant and useful to your visitors.
Here are some additional tips for deciding which pages not to index:
Consider your audience: Think about the type of content that your visitors are looking for. If a page is not relevant to their interests, then there is no need to index it.
Use your analytics: Look at your website analytics to see which pages are the most popular. These are the pages that you should focus on indexing.
Get feedback from your visitors: Ask your visitors what type of content they are looking for. This feedback can help you decide which pages to index and which pages to exclude.
By following these tips, you can make sure that your website is indexed by search engines in a way that is beneficial to your visitors.
-
Indexing decisions for web pages play a crucial role in search engine optimization (SEO) and overall website management. There are certain types of pages that you may want to prevent search engines from indexing to maintain the quality of your website's search engine results and to avoid potential SEO issues. Here are some best practices for types of pages not to index:
Thin Content Pages: Avoid indexing pages with minimal or low-quality content. Such pages can include placeholder pages, duplicate content, or pages with very little text. Thin content can harm your website's SEO.
Internal Search Result Pages: Search engines can sometimes index internal search result pages, which can lead to duplicate content issues. Use the "noindex" meta tag to prevent indexing of these pages.
Tag and Category Pages: If you have a blog or a content-heavy website, tag and category pages may contain duplicate or low-value content. Consider using the "noindex" tag for these pages.
Thank You and Confirmation Pages: Pages that users see after completing a form or making a purchase are often not useful for search engine results. Prevent these pages from being indexed to avoid cluttering search results.
Private or Confidential Pages: Pages with sensitive information or private data should never be indexed. Make sure to use proper authentication and access controls to protect these pages.
Duplicate Content Pages: If you have multiple versions of the same content (e.g., print-friendly versions, mobile versions), use canonical tags to indicate the preferred version and prevent duplicate content issues.
Session ID or URL Parameters: Pages with session IDs or excessive URL parameters can create many duplicate URLs. Use URL canonicalization techniques or robots.txt to prevent indexing of unnecessary variations.
Login Pages and Admin Sections: Prevent search engines from indexing login pages and admin sections of your website to maintain security and keep sensitive information hidden.
Temporary or Under-Construction Pages: If you're working on a page that's not ready for public viewing, use the "noindex" tag to prevent it from appearing in search results.
404 Error Pages: While 404 error pages should not be indexed, it's essential to provide a helpful 404 page that guides users to relevant content or the homepage.
Pagination Pages: For paginated content like articles split across multiple pages, it's often best to let search engines index the main content and use rel="prev" and rel="next" tags to indicate the paginated structure without indexing each page individually.
Regards : Epicsprtsx
-
I want to extend my gratitude to the author for this comprehensive guide on best practices for managing indexing in SEO. As someone deeply involved in digital marketing and SEO, I find this topic to be of utmost importance. In today's fast-paced online landscape, it's crucial to make informed decisions about which pages to index and which to keep out of search engine results pages (SERPs).
One of the key takeaways from this article is the emphasis on optimizing crawl budget. Google's crawl budget is a finite resource, and ensuring that search engines allocate it wisely can significantly impact a website's overall performance. The author rightly points out that preventing the indexing of pages that don't add substantial value to users can help in this regard.
One of the practices highlighted in the article is the use of the "noindex" meta tag. This is a simple yet effective way to communicate to search engines that specific pages should not be included in their index. I appreciate the step-by-step instructions provided on how to implement this tag properly. This can be particularly helpful for those new to SEO.
Additionally, the article's discussion on using robots.txt to disallow crawling of certain pages is a valuable strategy. However, as mentioned, it's important to exercise caution when using this method to prevent accidentally blocking important pages. The emphasis on regularly monitoring the robots.txt file and conducting thorough testing is a crucial piece of advice. This shows that the author is not just focused on prevention but also on maintaining site health and visibility.
Another aspect that I found intriguing is the section on "thin content" pages. Identifying and addressing these pages is essential for maintaining the quality of a website. It's great to see practical recommendations on how to handle such pages, including updating them with relevant content or redirecting them to more relevant pages. This demonstrates a commitment to providing the best possible user experience, which is at the core of SEO success.
Furthermore, the article delves into the nuances of handling duplicate content, an issue that many SEO practitioners encounter. The explanation of canonical tags and their role in consolidating duplicate content signals to search engines is spot on. The emphasis on regular audits to identify and resolve duplicate content issues is a proactive approach that can prevent potential ranking and indexing problems down the road.
I would like to add that, in my experience, it's also important to stay updated with Google's guidelines and algorithm changes. Google's algorithms are constantly evolving, and what works today may not be as effective tomorrow. Therefore, staying informed through resources like Google Webmaster Guidelines and reputable SEO news sources is essential.
In conclusion, this article provides a wealth of practical information and strategies for managing indexing in SEO. It's evident that the author has a deep understanding of the subject matter and is committed to helping SEO professionals make informed decisions. I look forward to reading more insightful articles from this source in the future. Thank you for sharing these valuable insights.
Team marketingratis.com
-
When it comes to optimizing your website for search engines, there are certain types of pages that you may want to consider not indexing. Here are a few examples of such pages:
Duplicate content: Pages that contain identical or substantially similar content to other pages on your site or elsewhere on the web. Search engines prefer unique and original content, so it's best to avoid indexing duplicate pages.
Thin content: Pages that lack substantial content or are of low quality. These can include pages that have little to no text, primarily consisting of images, videos, or advertisements. Search engines tend to prioritize pages with valuable and informative content.
Temporary or staging pages: Pages created during the development or testing phase of your website, which may not be relevant or useful to search engine users. It's a good idea to prevent these pages from being indexed to avoid confusion or negative impacts on search engine rankings.
Private or sensitive information: Pages that include personal information, login pages, or any content that should be restricted to authorized users only. Preventing indexing of such pages can help maintain privacy and security.
To prevent search engines from indexing specific pages, you can use the "robots.txt" file directives or the HTML "noindex" meta tag. These methods allow you to control which pages search engines can or cannot index.
Remember, it's essential to regularly review and update your website's indexing strategies to ensure optimal visibility and user experience.
-
The article does a great job of highlighting various scenarios where you should consider using the "noindex" meta tag or other techniques to prevent certain pages from appearing in search engine results. (Canada PR)Whether it's duplicate content, thin or low-quality pages, internal search result pages, or sensitive information, this post provides valuable insights and actionable tips to help improve your website's SEO and user experience.
-
Pages with duplicate content, like printer-friendly versions, should be set as "noindex" to prevent confusion. Thin or low-quality content pages, such as placeholders or login screens, should also be excluded. Internal search results, tag/category pages, and user-generated content areas might be better off without indexing due to potential duplicate content e.g https://smamepestimate.com/ or spam concerns. Thank you/confirmation pages, sensitive content, and paginated/sorted versions of content can also benefit from not being indexed.
-
@donsilvernail What should I do with pages that I've de-indexed intentionally?
Like contact us and privacy policy generator I have to do it on my personal blog Footyware. Can I interlink it with my other pages too like homepage etc please guide. -
Need to be clear on the purpose of "no-index". Search engines will still crawl the page, but in theory will not be published in the index. Some search engines may still choose to index the page despite no-index tag. Also that page will still be publicly accessible on your website.
As already noted a couple of times I would be very slow to noindex any page.
I can't think of very many applications where it would be used. The way I view it is either something is public or its private, if it's public you properly want search engines to find it, or if it's private it should be locked away behind a username and password.
-
Hi Richard,
Some archive pages in WordPress can produce significant traffic. Especial if the articles that reside under the archive are informative and the tag or category you use is a good keyword and provides value. So i only no index archives that have no real value.
Contact forms are up to you. Does the form sit on a landing page you want visitors? Or is it an internal link for data collection. A determination on what should be indexed or no indexed is what pages bring value to potential visitors. Many internal search pages bring no value to a user searching for your content on google. So these could be no index. User archives could be no index especial if the user is not an author of content on your site.
Thanks,
Don Silvernal
-
Hi there,
Really any pages that you would not want returned to a user in the SERPs. Does the site contain sensitive personal information in some sort of customer profile? If so, you would want to index these pages.
I would not noindex contact form pages (valuable for users to be able to find) but internal search pages would be a good candidate as well as 'thank you' pages. If you have an ecommerce website, noindexing the shopping cart would be another smart idea.
As for archive pages, I tend to handle these with a canonical tag.
Hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
What is the best tool for getting a SiteMap url for a website with over 4k pages?
I have just migrated my website from HUGO to Wordpress and I want to submit the sitemap to Google Search Console (because I haven't done so in a couple years). It looks like there are many tools for getting a sitemap file built. But I think they probably vary in quality. Especially considering the size of my site.
Technical SEO | | DanKellyCockroach2 -
Best Practices For Angular Single Page Applications & Progressive Web Apps
Hi Moz Community, Is there a proper way to do SPA (client side rendered) and PWA without having a negative impact on SEO? Our dev team is currently trying to covert most of our pages to Angular single page application client side rendered. I told them we should use a prerendering service for users that have JS disabled or use server side rendering instead since this would ensure that most web crawlers would be able to render and index all the content on our pages even with all the heavy JS use. Is there an even better way to do this or some best practices? In terms of the PWA that they want to add along with changing the pages to SPA, I told them this is pretty much separate from SPA's because they are not dependent. Adding a manifest and service worker to our site would just be an enhancement. Also, if we do complete PWA with JS for populating content/data within the shell, meaning not just the header and footer, making the body a template with dynamic JS as well would that effect our SEO in any way, any best practices here as well? Thanks!
Technical SEO | | znotes0 -
Website SEO Product Pages - Condense Product Pages
We are managing a website that has seen consistently dropping rankings over the last 2 years (http://www.independence-bunting.com/). Our long term strategy has been purely content-based and is of high quality, but isn’t seeing the desired results. It is an ecommerce site that has a lot of pages, most of which are category or product pages. Many of the product pages have duplicate or thin content, which we currently see as one of the primary reasons for the ranking drops.The website has many individual products which have the same fabric and size options, but have different designs. So it is difficult to write valuable content that differs between several products that have similar designs. Right now each of the different designs has its own product page. We have a dilemma, because our options are:A.Combine similar designs of the product into one product page where the customer must choose a design, a fabric, and a size before checking out. This way we can have valuable content and don’t have to duplicate that content on other pages or try to find more to say about something that there really isn’t anything else to say about. However, this process will remove between 50% and 70% of the pages on the website. We know number of indexed pages is important to search engines and if they suddenly see that half of our pages are gone, we may cause more negative effects despite the fact that we are in fact aiming to provide more value to the user, rather than less.B.Leave the product pages alone and try to write more valuable content for each product page, which will be difficult because there really isn’t that much more to say, or more valuable ways to say it. This is the “safe” option as it means that our negative potential impact is reduced but we won’t necessarily see much positive trending either. C.Test solution A on a small percentage of the product categories to see any impact over the next several months before making sitewide updates to the product pages if we see positive impact, or revert to the old way if we see negative impact.Any sound advice would be of incredible value at this point, as the work we are doing isn’t having the desired effects and we are seeing consistent dropping rankings at this point.Any information would be greatly appreciated. Thank you,
Technical SEO | | Ed-iOVA0 -
How do I get my pages to go from "Submitted" to "Indexed" in Google Webmaster Tools?
Background: I recently launched a new site and it's performing much better than the old site in terms of bounce rate, page view, pages per session, session duration, and conversions. As suspected, sessions, users, and % new sessions are all down. Which I'm okay with because the the old site had a lot of low quality traffic going to it. The traffic we have now is much more engaged and targeted. Lastly, the site was built using Squarespace and was launched the middle of August. **Question: **When reviewing Google Webmaster Tools' Sitemaps section, I noticed it says 57 web pages Submitted, but only 5 Indexed! The sitemap that's submitted seems to be all there. I'm not sure if this is a Squarespace thing or what. Anyone have any ideas? Thanks!!
Technical SEO | | Nate_D0 -
Unnecessary pages getting indexed in Google for my blog
I have a blog dapazze.com and I am suffering from a problem for a long time. I found out that Google have indexed hundreds of replytocom links and images attachment pages for my blog. I had to remove these pages manually using the URL removal tool. I had used "Disallow: ?replytocom" in my robots.txt, but Google disobeyed it. After that, I removed the parameter from my blog completely using the SEO by Yoast plugin. But now I see that Google has again started indexing these links even after they are not present in my blog (I use #comment). Google have also indexed many of my admin and plugin pages, whereas they are disallowed in my robots.txt file. Have a look at my robots.txt file here: http://dapazze.com/robots.txt Please help me out to solve this problem permanently?
Technical SEO | | rahulchowdhury0 -
Properly Moving Blog from Index to its Own Page
Right now I have a website that is exclusively a blog. I want to create pages outside of the blog and move the blog to a page other than the index file e.g.) from domain.com to domain.com/blog I will have the blog post pages stay in the root directory. e.g.) domain.com/blog-post Any suggestions how to properly tell SE's and other websites that the blog has moved?
Technical SEO | | Bartell0 -
What is the best way to deal with pages whose content changes?
My site features businesses that offers activities for kids. Each business has its own page on my site. Business pages contains a listing of different activities that organization is putting on (such as events, summer camps, drop-in activities). Some businesses only offer seasonal activities (for example, during Christmas break and summer camps). The rest of the year, the business has no activities -- the page is empty. This is creating 2 problems. It's poor user experience (which I can fix no problem) but it also is thin content and sometimes treated as duplicate content. What's the best way to deal with pages whose content can be quite extensive at certain points of the year and shallow or empty at other parts? Should I include a meta ROBOTS tag to not index when there is no content, and change the tag to index when there is content? Should I just ignore this problem? Should I remove the page completely and do a redirect? Would love to know people's thoughts.
Technical SEO | | ChatterBlock0 -
Optimal / Best Practice Title tag
Hi Guys, Am I write in saying google will take / create many variable from your title tag? Graphic, Web Design and Online Marketing in Ireland | Company Name results: Graphic Design, Web design, Web design in Ireland, Online Marketing in Ireland, Online Marketing, Graphic and Web Design, etc etc. (plus lots of long tail there as well). Would this be considered the optimal way as 'Design' is the common denominator for Graphic & Web. Then Ireland can be common to every other keyword such as Graphic design, Web design, Online Marketing. (in ireland) The reason why I ask is: lately I've notice title tags being stuffed with keywords and don't actually read correctly in the SERP My suggested way could have more benefits plus it reads well. Your thoughts, thanks.
Technical SEO | | Socialdude0