Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Best practices for types of pages not to index
-
Trying to better understand best practices for when and when not use a content="noindex". Are there certain types of pages that we shouldn't want Google to index? Contact form pages, privacy policy pages, internal search pages, archive pages (using wordpress). Any thoughts would be appreciated.
-
Certainly! When it comes to SEO (Search Engine Optimization), there are certain types of pages that you may want to prevent search engines from indexing. This can help ensure that only your most relevant and valuable content is displayed in search engine results. Here are some best practices for types of pages not to index:
Duplicate Content Pages:
Avoid indexing pages with duplicate content, as search engines prefer unique content.
Use canonical tags to indicate the preferred version of a page.
Thin or Low-Quality Content Pages:Pages with little to no valuable content may harm your site's overall SEO.
Consider adding substantial content to these pages or use meta tags to prevent indexing.
Internal Search Results Pages:Exclude internal search results pages from indexing, as they may lead to a poor user experience in search results.
Use the robots.txt file to disallow crawling of these pages.
Thank You and Confirmation Pages:Pages that users see after completing a form submission or transaction may not provide significant value to search engine users.
Use the noindex meta tag to prevent indexing of thank you and confirmation pages.
Login and Account Pages:Secure pages containing login forms or user account information to prevent unauthorized access.
Use the robots.txt file to disallow crawling of these pages.
Tag and Category Pages:Depending on your content management system, tag and category pages may be automatically generated. These can sometimes result in duplicate content issues.
Use the noindex meta tag or canonical tags as appropriate.
Paginated Pages:For large sets of paginated content, consider only indexing the main paginated page and using the rel="next" and rel="prev" tags to indicate the paginated structure.
Privacy Policy and Terms of Service Pages:While it's important to have these pages, they might not need to be indexed.
Use the noindex meta tag if you don't want search engines to index these legal pages.
Media Files and Non-HTML Content:Files like PDFs, images, and other non-HTML content may not need to be indexed.
Use appropriate meta tags or header directives to prevent indexing.
Test and Development Pages:Pages used for testing or development purposes should not be indexed.
Use authentication or the robots.txt file to block search engine bots from accessing these pages.
Always keep in mind that SEO best practices evolve, and it's essential to stay updated with the latest recommendations from search engines. Regularly check your website's performance in search engine results and adjust your indexing strategy accordingly. -
Best practices for determining which types of pages not to index involve strategic decisions to enhance the overall performance and relevance of your website on search engines. Here are some key considerations:
Thin or Low-Quality Content:
Recommendation: Identify and exclude pages with thin or low-quality content that doesn't provide substantial value to users. Focus on creating high-quality, informative content that aligns with user intent.
Duplicate Content:
Recommendation: Avoid indexing pages with duplicate content, as it can lead to confusion for search engines and may result in lower rankings. Use canonical tags to specify the preferred version of the content.
Internal Search Result Pages:
Recommendation: Exclude internal search result pages from indexing, as they often lead to duplicate content issues. Ensure that search engines focus on the primary content pages of your site.
Archive or Staging Pages:
Recommendation: Prevent search engines from indexing archive or staging pages. Use robots.txt or meta tags to disallow indexing of such pages to maintain the integrity of your live content.
Thank You and Confirmation Pages:
Recommendation: Non-essential pages like thank you or confirmation pages for form submissions may not need indexing. Exclude these pages to avoid unnecessary clutter in search engine results.
Login or Session-Specific Pages:
Recommendation: Exclude pages that require user authentication or are session-specific. This prevents search engines from indexing content that's not meant for public access.
Paginated Pages:
Recommendation: For paginated content, consider using rel="next" and rel="prev" tags to signal the relationship between pages. This helps search engines understand the structure without indexing each individual page.
Category or Tag Pages:
Recommendation: Depending on your website structure, category or tag pages may not need indexing. Ensure that these pages don't dilute the overall relevance of your site and use noindex tags if necessary.
Privacy Policy, Terms of Service, and Legal Pages:
Recommendation: While important for compliance, legal and policy pages may not require indexing in search results. Use noindex tags for these pages, allowing them to serve their purpose without being prominent in search listings.
Dynamic URLs with Parameters:
Recommendation: Exclude dynamically generated pages with URL parameters that don't represent unique content. Utilize canonical tags or parameter handling in Google Search Console to manage these pages.
Unnecessary Media or File Attachment Pages:
Recommendation: Media or file attachment pages may not need indexing. Use noindex tags to prevent these pages from appearing in search results while still providing access to the media itself.
Regularly audit and monitor your site's performance in search engine results to ensure that the selected pages for non-indexing align with your SEO strategy and user experience goals. Always consider the specific needs and structure of your website when implementing these best practices.
Read My Recent post here : PBU in Football
-
Here are some best practices for types of pages not to index:
Duplicate content pages. If you have multiple pages with the same or similar content, it's generally a good idea to avoid indexing all of them. This could include printer-friendly versions, alternate language versions, or slight variations of the same content.
Thin or low-quality content pages. Pages with little or no content, or pages with content that is poorly written or irrelevant to your target audience, should not be indexed.
Internal search results pages. These pages are typically not meant for users to land on directly, and they can clutter up search engine results pages (SERPs).
Privacy and policy pages. These pages are typically not relevant to search users, and they can contain sensitive information that should not be indexed.
Thank-you pages. These pages are typically displayed after a user submits a form or makes a purchase, and they are not meant for indexing.
Login and checkout pages. These pages are typically not relevant to search users, and they can contain sensitive information that should not be indexed.
Staging or test pages. These pages are not meant to be seen by the public, and they can clutter up SERPs.
Paginated pages. If your paginated pages contain the same content as your main product or category pages, you may want to consider noindexing them. This will help to avoid duplicate content issues.
You can use a number of methods to prevent pages from being indexed, including:
Robots.txt file. You can use your robots.txt file to block search engines from crawling certain pages on your website.
Noindex meta tag. You can add a noindex meta tag to the header of a page to prevent it from being indexed.
Canonical tags. You can use canonical tags to specify which version of a page is the preferred version for search engines. This can be helpful for preventing duplicate content issues.
It's important to note that noindexing pages is not always necessary. For example, if you have a blog with a lot of high-quality content, you may want to index all of your pages, even if they have similar content. However, if you have a lot of low-quality or irrelevant pages on your website, it's a good idea to noindex them to avoid harming your SEO.
-
Thanks to all of you for sending me valuable information.
-
To prevent specific types of web pages from being indexed by search engines, follow these best practices: Use the robots.txt file to disallow indexing for entire sections of your website or directories. On individual pages, utilize the meta robots tag to specify "noindex" or "nofollow" directives. Employ the X-Robots-Tag HTTP header to communicate indexing preferences, either at the server level or on specific pages. Password-protect pages that should be accessible only to authorized users. Implement canonical tags to indicate the preferred version of a page. Include only desired pages in your XML sitemap. Maintain a clean URL structure, and use "noindex" directives in robots meta headers for dynamic or user-generated content. For pages you want completely removed from search results, return 404 or 410 HTTP status codes. Regularly monitor indexed pages using tools like Google Search Console to ensure compliance with your indexing preferences while considering the potential impact on SEO and user experience.
-
When it comes to search engine optimization (SEO), there are certain types of pages that you may consider excluding from being indexed by search engines. Here are some common examples:
Duplicate content pages: If you have multiple pages with similar or identical content, it's generally a good idea to avoid indexing all of them. This could include printer-friendly versions, alternate language versions, or slight variations of the same content.
Temporary or seasonal pages: Pages that are only relevant for a limited time, such as seasonal promotions or special event pages, may not need to be indexed. Once the event or promotion has passed, you can remove them from being indexed to prevent clutter in search engine results.
Private or internal pages: If you have pages that are intended for internal use only, such as employee login pages, private user profiles, or administrative sections, it's typically best to exclude them from indexing. This ensures that sensitive or irrelevant content doesn't appear in search results.
Thin or low-quality pages: Pages with minimal or insufficient content, such as placeholder pages, thin affiliate pages, or low-quality auto-generated content, might not provide much value to search engine users. It's generally better to improve or remove such pages rather than indexing them.
Pagination and sorting pages: Pages that only differ in sorting, filtering, or pagination functionality, such as category listings or search result pages, may not need individual indexing. In these cases, it's often recommended to use canonical tags or URL parameters to consolidate them into a single indexed page.
Remember, these are general recommendations, and the specific needs of your website may vary. It's always a good idea to consult with an SEO professional to understand what pages should or shouldn't be indexed based on your unique circumstances.
things i cant accomplish ai seems interesting being looking at this blog
https://givevaluefirst.com/artificial-intelligence-for-dummies/
-
To prevent certain types of pages from appearing in search engine results, use methods like robots.txt, meta robots tags, or canonicalization. Common pages to exclude include duplicates, low-quality content, search result pages, login/profile pages, thank you pages, and outdated content. Be cautious when choosing which pages to exclude to avoid affecting your site's SEO and user experience.
-
Best practices for preventing indexing of certain types of pages on your website include:
Avoid indexing duplicate content pages.
Exclude pages with thin or low-quality content.
Do not index internal search results pages.
Privacy and policy pages are typically not meant for indexing.
Consider "noindexing" tag or category archive pages.
Author pages may be "noindexed" if they lack substantial content.
"Thank-you" pages after form submissions or purchases can often be excluded.
Dynamic parameters or session IDs should not be indexed.
Pagination pages can be "noindexed" if they duplicate content.
Login or registration pages often don't need indexing.
Implement these practices using "noindex" meta tags or "robots.txt" directives while being cautious not to inadvertently block essential pages. Regularly monitor indexing status through tools like Google Search Console. -
Best practices for preventing indexing of certain types of pages on your website include:
- Avoid indexing duplicate content pages.
- Exclude pages with thin or low-quality content.
- Do not index internal search results pages.
- Privacy and policy pages are typically not meant for indexing.
- Consider "noindexing" tag or category archive pages.
- Author pages may be "noindexed" if they lack substantial content.
- "Thank-you" pages after form submissions or purchases can often be excluded.
- Dynamic parameters or session IDs should not be indexed.
- Pagination pages can be "noindexed" if they duplicate content.
- Login or registration pages often don't need indexing.
Implement these practices using "noindex" meta tags or "robots.txt" directives while being cautious not to inadvertently block essential pages. Regularly monitor indexing status through tools like Google Search Console.
-
I try not to index terms and conditions pages and privacy policy. Then there's the "Thank you" pages that might have conversion tracking pixels on. I do this for a few sites.
-
I have created a new website. I am new to blogging, so I need help with indexing and indexing. Some experts say we need to index our all pages and a few blog posts are just the opposite of that opinion.
I have a few pages like Affiliate Disclosures, Contact, About, and Policy.
Should I index them or not? -
Duplicate Content: Avoid indexing pages with duplicate or thin content to prevent SEO issues.
Private or Confidential Pages: Secure login, checkout, or admin pages should not be indexed.
Thank You Pages: Exclude pages users see after form submissions or purchases.
Low-Value Pages: Hide pages with low-quality or outdated content from indexing.
Temporary Pages: Prevent indexing of staging, test, or under-construction pages.
Pagination: Use rel="nofollow" for paginated pages to consolidate value.
Canonicalization: Set canonical tags to specify preferred URLs for similar content.
Sitemaps: Exclude non-essential pages from your sitemap to control indexing.
Non-Content Files: Don't index PDFs, images, or other non-HTML files.
Disallowed in Robots.txt: Use robots.txt to block search engines from indexing unwanted pages.
-
Certainly! Best practices for types of pages not to index are essential for optimizing your website's SEO performance. By carefully selecting which pages to exclude from search engine indexing, you can improve crawl budget allocation, enhance user experience, and maintain the quality of your site. These practices typically include:
Duplicate Content: Identifying and addressing duplicate content issues by using canonical tags to consolidate signals to search engines.
Thin Content: Evaluating and improving pages with thin or low-quality content by either costume updating them with relevant information or redirecting them to more pertinent pages.
Private or Internal Pages: Ensuring that private or internal pages, such as login pages or admin sections, are not indexed to prevent them from appearing in search results.
Search Result Pages: Excluding search result pages from indexing to prevent user-generated queries from appearing in SERPs.
Media Files: Preventing indexing of media files like images, videos, and PDFs, as they may not provide valuable information in search results.
-
There are a few types of pages that you may not want to index, for a variety of reasons. Here are some of the best practices for types of pages not to index:
Pages that are not relevant to your website's visitors: If a page is not relevant to the content of your website, or if it is not likely to be of interest to your visitors, then there is no reason to index it. This could include pages such as login pages, error pages, and internal pages that are only used by administrators.
Pages that are duplicate content: If a page is duplicate content of another page on your website, then there is no need to index both pages. This could include pages that are generated dynamically, such as search results pages or product pages.
Pages that are not secure: If a page is not secure, such as a page that uses HTTP instead of HTTPS, then you may not want to index it. This is because search engines may flag these pages as insecure, which could deter visitors from visiting your website.
Pages that are frequently updated: If a page is frequently updated, such as a blog page, then you may not want to index it. This is because the search engines will have to crawl the page more often, which could slow down your website.
Pages that are not mobile-friendly: If a page is not mobile-friendly, then you may not want to index it. This is because more and more people are using mobile devices to access the internet, and search engines are starting to favor mobile-friendly websites.
By following these best practices, you can ensure that your website is indexed by search engines only with the pages that are most relevant and useful to your visitors.
Here are some additional tips for deciding which pages not to index:
Consider your audience: Think about the type of content that your visitors are looking for. If a page is not relevant to their interests, then there is no need to index it.
Use your analytics: Look at your website analytics to see which pages are the most popular. These are the pages that you should focus on indexing.
Get feedback from your visitors: Ask your visitors what type of content they are looking for. This feedback can help you decide which pages to index and which pages to exclude.
By following these tips, you can make sure that your website is indexed by search engines in a way that is beneficial to your visitors.
-
Indexing decisions for web pages play a crucial role in search engine optimization (SEO) and overall website management. There are certain types of pages that you may want to prevent search engines from indexing to maintain the quality of your website's search engine results and to avoid potential SEO issues. Here are some best practices for types of pages not to index:
Thin Content Pages: Avoid indexing pages with minimal or low-quality content. Such pages can include placeholder pages, duplicate content, or pages with very little text. Thin content can harm your website's SEO.
Internal Search Result Pages: Search engines can sometimes index internal search result pages, which can lead to duplicate content issues. Use the "noindex" meta tag to prevent indexing of these pages.
Tag and Category Pages: If you have a blog or a content-heavy website, tag and category pages may contain duplicate or low-value content. Consider using the "noindex" tag for these pages.
Thank You and Confirmation Pages: Pages that users see after completing a form or making a purchase are often not useful for search engine results. Prevent these pages from being indexed to avoid cluttering search results.
Private or Confidential Pages: Pages with sensitive information or private data should never be indexed. Make sure to use proper authentication and access controls to protect these pages.
Duplicate Content Pages: If you have multiple versions of the same content (e.g., print-friendly versions, mobile versions), use canonical tags to indicate the preferred version and prevent duplicate content issues.
Session ID or URL Parameters: Pages with session IDs or excessive URL parameters can create many duplicate URLs. Use URL canonicalization techniques or robots.txt to prevent indexing of unnecessary variations.
Login Pages and Admin Sections: Prevent search engines from indexing login pages and admin sections of your website to maintain security and keep sensitive information hidden.
Temporary or Under-Construction Pages: If you're working on a page that's not ready for public viewing, use the "noindex" tag to prevent it from appearing in search results.
404 Error Pages: While 404 error pages should not be indexed, it's essential to provide a helpful 404 page that guides users to relevant content or the homepage.
Pagination Pages: For paginated content like articles split across multiple pages, it's often best to let search engines index the main content and use rel="prev" and rel="next" tags to indicate the paginated structure without indexing each page individually.
Regards : Epicsprtsx
-
I want to extend my gratitude to the author for this comprehensive guide on best practices for managing indexing in SEO. As someone deeply involved in digital marketing and SEO, I find this topic to be of utmost importance. In today's fast-paced online landscape, it's crucial to make informed decisions about which pages to index and which to keep out of search engine results pages (SERPs).
One of the key takeaways from this article is the emphasis on optimizing crawl budget. Google's crawl budget is a finite resource, and ensuring that search engines allocate it wisely can significantly impact a website's overall performance. The author rightly points out that preventing the indexing of pages that don't add substantial value to users can help in this regard.
One of the practices highlighted in the article is the use of the "noindex" meta tag. This is a simple yet effective way to communicate to search engines that specific pages should not be included in their index. I appreciate the step-by-step instructions provided on how to implement this tag properly. This can be particularly helpful for those new to SEO.
Additionally, the article's discussion on using robots.txt to disallow crawling of certain pages is a valuable strategy. However, as mentioned, it's important to exercise caution when using this method to prevent accidentally blocking important pages. The emphasis on regularly monitoring the robots.txt file and conducting thorough testing is a crucial piece of advice. This shows that the author is not just focused on prevention but also on maintaining site health and visibility.
Another aspect that I found intriguing is the section on "thin content" pages. Identifying and addressing these pages is essential for maintaining the quality of a website. It's great to see practical recommendations on how to handle such pages, including updating them with relevant content or redirecting them to more relevant pages. This demonstrates a commitment to providing the best possible user experience, which is at the core of SEO success.
Furthermore, the article delves into the nuances of handling duplicate content, an issue that many SEO practitioners encounter. The explanation of canonical tags and their role in consolidating duplicate content signals to search engines is spot on. The emphasis on regular audits to identify and resolve duplicate content issues is a proactive approach that can prevent potential ranking and indexing problems down the road.
I would like to add that, in my experience, it's also important to stay updated with Google's guidelines and algorithm changes. Google's algorithms are constantly evolving, and what works today may not be as effective tomorrow. Therefore, staying informed through resources like Google Webmaster Guidelines and reputable SEO news sources is essential.
In conclusion, this article provides a wealth of practical information and strategies for managing indexing in SEO. It's evident that the author has a deep understanding of the subject matter and is committed to helping SEO professionals make informed decisions. I look forward to reading more insightful articles from this source in the future. Thank you for sharing these valuable insights.
Team marketingratis.com
-
When it comes to optimizing your website for search engines, there are certain types of pages that you may want to consider not indexing. Here are a few examples of such pages:
Duplicate content: Pages that contain identical or substantially similar content to other pages on your site or elsewhere on the web. Search engines prefer unique and original content, so it's best to avoid indexing duplicate pages.
Thin content: Pages that lack substantial content or are of low quality. These can include pages that have little to no text, primarily consisting of images, videos, or advertisements. Search engines tend to prioritize pages with valuable and informative content.
Temporary or staging pages: Pages created during the development or testing phase of your website, which may not be relevant or useful to search engine users. It's a good idea to prevent these pages from being indexed to avoid confusion or negative impacts on search engine rankings.
Private or sensitive information: Pages that include personal information, login pages, or any content that should be restricted to authorized users only. Preventing indexing of such pages can help maintain privacy and security.
To prevent search engines from indexing specific pages, you can use the "robots.txt" file directives or the HTML "noindex" meta tag. These methods allow you to control which pages search engines can or cannot index.
Remember, it's essential to regularly review and update your website's indexing strategies to ensure optimal visibility and user experience.
-
The article does a great job of highlighting various scenarios where you should consider using the "noindex" meta tag or other techniques to prevent certain pages from appearing in search engine results. (Canada PR)Whether it's duplicate content, thin or low-quality pages, internal search result pages, or sensitive information, this post provides valuable insights and actionable tips to help improve your website's SEO and user experience.
-
Pages with duplicate content, like printer-friendly versions, should be set as "noindex" to prevent confusion. Thin or low-quality content pages, such as placeholders or login screens, should also be excluded. Internal search results, tag/category pages, and user-generated content areas might be better off without indexing due to potential duplicate content e.g https://smamepestimate.com/ or spam concerns. Thank you/confirmation pages, sensitive content, and paginated/sorted versions of content can also benefit from not being indexed.
-
@donsilvernail What should I do with pages that I've de-indexed intentionally?
Like contact us and privacy policy generator I have to do it on my personal blog Footyware. Can I interlink it with my other pages too like homepage etc please guide. -
Need to be clear on the purpose of "no-index". Search engines will still crawl the page, but in theory will not be published in the index. Some search engines may still choose to index the page despite no-index tag. Also that page will still be publicly accessible on your website.
As already noted a couple of times I would be very slow to noindex any page.
I can't think of very many applications where it would be used. The way I view it is either something is public or its private, if it's public you properly want search engines to find it, or if it's private it should be locked away behind a username and password.
-
Hi Richard,
Some archive pages in WordPress can produce significant traffic. Especial if the articles that reside under the archive are informative and the tag or category you use is a good keyword and provides value. So i only no index archives that have no real value.
Contact forms are up to you. Does the form sit on a landing page you want visitors? Or is it an internal link for data collection. A determination on what should be indexed or no indexed is what pages bring value to potential visitors. Many internal search pages bring no value to a user searching for your content on google. So these could be no index. User archives could be no index especial if the user is not an author of content on your site.
Thanks,
Don Silvernal
-
Hi there,
Really any pages that you would not want returned to a user in the SERPs. Does the site contain sensitive personal information in some sort of customer profile? If so, you would want to index these pages.
I would not noindex contact form pages (valuable for users to be able to find) but internal search pages would be a good candidate as well as 'thank you' pages. If you have an ecommerce website, noindexing the shopping cart would be another smart idea.
As for archive pages, I tend to handle these with a canonical tag.
Hope this helps!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Japanese URL-structured sitemap (pages) not being indexed by Bing Webmaster Tools
Hello everyone, I am facing an issue with the sitemap submission feature in Bing Webmaster Tools for a Japanese language subdirectory domain project. Just to outline the key points: The website is based on a subdirectory URL ( example.com/ja/ ) The Japanese URLs (when pages are published in WordPress) are not being encoded. They are entered in pure Kanji. Google Webmaster Tools, for instance, has no issues reading and indexing the page's URLs in its sitemap submission area (all pages are being indexed). When it comes to Bing Webmaster Tools it's a different story, though. Basically, after the sitemap has been submitted ( example.com/ja/sitemap.xml ), it does report an error that it failed to download this part of the sitemap: "page-sitemap.xml" (basically the sitemap featuring all the sites pages). That means that no URLs have been submitted to Bing either. My apprehension is that Bing Webmaster Tools does not understand the Japanese URLs (or the Kanji for that matter). Therefore, I generally wonder what the correct way is to go on about this. When viewing the sitemap ( example.com/ja/page-sitemap.xml ) in a web browser, though, the Japanese URL's characters are already displayed as encoded. I am not sure if submitting the Kanji style URLs separately is a solution. In Bing Webmaster Tools this can only be done on the root domain level ( example.com ). However, surely there must be a way to make Bing's sitemap submission understand Japanese style sitemaps? Many thanks everyone for any advice!
Technical SEO | | Hermski0 -
What is SEO best practice to implement a site logo as an SVG?
What is SEO best practice to implement a site logo as an SVG?
Technical SEO | | twisme
Since it is possible to implement a description for SVGs it seems that it would be possible to use that for the site name. <desc>sitename</desc>
{{ STUFF }} There is also a title tag for SVGs. I’ve read in a thread from 2015 that sometimes it gets confused with the title tag in the header (at least by Moz crawler) which might cause trouble. What is state of the art here? Any experiences and/or case studies with using either method? <title>sitename</title>
{{ STUFF }} However, to me it seems either way that best practice in terms of search engines being able to crawl is to load the SVG and implement a proper alt tag: What is your opinion about this? Thanks in advance.1 -
Discrepancy in actual indexed pages vs search console
Hi support, I checked my search console. It said that 8344 pages from www.printcious.com/au/sitemap.xml are indexed by google. however, if i search for site:www.printcious.com/au it only returned me 79 results. See http://imgur.com/a/FUOY2 https://www.google.com/search?num=100&safe=off&biw=1366&bih=638&q=site%3Awww.printcious.com%2Fau&oq=site%3Awww.printcious.com%2Fau&gs_l=serp.3...109843.110225.0.110430.4.4.0.0.0.0.102.275.1j2.3.0....0...1c.1.64.serp..1.0.0.htlbSGrS8p8 Could you please advise why there is discrepancy? Thanks.
Technical SEO | | Printcious0 -
What are best options for website built with navigation drop-down menus in JavaScript, to get those menus indexed by Google?
This concerns f5.com, a large website with navigation menus that drop down when hovered over. The sub nav items (example: “DDoS Protection”) are not cached by Google and therefore do not distribute internal links properly to help those sub-pages rank well. Best option naturally is to change the nav menus from JS to CSS but barring that, is there another option? Will Schema SiteNavigationElement work as an alternate?
Technical SEO | | CarlLarson0 -
What is the best way to redirect visitors to certain pages of your site based on their location?
One website I manage wants to redirect users to state specific pages based on their location. What is the best way to accomplish this? For example a user enters the through site.com but they are in Colorado so we want to direct them to site.com/colorado.
Technical SEO | | Firestarter-SEO0 -
How to create site map for large site (ecommerce type) that has 1000's if not 100,000 of pages.
I know this is kind of a newbie question but I am having an amazing amount of trouble creating a sitemap for our site Bestride.com. We just did a complete redesign (look and feel, functionality, the works) and now I am trying to create a site map. Most of the generators I have used "break" after reaching some number of pages. I am at a loss as to how to create the sitemap. Any help would be greatly appreciated! Thanks
Technical SEO | | BestRide0 -
Unnecessary pages getting indexed in Google for my blog
I have a blog dapazze.com and I am suffering from a problem for a long time. I found out that Google have indexed hundreds of replytocom links and images attachment pages for my blog. I had to remove these pages manually using the URL removal tool. I had used "Disallow: ?replytocom" in my robots.txt, but Google disobeyed it. After that, I removed the parameter from my blog completely using the SEO by Yoast plugin. But now I see that Google has again started indexing these links even after they are not present in my blog (I use #comment). Google have also indexed many of my admin and plugin pages, whereas they are disallowed in my robots.txt file. Have a look at my robots.txt file here: http://dapazze.com/robots.txt Please help me out to solve this problem permanently?
Technical SEO | | rahulchowdhury0 -
Home Page .index.htm and .com Duplicate Page Content/Title
I have been whittling away at the duplicate content on my clients' sites, thanks to SEOmoz's pro report, and have been getting push back from the account manager at register.com (the site was built here and the owner doesn't want to move it). He says these are the exact same page and he can't access one to redirect to the other. Any suggestions? The SEOmoz report says there is duplicate content on both these urls: Durango Mountain Biking | Durango Mountain Resort - Cascade Village http://www.cascadevillagehotel.com/index.htm Durango Mountain Biking | Durango Mountain Resort - Cascade Village http://www.cascadevillagehotel.com/ Your help is greatly appreciated! Sheryl
Technical SEO | | TOMMarketingLtd.0