Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Should I remove the ?replytocom variables in wordpress?
-
I'm using Yoast's wordpress plugin and there is an option to remove the replytocom variables. I'm curious what everyone's thoughts were on that, and if I should do it.
Here's the site if you need to see it.
Thanks!
-
Hey, guys, this is a very old post but because it might be useful to people in the future I thought I would update the URL that Ryan did a. great job posting but no longer goes to the correct place. No one can control what I third-party site changes there URL structure to rite?
you can use this plug-in called
?replytocom= replaced with #replytocom=
-
Mine is not indexed in the Google Search, but in Google Webmasters and SEOmoz they showed the error !
Should i remove those links via URL parameters ?
Not indexed, someone here told me some plugins are helping you, i have All-in seo plugin and removed the comluv plugin(for more spam in my blog). other plugins are like share social network That's it.
Any help for me and my blog will be much appreciated !
-
Thanks Ryan! Have a great 4th of July!
-
What do people not using this plug in do? I'm assuming not many people do this, right?
I presume they accept the WP default options. Our practice and understanding of SEO is what allows us to analyze and make decisions regarding tidbits such as the one you mentioned.
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
I do think there is a benefit. You are impacting a LOT of links. Every comment on your site. It may be a tiny 1% benefit type of thing, but the change applies site wide and will presumably be in place for years.
-
I''ll try to find the link where he talks about using pages instead of posts and share it. Curious to hear your thoughts on it.
I'll go ahead and select that option, thanks for your help. (On a side note, what do people not using this plug in do? I'm assuming not many people do this, right?)
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
-
Regarding the new pages instead of posts idea, do you have a link to share?
Regarding the comment url, the page with the comment should be fully indexed either way. By changing the link, you are helping search engines better understand your site. The comment links do not represent a new page or new information.
Google clearly understands WP sites exceptionally well. I am confident you can choose various options and they will still understand those links represent comments. With that said, I would still go with Yoast on this one.
Actually, SEOmoz does it too. Take a look at their blog comments.
-
Thanks for taking the time to check into it. One I'm concerned with is how this will effect long tail seo / indexing of the comments. How will this effect my organic traffic? (will it hurt it?)
I don't see these sorts of pages coming up in google now, so I'm not sure what selecting that option does (and how it effects the site.)
Yoast does a few things different with his site, and I don't always follow his lead. For example he suggets making new pages instead of new posts for your blog posts. He's the only one I've ever heard say this, or do this.
-
I just took a look at Yoast's site and I now better understand the option to remove the variables. I recommend selecting that option. From the Yoast site:
method remove_reply_to_com [line 939]
string remove_reply_to_com( string $link)
Removes the ?replytocom variable from the link, replacing it with a #comment- <number>anchor.</number> Tags: access: public Parameters: string $link The comment link as a string.
Example: http://yoast.com/user-contact-fields-wordpress/#comment-110294
-
hmm..thanks for the feedback. So do you suggest not blocking those? (and I'll message yoast also and see what his thoughts are.)
Thanks.
-
I understand the logic behind blocking removing the variables. They are a lot of extra links on the page which some webmasters might prefer to manage.
What I would prefer is to reform the link so it was something like: http://noahsdad.com/treadmill-training-progress#replytocom=22729
I am guessing the "respond" portion of the URL acts as if someone pressed the reply button which seems unnecessary. If someone clicks the link whether in search results or otherwise and is taken directly to the comment, they should be quite happy. If they wish to reply they can hit the reply button.
Google ignores anything after the # character in a URL. Therefore Google would see these as simply a link to the page which should already be indexed.
Perhaps you can ask Yoast about his thoughts.
-
Thanks for the kind words, I agree, he is a cutie.
Will blocking those cause the comments not to be indexed though?
-
Yup - removing those will save you the trouble of duplicate content - since Google by default is crawling those as different URLs. By default, if you have comments enabled, there's a link at the bottom of posts with that parameter in the url (the same as the blog post url - see here ---> http://noahsdad.com/treadmill-training-progress/?replytocom=22729#respond ).
Noah is cute!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Robots.txt blocked internal resources Wordpress
Hi all, We've recently migrated a Wordpress website from staging to live, but the robots.txt was deleted. I've created the following new one: User-agent: *
Intermediate & Advanced SEO | | Mat_C
Allow: /
Disallow: /wp-admin/
Disallow: /wp-includes/
Disallow: /wp-content/plugins/
Disallow: /wp-content/cache/
Disallow: /wp-content/themes/
Allow: /wp-admin/admin-ajax.php However, in the site audit on SemRush, I now get the mention that a lot of pages have issues with blocked internal resources in robots.txt file. These blocked internal resources are all cached and minified css elements: links, images and scripts. Does this mean that Google won't crawl some parts of these pages with blocked resources correctly and thus won't be able to follow these links and index the images? In other words, is this any cause for concern regarding SEO? Of course I can change the robots.txt again, but will urls like https://example.com/wp-content/cache/minify/df983.js end up in the index? Thanks for your thoughts!2 -
Mass Removal Request from Google Index
Hi, I am trying to cleanse a news website. When this website was first made, the people that set it up copied all kinds of articles they had as a newspaper, including tests, internal communication, and drafts. This site has lots of junk, but this kind of junk was on the initial backup, aka before 1st-June-2012. So, removing all mixed content prior to that date, we can have pure articles starting June 1st, 2012! Therefore My dynamic sitemap now contains only articles with release date between 1st-June-2012 and now Any article that has release date prior to 1st-June-2012 returns a custom 404 page with "noindex" metatag, instead of the actual content of the article. The question is how I can remove from the google index all this junk as fast as possible that is not on the site anymore, but still appears in google results? I know that for individual URLs I need to request removal from this link
Intermediate & Advanced SEO | | ioannisa
https://www.google.com/webmasters/tools/removals The problem is doing this in bulk, as there are tens of thousands of URLs I want to remove. Should I put the articles back to the sitemap so the search engines crawl the sitemap and see all the 404? I believe this is very wrong. As far as I know this will cause problems because search engines will try to access non existent content that is declared as existent by the sitemap, and return errors on the webmasters tools. Should I submit a DELETED ITEMS SITEMAP using the <expires>tag? I think this is for custom search engines only, and not for the generic google search engine.
https://developers.google.com/custom-search/docs/indexing#on-demand-indexing</expires> The site unfortunatelly doesn't use any kind of "folder" hierarchy in its URLs, but instead the ugly GET params, and a kind of folder based pattern is impossible since all articles (removed junk and actual articles) are of the form:
http://www.example.com/docid=123456 So, how can I bulk remove from the google index all the junk... relatively fast?0 -
Removing duplicate content
Due to URL changes and parameters on our ecommerce sites, we have a massive amount of duplicate pages indexed by google, sometimes up to 5 duplicate pages with different URLs. 1. We've instituted canonical tags site wide. 2. We are using the parameters function in Webmaster Tools. 3. We are using 301 redirects on all of the obsolete URLs 4. I have had many of the pages fetched so that Google can see and index the 301s and canonicals. 5. I created HTML sitemaps with the duplicate URLs, and had Google fetch and index the sitemap so that the dupes would get crawled and deindexed. None of these seems to be terribly effective. Google is indexing pages with parameters in spite of the parameter (clicksource) being called out in GWT. Pages with obsolete URLs are indexed in spite of them having 301 redirects. Google also appears to be ignoring many of our canonical tags as well, despite the pages being identical. Any ideas on how to clean up the mess?
Intermediate & Advanced SEO | | AMHC0 -
Dev Subdomain Pages Indexed - How to Remove
I own a website (domain.com) and used the subdomain "dev.domain.com" while adding a new section to the site (as a development link). I forgot to block the dev.domain.com in my robots file, and google indexed all of the dev pages (around 100 of them). I blocked the site (dev.domain.com) in robots, and then proceeded to just delete the entire subdomain altogether. It's been about a week now and I still see the subdomain pages indexed on Google. How do I get these pages removed from Google? Are they causing duplicate content/title issues, or does Google know that it's a development subdomain and it's just taking time for them to recognize that I deleted it already?
Intermediate & Advanced SEO | | WebServiceConsulting.com0 -
Effect of Removing Footer Links In all Pages Except Home Page
Dear MOZ Community: In an effort to improve the user interface of our business website (a New York CIty commercial real estate agency) my designer eliminated a standardized footer containing links to about 20 pages. The new design maintains this footer on the home page, but all other pages (about 600 eliminate the footer). The new design does a very good job eliminating non essential items. Most of the changes remove or reduce the size of unnecessary design elements. The footer removal is the only change really effect the link structure. The new design is not launched yet. Hoping to receive some good advice from the MOZ community before proceeding My concern is that removing these links could have an adverse or unpredictable effect on ranking. Last Summer we launched a completely redesigned version of the site and our ranking collapsed for 3 months. However unlike the previous upgrade this modifications does not URL names, tags, text or any major element. Only major change is the footer removal. Some of the footer pages provide good (not critical) info for visitors. Note the footer will still appear on the home page but will be removed on the interior pages. Are we risking any detrimental ranking effect by removing this footer? Can we compensate by adding text links to these pages if the links from the footer are removed? Seems irregular to have a home page footer but no footer on the other pages. Are we inviting any downgrade, penalty, adverse SEO effect by implementing this? I very much like the new design but do not want to risk a fall in rank and traffic. Thanks for your input!!!
Intermediate & Advanced SEO | | Kingalan1
Alan0 -
Remove URLs that 301 Redirect from Google's Index
I'm working with a client who has 301 redirected thousands of URLs from their primary subdomain to a new subdomain (these are unimportant pages with regards to link equity). These URLs are still appearing in Google's results under the primary domain, rather than the new subdomain. This is problematic because it's creating an artificial index bloat issue. These URLs make up over 90% of the URLs indexed. My experience has been that URLs that have been 301 redirected are removed from the index over time and replaced by the new destination URL. But it has been several months, close to a year even, and they're still in the index. Any recommendations on how to speed up the process of removing the 301 redirected URLs from Google's index? Will Google, or any search engine for that matter, process a noindex meta tag if the URL's been redirected?
Intermediate & Advanced SEO | | trung.ngo0 -
How Do I Generate a Sitemap for a Large Wordpress Site?
Hello Everyone! I am working with a Wordpress site that is in Google news (i.e. everyday we have about 30 new URLs to add to our sitemap) The site has years of articles, resulting in about 200,000 pages on the site. Our strategy so far has been use a sitemap plugin that only generates the last few months of posts, however we want to improve our SEO and submit all the URLs in our site to search engines. The issue is the plugins we've looked at generate the sitemap on-the-fly. i.e. when you request the sitemap, the plugin then dynamically generates the sitemap. Our site is so large that even a single request for our sitemap.xml ties up tons of server resources and takes an extremely long time to generate the sitemap (if the page doesn't time out in the process). Does anyone have a solution? Thanks, Aaron
Intermediate & Advanced SEO | | alloydigital0 -
Robots.txt & url removal vs. noindex, follow?
When de-indexing pages from google, what are the pros & cons of each of the below two options: robots.txt & requesting url removal from google webmasters Use the noindex, follow meta tag on all doctor profile pages Keep the URLs in the Sitemap file so that Google will recrawl them and find the noindex meta tag make sure that they're not disallowed by the robots.txt file
Intermediate & Advanced SEO | | nicole.healthline0