Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Should I remove the ?replytocom variables in wordpress?
-
I'm using Yoast's wordpress plugin and there is an option to remove the replytocom variables. I'm curious what everyone's thoughts were on that, and if I should do it.
Here's the site if you need to see it.
Thanks!
-
Hey, guys, this is a very old post but because it might be useful to people in the future I thought I would update the URL that Ryan did a. great job posting but no longer goes to the correct place. No one can control what I third-party site changes there URL structure to rite?
you can use this plug-in called
?replytocom= replaced with #replytocom=
-
Mine is not indexed in the Google Search, but in Google Webmasters and SEOmoz they showed the error !
Should i remove those links via URL parameters ?
Not indexed, someone here told me some plugins are helping you, i have All-in seo plugin and removed the comluv plugin(for more spam in my blog). other plugins are like share social network That's it.
Any help for me and my blog will be much appreciated !
-
Thanks Ryan! Have a great 4th of July!
-
What do people not using this plug in do? I'm assuming not many people do this, right?
I presume they accept the WP default options. Our practice and understanding of SEO is what allows us to analyze and make decisions regarding tidbits such as the one you mentioned.
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
I do think there is a benefit. You are impacting a LOT of links. Every comment on your site. It may be a tiny 1% benefit type of thing, but the change applies site wide and will presumably be in place for years.
-
I''ll try to find the link where he talks about using pages instead of posts and share it. Curious to hear your thoughts on it.
I'll go ahead and select that option, thanks for your help. (On a side note, what do people not using this plug in do? I'm assuming not many people do this, right?)
You think there is any benefit to doing it, or just one of those "hey why not" sort of things?
-
Regarding the new pages instead of posts idea, do you have a link to share?
Regarding the comment url, the page with the comment should be fully indexed either way. By changing the link, you are helping search engines better understand your site. The comment links do not represent a new page or new information.
Google clearly understands WP sites exceptionally well. I am confident you can choose various options and they will still understand those links represent comments. With that said, I would still go with Yoast on this one.
Actually, SEOmoz does it too. Take a look at their blog comments.
-
Thanks for taking the time to check into it. One I'm concerned with is how this will effect long tail seo / indexing of the comments. How will this effect my organic traffic? (will it hurt it?)
I don't see these sorts of pages coming up in google now, so I'm not sure what selecting that option does (and how it effects the site.)
Yoast does a few things different with his site, and I don't always follow his lead. For example he suggets making new pages instead of new posts for your blog posts. He's the only one I've ever heard say this, or do this.
-
I just took a look at Yoast's site and I now better understand the option to remove the variables. I recommend selecting that option. From the Yoast site:
method remove_reply_to_com [line 939]
string remove_reply_to_com( string $link)
Removes the ?replytocom variable from the link, replacing it with a #comment- <number>anchor.</number> Tags: access: public Parameters: string $link The comment link as a string.
Example: http://yoast.com/user-contact-fields-wordpress/#comment-110294
-
hmm..thanks for the feedback. So do you suggest not blocking those? (and I'll message yoast also and see what his thoughts are.)
Thanks.
-
I understand the logic behind blocking removing the variables. They are a lot of extra links on the page which some webmasters might prefer to manage.
What I would prefer is to reform the link so it was something like: http://noahsdad.com/treadmill-training-progress#replytocom=22729
I am guessing the "respond" portion of the URL acts as if someone pressed the reply button which seems unnecessary. If someone clicks the link whether in search results or otherwise and is taken directly to the comment, they should be quite happy. If they wish to reply they can hit the reply button.
Google ignores anything after the # character in a URL. Therefore Google would see these as simply a link to the page which should already be indexed.
Perhaps you can ask Yoast about his thoughts.
-
Thanks for the kind words, I agree, he is a cutie. Will blocking those cause the comments not to be indexed though?
-
Yup - removing those will save you the trouble of duplicate content - since Google by default is crawling those as different URLs. By default, if you have comments enabled, there's a link at the bottom of posts with that parameter in the url (the same as the blog post url - see here ---> http://noahsdad.com/treadmill-training-progress/?replytocom=22729#respond ).
Noah is cute!
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Should I use noindex or robots to remove pages from the Google index?
I have a Magento site and just realized we have about 800 review pages indexed. The /review directory is disallowed in robots.txt but the pages are still indexed. From my understanding robots means it will not crawl the pages BUT if the pages are still indexed if they are linked from somewhere else. I can add the noindex tag to the review pages but they wont be crawled. https://www.seroundtable.com/google-do-not-use-noindex-in-robots-txt-20873.html Should I remove the robots.txt and add the noindex? Or just add the noindex to what I already have?
Intermediate & Advanced SEO | | Tylerj0 -
6 .htaccess Rewrites: Remove index.html, Remove .html, Force non-www, Force Trailing Slash
i've to give some information about my website Environment 1. i have static webpage in the root. 2. Wordpress installed in sub-dictionary www.domain.com/blog/ 3. I have two .htaccess , one in the root and one in the wordpress
Intermediate & Advanced SEO | | NeatIT
folder. i want to www to non on all URLs Remove index.html from url Remove all .html extension / Re-direct 301 to url
without .html extension Add trailing slash to the static webpages / Re-direct 301 from non-trailing slash Force trailing slash to the Wordpress Webpages / Re-direct 301 from non-trailing slash Some examples domain.tld/index.html >> domain.tld/ domain.tld/file.html >> domain.tld/file/ domain.tld/file.html/ >> domain.tld/file/ domain.tld/wordpress/post-name >> domain.tld/wordpress/post-name/ My code in ROOT htaccess is <ifmodule mod_rewrite.c="">Options +FollowSymLinks -MultiViews RewriteEngine On
RewriteBase / #removing trailing slash
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule ^(.*)/$ $1 [R=301,L] #www to non
RewriteCond %{HTTP_HOST} ^www.(([a-z0-9_]+.)?domain.com)$ [NC]
RewriteRule .? http://%1%{REQUEST_URI} [R=301,L] #html
RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule ^([^.]+)$ $1.html [NC,L] #index redirect
RewriteCond %{THE_REQUEST} ^[A-Z]{3,9}\ /index.html\ HTTP/
RewriteRule ^index.html$ http://domain.com/ [R=301,L]
RewriteCond %{THE_REQUEST} .html
RewriteRule ^(.*).html$ /$1 [R=301,L]</ifmodule> The above code do 1. redirect www to non-www
2. Remove trailing slash at the end (if exists)
3. Remove index.html
4. Remove all .html
5. Redirect 301 to filename but doesn't add trailing slash at the end0 -
Mass Removal Request from Google Index
Hi, I am trying to cleanse a news website. When this website was first made, the people that set it up copied all kinds of articles they had as a newspaper, including tests, internal communication, and drafts. This site has lots of junk, but this kind of junk was on the initial backup, aka before 1st-June-2012. So, removing all mixed content prior to that date, we can have pure articles starting June 1st, 2012! Therefore My dynamic sitemap now contains only articles with release date between 1st-June-2012 and now Any article that has release date prior to 1st-June-2012 returns a custom 404 page with "noindex" metatag, instead of the actual content of the article. The question is how I can remove from the google index all this junk as fast as possible that is not on the site anymore, but still appears in google results? I know that for individual URLs I need to request removal from this link
Intermediate & Advanced SEO | | ioannisa
https://www.google.com/webmasters/tools/removals The problem is doing this in bulk, as there are tens of thousands of URLs I want to remove. Should I put the articles back to the sitemap so the search engines crawl the sitemap and see all the 404? I believe this is very wrong. As far as I know this will cause problems because search engines will try to access non existent content that is declared as existent by the sitemap, and return errors on the webmasters tools. Should I submit a DELETED ITEMS SITEMAP using the <expires>tag? I think this is for custom search engines only, and not for the generic google search engine.
https://developers.google.com/custom-search/docs/indexing#on-demand-indexing</expires> The site unfortunatelly doesn't use any kind of "folder" hierarchy in its URLs, but instead the ugly GET params, and a kind of folder based pattern is impossible since all articles (removed junk and actual articles) are of the form:
http://www.example.com/docid=123456 So, how can I bulk remove from the google index all the junk... relatively fast?0 -
How to 301 redirect old wordpress category?
Hi All, In order to avoid duplication errors we've decided to redirect old categories (merge some categories).
Intermediate & Advanced SEO | | BeytzNet
In the past we have been very generous with the number of categories we assigned each post. One category needs to be redirected back to blog home (removed completely) while a couple others should be merged. Afterwords we will re-categorize some of the old posts. What is the proper way to do so?
We are not technical, Is there a plugin that can assist? Thanks0 -
Wordpress blog in a subdirectory not being indexed by Google
HI MozzersIn my websites sitemap.xml, pages are listed, such as /blog/ and /blog/textile-fact-or-fiction-egyptian-cotton-explained/These pages are visible when you visit them in a browser and when you use the Google Webmaster tool - Fetch as Google to view them (see attachment), however they aren't being indexed in Google, not even the root directory for the blog (/blog/) is being indexed, and when we query:site: www.hilden.co.uk/blog/ It returns 0 results in Google.Also note that:The Wordpress installation is located at /blog/ which is a subdirectory of the main root directory which is managed by Magento. I'm wondering if this causing the problem.Any help on this would be greatly appreciated!AnthonyToTOHuj.png?1
Intermediate & Advanced SEO | | Tone_Agency0 -
Best practice for removing indexed internal search pages from Google?
Hi Mozzers I know that it’s best practice to block Google from indexing internal search pages, but what’s best practice when “the damage is done”? I have a project where a substantial part of our visitors and income lands on an internal search page, because Google has indexed them (about 3 %). I would like to block Google from indexing the search pages via the meta noindex,follow tag because: Google Guidelines: “Use robots.txt to prevent crawling of search results pages or other auto-generated pages that don't add much value for users coming from search engines.” http://support.google.com/webmasters/bin/answer.py?hl=en&answer=35769 Bad user experience The search pages are (probably) stealing rankings from our real landing pages Webmaster Notification: “Googlebot found an extremely high number of URLs on your site” with links to our internal search results I want to use the meta tag to keep the link juice flowing. Do you recommend using the robots.txt instead? If yes, why? Should we just go dark on the internal search pages, or how shall we proceed with blocking them? I’m looking forward to your answer! Edit: Google have currently indexed several million of our internal search pages.
Intermediate & Advanced SEO | | HrThomsen0 -
Removed Site-wide links
Hi there, I have recently removed quite a lot of site-wide links leaving the only link on homepage's of some websites, since doing this I have seen a dramatic drop on my keywords, going from position 2-3 to nowhere. Has anyone else experienced anything like this, should I expect to see a return on these keywords? Thanks
Intermediate & Advanced SEO | | Paul780 -
Robots.txt is blocking Wordpress Pages from Googlebot?
I have a robots.txt file on my server, which I did not develop, it was done by the web designer at the company before me. Then there is a word press plugin that generates a robots.txt file. How Do I unblock all the wordpress pages from googlebot?
Intermediate & Advanced SEO | | ENSO0