Duplicate Errors from Wordpress login redirects
-
I've some Duplicate issues showing up in Moz Analytics which are due to a Q&A plugin being used on a Wordpress website which prompts the user to login. There's a number of links looking like the one shown below, which lead to the login page:
www.website.com/wp-login.php?redirect_to=http%3A%2F%2Fwww.website.com%question%2....
What's the best way to deal with this?
--
extra info:
-
this is only showing up in Moz Analytics. Google Webmaster Tools reports no duplicates.. I'm guessing this is maybe down to the 'redirect_to' parameter being effective in grouping the URLs for Googlebot.
-
currently the wplogin and consequent redirects are 'noindex, follow' - I cannot see where this is being generated from in wp-login.php to change this to nofollow (if this will solve it).
-
-
Yea I'd already blocked some duplicates from a BuddyPress issue, so I didn't want to just jump in and block straight away without some further investigation. Good to know that's the best solution to keep things clean. Cheers for answering Dan.
-
Greg
That's right, the best way is to block crawling with robots.txt - makes sense to keep crawling clean and efficient. If you're using Yoast you can edit robots.txt right in there, or you can do via FTP.
-
Thanks Dan, that's really helpful...
Webmaster Tools reports no crawl issues or anything strange and the crawled pages matches the site size. I've performed a Screaming Frog crawl with the suggested settings, and it IS seeing those redirected pages - 48 in total - which matches the number Moz Analytics is reporting.
The actual page these redirected URL's end up at is a CATCHA page - This is an unneeded layer of 'extra' security put in place by the hosting company after the spate of Wordpress hacks last year. (Cookie'd users who have recently passed the CATCHA would arrive at the Wordpress login page) As such, we don't have any control over the code on that page or anything.
So I guess that even though WMT isn't complaining about these duplicates, to keep things clean and tidy then blocking with robots.txt is the solution huh.
-
Greg
Generally if you're not seeing this in Webmaster Tools or Screaming Frog (have you tried a crawl there yet?) then it's probably not an issue. Crawl it with Screaming Frog, and if you maintain the default settings (honor robots.txt and don't follow nofollows) and set to Googlebot, this will be a pretty accurate representation of what Google is doing. If the pages don't pop up, you should be fine.
Also, check webmaster tools for "crawl stats" - on average, is Google crawling an abnormal amount of pages compared to the "normal" site size?
If it is a problem, you can always block them with robots.txt
-Dan
-
Always for Login and Logout pages use ( NOINDEX - NOFOLLOW ) so you will not face problems like this again.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Duplicate Content
Let's say a blog is publishing original content. Now let's say a second blog steals that original content via bot and publishes it as it's own. Now further assume the original blog doesn't notice this for several years. How much damage could this do to blog A for Google results? Any opinions?
Intermediate & Advanced SEO | | CYNOT0 -
How to find the redirects on website
I want to find the complete internal redirects on website. Just internally linked. How to find such?
Intermediate & Advanced SEO | | vtmoz0 -
Redirecting a Few URLs to a New Domain
We are in the process of buying the blog section of a site. Let's say Site A is buying Site B. We have taken the content from Site B and replicated it on Site A, along with the exact url besides the TLD. We then issued 301 redirects from Site B to Site A and initiated a crawl on those original Site B urls so Google would understand they are now redirecting to Site A. The new urls for Site A, with the same content are now showing up in Google's index if we do a site:SiteA.com search on the big G. Anyone have any experience with this as to how long before Site A urls should replace Site B urls in the search results? I undestand there may be a ranking difference and CTR difference based on domain bias, etc... I'm just asking if everything goes as planned and there isn't a huge issue, does the process take weeks or months?
Intermediate & Advanced SEO | | seoaustin0 -
Robots.txt and redirected backlinks
Hey there, since a client's global website has a very complex structure which lead to big duplicate content problems, we decided to disallow crawler access and instead allow access to only a few relevant subdirectories. While indexing has improved since this I was wondering if we might have cut off link juice. Since several backlinks point to the disallowed root directory and are from there redirected (301) to the allowed directory I was wondering if this could cause any problems? Example: If there is a backlink pointing to example.com (disallowed in robots.txt) and is redirected from there to example.com/uk/en (allowed in robots.txt). Would this cut off the link juice? Thanks a lot for your thoughts on this. Regards, Jochen
Intermediate & Advanced SEO | | Online-Marketing-Guy0 -
Should i fix 404 my errors?
We have about 250, 404 errors due to changing alot of page names throughout our site. I've read some articles saying to leave them and eventually they will go away. Normally I would do a 301 redirect. What's the best solution?
Intermediate & Advanced SEO | | JimDirectMailCoach0 -
Blog Duplicate Content
Hi, I have a blog, and like most blogs I have various search options (subject matter, author, archive, etc) which produce the same content via different URLs. Should I implement the rel-canonical tag AND the meta robots tag (noindex, follow) on every page of duplicate blog content, or simply choose one or the other? What's best practice? Thanks Mozzers! Luke
Intermediate & Advanced SEO | | McTaggart0 -
Canonical Not Fixing Duplicate Content
I added a canonical tag to the home page last month, but I am still showing duplicate content for the home page. Here is the tag I added: What am I missing? Duplicate-Content.jpg
Intermediate & Advanced SEO | | InnoInsulation0 -
Wordpress Titles
My question is about long url titles, my client is using wordpress and the rankings are going well apart from two, which for some reason just wont move. After using some of the tools available on SEO moz which i have found very helpful I have spotted a re-occuring warning throughout the site, the titles, in word press you have this setting (below) page title : %page_title% | %blog_title% My question is my client has quite good brand online but I done want to impact this. The problem I have is that I have a Keyword in the title then the clients company name which is three words and takes up a lot of space. I am thinking about removing this but in two minds so i was kinda hoping for a bit of advice as this looks like a standard in wordpress. Mike
Intermediate & Advanced SEO | | TomBarker820