De-indexing and SSL question
-
Few days ago Google indexed hundreds of my directories by mistake (error with plugins/host), my traffic dropped as a consequence.
Anyway I fixed that and submitted a URL removal request. Now just waiting things to go back to normality.
Meantime I was supposed to move my website to HTTPS this week.
Question:
Should I wait until this indexing error has been fixed or I may as well go ahead with the SSL move?
-
Let me know if you can let me see the domain?
-
PS I agree it would be nice to hear from more people I am sure you will by tomorrow.
-
I respect your want to get this right
“A few days ago, Google indexed hundreds of my directories by mistake (error with plugins/host); my traffic dropped as a consequence.
_ Anyway, I fixed that and submitted a URL removal request. Now just waiting for things to go back to normality.” _
in my opinion, you can see the change happen to the https URLs & you want to be sure that https will not let your unwanted directory back in Google's index.
I think unless you can show me your site better Search & replace will be more comfortable & faster.
Yes, I have used better search & replace many times; it’s a great tool.
Make sure you back up your site before you do a search & replace.
https://wordpress.org/plugins/better-search-replace/
I think you should be able to see everything needed and more from this great how-to the person who made it is a friend & they don't just show you how to do this on Kinsta they show you how to use Apache & Nginx as well.
It covers everything you might want to think about managed wordpress hosting if this is something you don't feel comfortable doing.
https://kinsta.com/blog/http-to-https/
I would look at Pagely, Pantheon Servebolt & Kinsta.
If it helps, I am ranked in the top 10 or 11 for assisting people on Moz. Unless you only blocked the HTTP version of your directories, then you will have the same results with https.
- Would you allow me to see the domain?
- Who is your hosting company?
Respectfully,
Tom Zickell
-
Hi Tom,
yes I have WordPress, I have read a few guides online about it.for the search and replace part, I found a plugin called Better Search Replace: https://www.wpbeginner.com/wp-tutorials/how-to-add-ssl-and-https-in-wordpress/
the instructions in the search engine journal article do the search-replace manually. The plugin should be better, have you tried it?
Good question. In all honesty, I think it's safe to go forward with an HTTPS migration.
I am a bit worried. I asked this question to many people in many forums and nobody replies. Which makes me wonder if it's something really bad. It would be nice to have additional opinions.
-
Good question. In all honesty, I think it's safe to go forward with an HTTPS migration.
A "URL removal request"
A simple move to https two different things that can be done at the same time.
if you have a certain set up please let me know I can give you better instructions on how to completely migrate over to HTTPS. Here are some of the basics below.
- https://www.searchenginejournal.com/https-migration-guide/195103/
- https://www.keycdn.com/blog/http-to-https
- https://gofishdigital.com/steps-in-website-https-migration/
if you're using WordPress
- https://www.searchenginejournal.com/wordpress-http-to-https/236969/
- Photograph https://imgur.com/q1RfhhH.jpg
Search & Replace in Files
To begin, search for instances of your domain pointing to HTTP URLs of your site.
Use the regex for **“www” **and without “www” URL cases and search for
http:\/\/(www\.)?yourdomain\.com
always search and replace your database if you have one.
remember to implement redirects
NGINX
Add the following to your Nginx config.
<code class=" language-nginx">server { listen 80; server_name domain.com www.domain.com; return 301 https://domain.com$request_uri; }</code>
Apache
Add the following to your
.htaccess
file.<code class=" language-apacheconf">RewriteEngine On RewriteCond %{HTTPS} off RewriteRule (.*) https://%{HTTP_HOST}%{REQUEST_URI} [R=301,L]</code>
6. Update your robots.txt File#
Update any hard-coded links or blocking rules you might have in your
robots.txt
that might still be pointing to HTTP directories or files- Update your robots.txt file
- Update your disavow file if you have one
- update Google search console
if you want a very quick easy way to implement HTTPS, redirects as well as certificates I recommend Cloudflare The free version should do it.
-
I hope this helps,
Tom
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Redirect indexed lightbox URLs?
Hello all, So I'm doing some technical SEO work on a client website and wanted to crowdsource some thoughts and suggestions. Without giving away the website name, here is the situation: The website has a dedicated /resources/ page. The bulk of the Resources are industry definitions, all encapsulated in colored boxes. When you click on the box, the definition opens in a lightbox with its own unique URL (Ex: /resources/?resource=augmented-reality). The information for these colored lightbox definitions is pulled from a normal resources page (Ex: /resources/augmented-reality/). Both of these URLs are indexed, leading to a lot of duplicate indexed content. How would you approach this? **Things to Consider: ** -Website is built on Wordpress with a custom theme.
Technical SEO | | Alces
-I have no idea how to even find settings for the lightbox (will be asking the client today).
-Right now my thought is to simply disallow the lightbox URL in robots.txt and hope Google will stop crawling and eventually drop from the index.
-I've considered adding the main resource page canonical to the lightbox URL, but it appears to be dynamically created and thus there is no place to access (outside of the FTP, I imagine?). I'm most rusty with stuff like this, so figured I'd appeal to the masses for some assistance. Thanks! -Brad0 -
Why are only a few of our pages being indexed
Recently rebuilt a site for an auctioneers, however it has a problem in that none of the lots and auctions are being indexed by Google on the new site, only the pages like About, FAQ, home, contact. Checking WMT shows that Google has crawled all the pages, and I've done a "Fetch as Google" on them and it loads up fine, so there's no crawling issues that is standing out. I've set the "URL Parameters" to no effect too. Also built a sitemap with all the lots in, pushed to Google which then crawled them all (massive spike in Crawl rate for a couple days), and still just indexing a handful of pages. Any clues to look into would be greatly appreciated. https://www.wilkinsons-auctioneers.co.uk/auctions/
Technical SEO | | Blue-shark0 -
301 redirect Question
Hi all, I have a client who has a domain lets say www.xyz.de which is redirected 301 to www.zyx.de. Now they're working on a relaunch and they want to use the www.xyz.de as their origibnal doman after that. So, at the end the www.zyx.de - which is indexed by Google - should be redirected to www.xyz.de. It vice versa. So the redirect becomes the original and the original becomes the redirect 😕 Is there anything we have to care off? Or will that run into the hell? Thanx. Seb.
Technical SEO | | TheHecksler0 -
IP addresses indexed?
I've met with a potential client who has a site with 1,000's of very specific part #'s which don't show in the SERP's on Google. They definitely have the issue of dynamic URL's - but the URL for the part # searches is an IP address rather than their domain name - example: 188.888.888.888/partssearch.php?pnum='1233445' I've not seen the IP address used like this for an external website - is this acceptable for SEO purposes? Thanks, Mark
Technical SEO | | DenverKelly0 -
Crawling and indexing content
If a page element (div, e.g.) is initially hidden and shown only by a hover descriptor or Javascript call, will Google crawl and index it’s content?
Technical SEO | | Mont0 -
Question on 301s
Hi Everyone, I have a questions on 301 redirects, i hope someone can give me some help on this. There was some 301 redirects made on some of the URLs at the beginning of the year, however we are now re-structuring the whole website, which means the URLs which had been given a 301 redirect are now getting another 301. The question is, should i delete the first 301 redirect from the htaccess file? Kind Regards
Technical SEO | | Paul780 -
Robots.txt questions...
All, My site is rather complicated, but I will try to break down my question as simply as possible. I have a robots.txt document in the root level of my site to disallow robot access to /_system/, my CMS. This looks like this: # /robots.txt file for http://webcrawler.com/
Technical SEO | | Horizon
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/ I have another robots.txt file in another level down, which is my holiday database - www.mysite.com/holiday-database/ - this is to disallow access to /holiday-database/ControlPanel/, my database CMS. This looks like this: **User-agent: ***
Disallow: /ControlPanel/ Am I correct in thinking that this file must also be in the root level, and not in the /holiday-database/ level? If so, should my new robots.txt file look like this: # /robots.txt file for http://webcrawler.com/
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/
Disallow: /holiday-database/ControlPanel/ Or, like this: # /robots.txt file for http://webcrawler.com/
# mail webmaster@webcrawler.com for constructive criticism **User-agent: ***
Disallow: /_system/
Disallow: /ControlPanel/ Thanks in advance. Matt0 -
Google indexing page with description
Hello, We rank fairly high for a lot of terms but Google is not indexing our descriptions properly. An example is with "arnold schwarzenegger net worth". http://www.google.ca/search?q=arnold+schwarzenegger+net+worth&ie=utf-8&oe=utf-8&aq=t&rls=org.mozilla:en-US:official&client=firefox-a When we add content, we throw up a placeholder page first. The content gets added with no body content and the page only contains the net worth amount of the celebrity. We then go back through and re-add the descriptions and profile bio shortly after. Will that affect how the pages are getting indexed and is there a way we can get Google to go back to the page and try to index the description so it doesn't just appear as a straight link? Thanks, Alex
Technical SEO | | Anti-Alex0