Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
What Tools Should I Use To Investigate Damage to my website
-
I would like to know what tools I should use and how to investigate damage to my website in2town.co.uk I hired a person to do some work to my website but they damaged it. That person was on a freelance platform and was removed because of all the complaints made about them. They also put in backdoors on websites including mine and added content.
I also had a second problem where my content was being stolen.
My site always did well and had lots of keywords in the top five and ten, but now they are not even in the top 200. This happened in January and feb.
When I write unique articles, they are not showing in Google and need to find what the problem is and how to fix it. Can anyone please help
-
Repairing website damage requires a structured approach. Start by assessing any issues caused by the freelancer using tools like Wordfence for WordPress to detect backdoors or malicious changes. It’s possible Google penalized you for whatever work the freelancer did. You might need to disavow toxic links they built, for example, or address other issues. Tools such as Google Alerts can help identify content duplicates for action.
For future reference, regular monitoring helps prevent unauthorized changes and stay informed about industry trends. If you monitor a web page for changes, you will ensure no unauthorized adjustments are made on your site. There are several ways to approach this, and several tools you can use; some will notify you when your webpage changes. Monitoring competitors for content trends can also guide your strategy and reveal potential areas for improvement.
-
use google search console and screaming frog
-
The best tool that i used to diagnose problems in my office interior designs service based website is the Google search console GSC. You can also use screaming frog or also use moz to analyze and solve issues
-
Page Freezer: Instantly preserve web pages and social media profiles to capture evidence of website damage
-
To investigate damage to your website, you should consider using tools like Google Search Console for monitoring search performance and detecting issues, website security scanners like Word fence to check for malware or vulnerabilities, and website auditing tools such as SEMrush or Screaming Frog for comprehensive analysis of technical SEO issues and website health.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Page Indexing without content
Hello. I have a problem of page indexing without content. I have website in 3 different languages and 2 of the pages are indexing just fine, but one language page (the most important one) is indexing without content. When searching using site: page comes up, but when searching unique keywords for which I should rank 100% nothing comes up. This page was indexing just fine and the problem arose couple of days ago after google update finished. Looking further, the problem is language related and every page in the given language that is newly indexed has this problem, while pages that were last crawled around one week ago are just fine. Has anyone ran into this type of problem?
Technical SEO | | AtuliSulava1 -
Good to use disallow or noindex for these?
Hello everyone, I am reaching out to seek your expert advice on a few technical SEO aspects related to my website. I highly value your expertise in this field and would greatly appreciate your insights.
Technical SEO | | williamhuynh
Below are the specific areas I would like to discuss: a. Double and Triple filter pages: I have identified certain URLs on my website that have a canonical tag pointing to the main /quick-ship page. These URLs are as follows: https://www.interiorsecrets.com.au/collections/lounge-chairs/quick-ship+black
https://www.interiorsecrets.com.au/collections/lounge-chairs/quick-ship+black+fabric Considering the need to optimize my crawl budget, I would like to seek your advice on whether it would be advisable to disallow or noindex these pages. My understanding is that by disallowing or noindexing these URLs, search engines can avoid wasting resources on crawling and indexing duplicate or filtered content. I would greatly appreciate your guidance on this matter. b. Page URLs with parameters: I have noticed that some of my page URLs include parameters such as ?variant and ?limit. Although these URLs already have canonical tags in place, I would like to understand whether it is still recommended to disallow or noindex them to further conserve crawl budget. My understanding is that by doing so, search engines can prevent the unnecessary expenditure of resources on indexing redundant variations of the same content. I would be grateful for your expert opinion on this matter. Additionally, I would be delighted if you could provide any suggestions regarding internal linking strategies tailored to my website's structure and content. Any insights or recommendations you can offer would be highly valuable to me. Thank you in advance for your time and expertise in addressing these concerns. I genuinely appreciate your assistance. If you require any further information or clarification, please let me know. I look forward to hearing from you. Cheers!0 -
How to get rid of bad backlinks
So I noticed my rankings going down and spam score going up. So under my spam score there are over 100 links for different websites but ALL redirect to semalt.com I researched it and it says they hijacked a bunched of backlinks but don't know much more. How can I get rid of all those backlinks? I was told I could use the disavow tool but apparently that can hurt my ranking as well. The semalt.com site has no backlink to me - it looks like the pages that have/had my backlinks, they've redirected to them. For instance this is one of the links http://www.oxvideos.xyz/indianantyphotoxxx 422005fb-9240-40a4-8b65-f5b1f5079dea-image.png
Link Building | | landlwoof40 -
Moz crawler is not able to crawl my website
Hi, i need help regarding Moz Can't Crawl Your Site i also share screenshot that Moz was unable to crawl your site on Mar 26, 2022. Our crawler was not able to access the robots.txt file on your site. This often occurs because of a server error from the robots.txt. Although this may have been caused by a temporary outage, we recommend making sure your robots.txt file is accessible and that your network and server are working correctly. Typically errors like this should be investigated and fixed by the site webmaster.
Technical SEO | | JasonTorney
my robts.txt also ok i checked it
Here is my website https://whiskcreative.com.au
just check it please as soon as possibe0 -
Unsolved SEO And Digital Marketing Training
Hi Everyone, I have a basic SEO and Digital Marketing knowledge and looking for a course /training which will teach me step by step SEO and tools need to use with hand on training. I have a website (https://gemslearninginstitute.com/) which I need bring in Google Packs and on the first page of Google. I have attended a few courses but none of them offered in depth knowledge with hands on training so whatever I do it is not producing results. Thanks
SEO Tactics | | fslpso0 -
Tool to Generate All the URLs on a Domain
Hi all, I've been using xml-sitemaps.com for a while to generate a list of all the URLs that exist on a domain. However, this tool only works for websites with under 500 URLs on a domain. The paid tool doesn't offer what we are looking for either. I'm hoping someone can help with a recommendation. We're looking for a tool that can: Crawl, and list, all the indexed URLs on a domain, including .pdf and .doc files (ideally in a .xls or .txt file) Crawl multiple domains with unlimited URLs (we have 5 websites with 500+ URLs on them) Seems pretty simple, but we haven't been able to find something that isn't tailored toward management of a single domain or that can crawl a huge volume of content.
Technical SEO | | timfrick0 -
Why is my website banned?
IMy website is Costume Machine at www.costumemachine.com . My site has been banned for 1 year now. I have requested that google reconsider my site 3 times without luck. The site is dynamic and basically pulls in feeds from affiliate sites. We have added over 1,500 pages of original content. The site has been running great since 2008 without any penalties. I don't think I got hit with any linking penalty. I cleaned up all questionable links last November when the penalty hit. Am I being hit with a "thin" site penalty? If that is the issue what is the best way to fix the problem?
Technical SEO | | tadden0 -
Does using parentheses affect the crawlers?
Quick question: if you using a parantheses around a keyword, do search bots still recognize the keyword? Fox ex: Welcome to a website about the National Basketball Association (NBA). Will the bots recognize that I'm trying to optimize to NBA and not (NBA)? Is this different for tags vs. actual body copy?
Technical SEO | | BPIAnalytics2