Panda, rankings and other non-sense issues
-
Hello everyone
I have a problem here. My website has been hit by Panda several times in the past, the first time back in 2011 (first Panda ever) and then another couple of times since then, and, lastly, the last June 2016 (either Panda or Phantom, not clear yet). In other words, it looks like my website is very prone to "quality" updates by big G:
http://www.virtualsheetmusic.com/
Still trying to understand how to get rid of Panda related issues once for all after so many years of tweaking and cleaning my website of possible duplicate or thin content (301 redirects, noindexed pages, canonicals, etc), and I have tried everything, believe me. You name it. We recovered several times though, but once in a while, we are still hit by that damn animal. It really looks like we are in the so called "grey" area of Panda, where we are "randomly" hit by it once in a while.
Interestingly enough, some of our competitors live joyful lives, at the top of the rankings, without caring at all about Panda and such, and I can't really make a sense of it.
Take for example this competitors of ours:
They have a much smaller catalog than ours, worse quality of offered music, thousands of duplicate pages, ads everywhere, and yet... they are able to rank 1st on the 1st page of Google for most of our keywords. And for most, I mean, 99.99% of them.
Take for example "violin sheet music", "piano sheet music", "classical sheet music", "free sheet music", etc... they are always first.
As I said, they have a much smaller website than ours, with a much smaller offering than ours, their content quality is questionable (not cured by professional musicians, and highly sloppy done content as well as design), and yet they have over 480,000 pages indexed on Google, mostly duplicate pages. They don't care about canonicals to avoid duplicate content, 301s, noindex, robot tags, etc, nor to add text or user reviews to avoid "thin content" penalties... they really don't care about anything of that, and yet, they rank 1st.
So... to all the experts out there, my question is: Why's that? What's the sense or the logic beyond that? And please, don't tell me they have a stronger domain authority, linking root domains, etc. because according to the duplicate and thin issues I see on that site, nothing can justify their positions in my opinion and, mostly, I can't find a reason why we instead are so much penalized by Panda and such kind of "quality" updates when they are released, whereas websites like that one (8notes.com) rank 1st making fun of all the mighty Panda all year around.
Thoughts???!!!
-
Thank you very much Julie, I really appreciated your words. I have wondered so many times what Google think of "quality", and why before us there are always very low quality websites distributing the exact same music for free (often copyrighted music, which is illegal) and most of those sites are full of ads. Is that quality?
We could open a new discussion thread on the "What is quality to Google?" topic, I think it'd be very popular!
Thank you again.
-
Thank you Donna, glad to know that I am not completely mad!!
As for the fact they have done a great job with title and alt tags, anchor texts, I agree, but you know what? That's another realm where I became paranoid for, the so called "over-optimization"... We used to have perfectly optimized titles, descriptions, H1,s ALTs, anchor text, etc... the whole enchilada perfectly optimized, then we began to lose rankings for an unknown reason (Panda? Over-optimization? Too many pages? What else?), and I began becoming paranoid about everything, so we started "de-optimizing" here and there, etc... Here is additional proof that when things are NOT clear, we all become paranoid and lose control on everything.
I may also add that some of the blame should be probably given to the SEO industry that has spread a lot of fear about all this stuff, without giving an absolute "quantification" of what means "too much optimization", or... too much duplicate content, or too much thing content, or too much bad links, etc... how much is "too much"? That's the question for which I am afraid there's not easy answer, but maybe they scared us too much about all this.
Thank you again, and please, let me know if you have any more ideas.
-
I agree with your observations. I don't see why you'd suffer a Panda penalty and your competition wouldn't.
My only other observation (again, not Panda related) is 8notes have made much more extensive use of link title and alt tags to reinforce their keywords and subject matter.
-
Oh, I also forgot to make you note that when we have more than one version for the same title and same instrument, we add a canonical to the first version.
For example, both items below:
http://www.virtualsheetmusic.com/score/HL-170563.html
http://www.virtualsheetmusic.com/score/HL-119157.html
Are canonicalized to this item:
http://www.virtualsheetmusic.com/score/HL-180308.html
Thanks
-
Thank you Donna!
Yes, I am aware we have much more segmentation, but why in the heck that should be bad? And, more importantly, why that should be worse than having much more actual duplicate content than they have if you consider they never use canonicals nor noindex to avoid duplicate issues.
Take for example one of their instrumental pages like this one:
https://www.8notes.com/violin/sheet_music/?orderby=5d
You can order the results by title, artist, level, etc... they don't even care to have duplicate issues by canonicalizing the URL when parameters are added to it. What do you think may be worse for Panda? I am just asking, I'd really like to understand what can be worse.
They have similar issues with their product pages, if you click on their tabs at the top, parameters get added to the URLs but no canonical is present there. We do have canonicals instead, as by Google's manual.
As for having duplicate titles (because of different instrumental versions), yes, that's something we have tackled in the past several times, tried to remove the duplicates (noindexed) but didn't help. I mean, we didn't notice any change by doing that waiting several months after the change. We just got much lower traffic because of that, but nothing positive.
Also, have a look at our competitor how many different version they have for some of the most popular titles:
Bach's Air on G: https://www.8notes.com/scores/air_on_the_g_string_bach_johann_sebastian.asp
Bach's Minuet: https://www.8notes.com/scores/minuet_bach_johann_sebastian.asp
Beethoven's Fur Elise: https://www.8notes.com/scores/fur_elise_beethoven_ludwig_van.asp
Beethoven's Ode to Joy: https://www.8notes.com/scores/ode_to_joy_(9th_symphony)_beethoven_ludwig_van.asp
Now, I understand we may have many more titles than them, our catalog is much bigger than theirs, but still they have similar issues, right? If so, why they have a very privileged spot in SERPs compared to us? I am sorry to sound like a broken record, but I am still not convinced the problem here is all this...
If you have the chance, I am eager to know your after thoughts... thank you again very much for your help and time Donna. Appreciated very much.
-
Yes, I see what you've done with genres, specials, etc. That looks good.
If I compare you to 8notes, you've got a lot more segmentation when it comes to instruments and those pages are NOT noindexed.
For example, you have 29 different versions of the "Dust in the Wind" sheet music page, all very similar. Here are a few:
- http://www.virtualsheetmusic.com/score/HL-328374.html (Dust in the Wind sheet music for violin)
- http://www.virtualsheetmusic.com/score/HL-328387.html (Dust in the Wind sheet music for trumpet solo)
- http://www.virtualsheetmusic.com/score/HL-301822.html (Dust in the Wind sheet music for choir and piano)
- http://www.virtualsheetmusic.com/score/HL-170563.html (Dust in the Wind sheet music for guitar (chords))
- http://www.virtualsheetmusic.com/score/HL-26501.html (Dust in the Wind sheet music for piano solo)
- http://www.virtualsheetmusic.com/score/HL-119157.html (Dust in the Wind sheet music for piano solo V2)
NOT noindexed doesn't mean they're getting indexed by Google. When I did a site command for http://www.virtualsheetmusic.com/score/HL-328387.html (Dust in the Wind sheet music for trumpet solo), it wasn't returned as a search result. When I searched for ""dust in the wind" trumpet solo virtualsheetmusic" it was not returned as a search result.
So maybe you need to consider noindexing all the instrument variations as well, but still offer them up on the site for visitors. I'd check analytics to see if anyone's landing on those pages from search. As I said earlier, I can understand why you'd want them indexed but if they're causing you more harm than good, you might have to balance that out.
I love a challenge but that's the best I can come up with Fabrizo.
-
Thank you Donna for your reply.
Well, I see what you mean, but if you look how those drill-downs are handled on our site, all pages generated dynamically that way are excluded by robots.txt. That's why I am puzzled to see the competitor's website having a similar kind of browsing without worrying about duplicate issues. Also, I apply any possible rule to reduce duplicate content as much as I can even for those few indexed pages such as the use of canonicals (when parameters are present in the URLs) or the use of rel="prev" or rel="next" for paginated content.
Please, let me know if that's what you meant or I am missing anything here.
Thank you again very much!
-
"I'd really like to know from you if you see anything on my website that could trigger a "Panda" kind of penalization, compared to my mentioned competitor above (8nots.com)."
_Key phrase being "compared to my mentioned competitor". Cause yes, I can see things that might trigger a Panda penalization. You have a lot of overlapping / duplicate content but so do your competitors. _
The only thing that comes to mind is if there's a threshold you're exceeding that your competitors aren't by virtue of the fact that you many (and more) ways to tag/filter your content, for example, genres, specials, and ensembles.
-
Yes, I agree with you, I don't see much logic beyond that. Of course, if they also count images, we have hundreds of thousands... we are a pretty big website, what do you expect, right?
My website situation makes people "scratch their heads", all the time...
I'd really like to know from you if you see anything on my website that could trigger a "Panda" kind of of penalization, compared to my mentioned competitor above (8nots.com). So far, no one on this thread has given me any hints on that.
Thank you again for your insights, I appreciated it very much.
-
I wonder if "indexed URLs" is an accurate label. I looked at the URLs Majestic found, and a significant number are redirected files, images, and mp3s.
You have a perplexing problem Fabrizo...
-
Donna, as a side note, I have no idea where Majestic pulled out over 944,000 indexed pages for our website. By spidering it with Screaming Frog we couldn't crawl more than 387,726 pages... unless they crawled all links dynamically generated by our internal search engine, which by the way, is blocked by the robots.txt file, therefore all those dynamic pages should be not counted.
Also, on the actual Google index, if you use the site: command, you'll see that Google has indexed just 123,000 pages from our site (because most of them are canonicalized), whereas you'll see over 545,000 for 8notes.com.
Actual data seems to be a little different...
-
Thank you Julie for your posting and for participating in this discussion.
Well, what you say might be true, but being an algorithmic penalization that shouldn't really happen... unless the system is flawed in some way to catch the wrong guys (every time?)
Also, thin and duplicate content is so much more obvious and noticeable on our competitors that makes me completely mad trying to find a logical explanation of why me and not them!
Unless, Panda or other similar "quality" updates are looking now for something else nobody has clearly understood yet...
-
I think it's like the Highway Patrol.
Everyone speeds, but everyone doesn't get caught. You point to a similar web page that hasn't been penalized, but there are many who have.
-
Thank you Kristen,
I have just put down a plan to re-architect our website that way and create "sub-categories" in the same way our competitor has done to push-up the main category pages as well (according to the "soloing" technique).
As I wrote yesterday below, this is also something clearly NOT related to Panda... just another needed tweak to the site.
Thank you again, appreciated your help!
-
Thank you Donna! Yes, I am aware of our different back link profile. We are a commercial website, therefore we have many backlinks from hundreds of affiliates... and that could cause issues, I am aware of that. I have worked a lot with my affiliate to put nofollow links where necessary, and to not pass page rank as much as possible... But again, we are talking about issues NOT Panda related... right?
So... again, this can't explain why the first Panda in 2011 as well as the last quality update released in June (was that really Panda?) has hit us hard. I am getting convinced that it is not Panda the beast hitting us once in a while, but something else... my point is: I could be under "several" penalties, ok, I get that... I could be under some Panda penalization, or other quality penalization, and maybe Penguin to some extent (I could never find a clear relation between my traffic loss and the release of Penguin updates though)... but if I am really in the eye of Panda when that happen, back to my original question, why my competitor has never been touched by the white & black bear when its content should be much more prone to Panda than mine? That's the whole point of my conversation here and the answer I am trying to find. I am trying to find a logical explanation of why my traffic dropped with the release of Panda updates, whereas my competitor wasn't touched at all.
Thank you again for your help, appreciated!
-
You're right. Your site and 8notes seem to be guilty of the same practices, assuming there's anything wrong with them, and yet it is ranking well. Although according to MajesticSEO, it has half the pages you have indexed (520,439 vs 944,432).
Your link profiles are significantly different. Again according to Majestic, you have way more backlinks (649,076 vs 234,122) but from half as many referring domains, IPs and subnets. You have 1/10th of the educational backlinks of 8notes. And the majority of your backlinks, roughly 55%, are nofollow whereas 90% of 8notes are the opposite (follow). 8notes seems to have more deep links as well.
Maybe it's worth looking a little more closely at your link profile?
8notes is also https. That might also have a bearing given you're both ecommerce.
-
In any case, I created this thread to discuss about Panda and its "possible" and "not-possible" implications... so, the discussion is still be open.
In the meantime I wait with hope for an answer to my questions above from Kristen (thank you again Kristen!), I'd like to get back to the topic: Panda. It is clear to me that my situation could be improved as Kristen has suggested above, but it is also clear that if so, that's nothing to do with Panda, isn't it? That's just about "content consolidation" and "topical relevance".
Then, back to my original discussion topic, what about the classical Panda issues such as "thin content", or "duplicate content"? As I wrote above, my competitor has plenty of that kind of content, but doesn't seem to have been touched by Panda whatsoever. So... what's the deal with Panda then? And in my particular niche, should I worry more about "topical relevance" and keep optimizing my site under other aspects (usability, user intent, etc), and stop worrying to much about thin and duplicate content?
If you were me, what would you do considering my competitor's evidence? How many other site owners like myself have become "paranoid" about Panda (wasting tons of time, resources and money) and have instead lost focus on other (probably more important!) issues such as topical relevance, content organization, usability, user intent, etc.
More thoughts on that?
-
This is a very good answer, thank you Kristen. The more I look at the "site structure" of my competitor compared to ours, the more I realize we need to work on that.
I have also started to think about the so called "siloing technique" Bruce Clay introduced a few years ago, and it looks like 8notes.com has done a very good job to follow that kind of concept, whereas we are probably "spreading" too much of our juice around thousands of different pages and categories... what are your thoughts on that?
Just a thought about the fact I have put those pages to be no indexed in the robots.txt file: If you look at those pages, you see they are generated dynamically from our internal search engine. And as you can see, you can filter results by clicking the filters on the left side of the page... which is a great thing for users, but can be problematic for search engines. Right? So.. that's why I decided to simply no-robot those pages, to avoid any possible indexing and crawling issues. So... how would you suggest tackling that problem? My first idea would be to create "static pages" for those dynamic pages linked from the category pages, and then block via no-robot the links of the filters on the left side... do you have any different ideas?
Thank you again for your help! Super-appreciated!
-
Thank you Kristen for your kind reply.
Yes, of course, I have considered that a thousand times. Do you really think that could cause so much trouble to make most of our rankings slip over the 10th or 20th page of the SERPS? Those pages you have mentioned on our website, are actually blocked by robots.txt, with the exception of the first page of course. My concern is about those first pages that should be able to rank anyway... unless you tell me that the contextual weight of the "subsequent" pages could play a role to "boost" the first page in some way if Google can spider and index them... but then, I'd be concerned with "too much similar or thin content" because by doing what our competitors are doing, I'd create thousands of additional pages with inside pretty much the same content (lists) organized in different ways... you see what I mean? Of course it seems to work for our competitor, but hence the contradiction and absurdity I was talking about above with the Panda algorithm: shouldn't all those thousands of extra and similar pages be bamboo for Panda?
I hope to have not confused you... I am just trying to find the elephant here that is causing the problem...
Thank you for your help!
-
Thank you Donna for your reply, but that's exactly why I posted my concerns here: How do I know if a page or a set of pages are "causing" me to not rank? That's the purpose of this discussion... how do I actually know that any action of that kind, by "nuking" pages, will be causing me more benefit than damage?
A couple of years ago, I thought to be under a Panda penalty, and I started to remove from the index many of our product pages that I thought were not bringing us traffic or had low user engagement... well, the only result we had was a steady decline of traffic because of the removed pages, that's all. I put all pages back after 6 months because we would have died miserably. After 6 months I concluded that we were NOT under a Panda penalty. IN fact, I put back all pages, traffic got back and we started to rank better than before (fortunate event??)
Also, if I we are under a Panda or similar penalty, why do we still rank well for some keywords? And, back again to square one: Why should we think to be under a Panda penalization if our content is actually less thin, less duplicate, and better handled with canonicals, noindex tages, etc, in order to cope with any possible Panda penalty than our well-ranked competitors??!
-
The only thing I can think to suggest is to look at how much inbound search traffic you are receiving on your category pages (e.g. genres, instruments, skill levels, exclusive, specials, etc.) to assess whether any of those could be noindexed. I understand why you'd want them indexed, but if they're causing you to not rank at all, then you might have to balance that out.
-
Thank you Julie, appreciated!! I really don't understand why most of the times our website content is buried in the search results, whereas "crappy" websites are shown prominently before us... do you call that "quality" big G??!!
Thank you again for your kind words
-
I feel bad for you. I don't have any answers, but I'm a singer, and your website is excellent. This is not an example of Google rewards quality.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Hreflang implementation issue
We are currently handling search for a global brand www.example.com which has presence in many countries worldwide. To help Google understand that there is an alternate version of the website available in another language, we have used “hreflang” tags. Also, there is a mother website (www.example.com/global) which is given the attribution of “x-default” in the “hreflang” tag. For Malaysia as a geolocation, the mother website is ranking instead of the local website (www.example.com/my) for majority of the products. The code used for “hreflang” tag execution, on a product page, being: These “hreflang” tags are also present in the XML sitemap of the website, mentioning them below: <loc>http://www.example.com/my/product_name</loc> <lastmod>2017-06-20</lastmod> Is this implementation of “hreflang” tags fine? As this implementation is true across all geo-locations, but the mother website is out-ranking me only in the Malaysia market. If the implementation is correct, what could be other reasons for the same ranking issue, as all other SEO elements have been thoroughly verified and they seem fine.
Intermediate & Advanced SEO | | Starcom_Search0 -
Not Ranking - Any Tips?
Hi guys, I have a client site that I've recently come onboard with that was published late last year, not really optimized for anything, and in a moderately, but not very, competitive, search space. Early April we optimized the home page and a couple of other pages and have since built about 5-6 (high quality, partial match) links to it, and a press release was done mid last month. The only other thing we did was change the site from non-www to www and set this as the preferred domain in Search Console. Over 6 weeks since that all began, and we're still not on the radar at all for any of our main keywords - nowhere. The only thing we are really ranking for is our brand name, but this is the wrong (press release, not home!) page, and it's bouncing a lot. All of the pages seem to be indexed, and we are ranking for one other (inconsequential) keyword, but 99 is the highest it has reached. An SEO friend told me to build some citations, but this is not a local business, nor are we trying to rank locally. Can anyone please suggest why it might be taking so long, and what else I could try? I imagine more links will help, but results from our outreach are hard to predict, so if there were another safe link type that could help me figure out whether this domain is in trouble or not ASAP, that would be ideal. Thanks very much in advance for any help you can provide. Ulla
Intermediate & Advanced SEO | | Ullamalm0 -
Have a Robots.txt Issue
I have a robots.txt file error that is causing me loads of headaches and is making my website fall off the SE grid. on MOZ and other sites its saying that I blocked all websites from finding it. Could it be as simple as I created a new website and forgot to re-create a robots.txt file for the new site or it was trying to find the old one? I just created a new one. Google's website still shows in the search console that there are severe health issues found in the property and that it is the robots.txt is blocking important pages. Does this take time to refresh? Is there something I'm missing that someone here in the MOZ community could help me with?
Intermediate & Advanced SEO | | primemediaconsultants0 -
Issue with Site Map - how critical would you rank this in terms of needing a fix?
A problem has been introduced onto our sitemap whereby previously excluded URLs are no longer being correctly excluded. These are returning a HTTP 400 Bad Request server response, although do correctly redirect to users. We have around 2300 pages of content, and around 600-800 of these previously excluded URLs, An example would be http://www.naturalworldsafaris.com/destinations/africa-and-the-indian-ocean/botswana/suggested-holidays/botswana-classic-camping-safari/Dates and prices.aspx (the page does correctly redirect to users). The site is currently being rebuilt and only has a life span of a few months. The cost our current developers have given us for resolving this is quite high with this in mind. I was just wondering: How much of a critical issue would you view this? Would it be sufficient (bearing in mind this is an interim measure) to change these pages so that they had a canonical or a redirect - they would however remain on the sitemap. Thanks
Intermediate & Advanced SEO | | KateWaite
Kate0 -
How to improve ranking of deep pages?
While this may sound like an obvious or stupid question at first...let me explain... We are an e-commerce website which sells one type of item nationally; for sake of an example which is similar to us, you can think of an e-commerce site that sells movie theater tickets in cities and towns across the country. Our home page ranks very well for the appropriate keywords as well as some of our state and city pages rank very well for local searches. However, while some state and city pages rank well for their respective local searches, others have a low page rank with some not even in the top 50 for their respective keywords. My question is that we aren't clear why some pages will rank well while others wont when the competition looks similar for those local searches. And in today's Panda/Penguin era we are unsure of how to get more of these state/city pages ranking better? For the record, we are quite strict about on-page SEO, 99% of our 5600 pages are crawled & we have minimum SEO errors from the SEOMoz crawls. Can anyone provide some feedback & thoughts?
Intermediate & Advanced SEO | | CTSupp0 -
How to rank for difficult terms
I am helping launch a new business venture that sells oak furniture in the UK. The key product terms are along the lines of; oak bed
Intermediate & Advanced SEO | | pretige12
oak dining table
oak chairs etc which are very popular search terms. Im used to being able to hang of local or branded search which arent available in this project. The domain is around 12 months old, it has few inbound links which we are working on now with an overseas freelancer (keeping an eye on them). The pages of the site are unique-content rich and the titles and other meta boxes are ticked. What else would the experts recommend?0 -
Ranking a site in the USA
I'm UK based and looking at setting up a site to rank in the USA. As I understand it a .com TLD is best but these are used worldwide so do I simply need to set the geotargeting to USA in webmaster tools? Or is there a better domain to use? With hosting the site in US and on page content related to US cities (I plan to create a page for each US city I operate in the the city name in the H1 tag) will that be enough for google to understand that the page should rank in the US version of google. Also how can I view Google USA search results - when I go to google.com it automatically redirects to google.co.uk and I can only change the location on the left hand side to UK cities. Any help much appreciated!
Intermediate & Advanced SEO | | SamCUK0 -
Keyword Ranking Question
I have recently hired a SEO company to help with our keyword. My question is what are the best tools to use to verify what that are reporting. I can do an unpersonalized search, but I am likely still getting the my local results. I have been using the SEOmoz rank tracker in the past but for some reason it is not able to retrieve results over the past day or so. Are there any other good tools to check ranking for an exact url at the for non-localized, non personalized results? Thanks for the suggestions.
Intermediate & Advanced SEO | | fertilityhealth0