What is the best way to eliminate this specific image low lying content?
-
The site in question is www.homeanddesign.com where we are working on recovering from some big traffic loss.
I finally have gotten the sites articles properly meta titled and descriptioned now I'm working on removing low lying content.
The way there CMS is built, images have their own page (every one that's clickable). So this leads to a lot of thin content that I think needs to be removed from the index. Here is an example:
http://www.homeanddesign.com/photodisplay.asp?id=3633
I'm considering the best way to remove it from the index but not disturb how users enjoy the site.
What are my options? Here is what I'm thinking:
-
add Disallow: /photodisplay to the robots.txt file
-
See if there is a way to make a lightbox instead of a whole new page for images. But this still leaves me with 100s of pages with just an image on there with backlinks, etc.
-
Add noindex tag to the photodisplay pages
-
-
Disallow: /photodisplay.asp?*
That should do it. But just to be safe you can add another one for:
Disallow: /photodisplay.asp
There is very, very, very, very little danger of you blocking your entire site from being crawled if you add those disallow statements to your robots.txt file. If you're an SEO your job is to "mess with" the robots.txt file. Furthermore, trying to dynamically change the robots meta tag to noindex based on page-type is going to be much more tricky and potentially dangerous than adding a line to the robots.txt file.
Don't forget to remove the pages from the index using the URL removal tool in GWT once the block has been added.
Also I'd stop linking to those pages. It is best practice not to link to pages that you don't want indexed if you can help it. I'd go the lightbox route you mentioned above. This is something I do on my Wordpress sites too.
Good luck!
-
Hi WIlliam,
I would personally go the route of adding the noindex tag to the photo pages. Messing with the robots.txt file would probably be quicker; however, I am a little hesitant about messing with the robots.txt tag if I don't have to... one slip and you could be blocking your whole site or an entire directory from being crawled vs specifically calling out each individual page using the noindex tag.
Lightboxes are fine, but like you say, you aren't really solving the problem of tons of other pages.
You could look into your CMS and see if there is a way to remove the automatically generated link to photodisplay.aspXXXX so that the images are still displayed with , but it doesn't add the <a href="">... you know?</a>
<a href="">Hope this helps.
Mike</a>
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
PDF best practices: to get them indexed or not? Do they pass SEO value to the site?
All PDFs have landing pages, and the pages are already indexed. If we allow the PDFs to get indexed, then they'd be downloadable directly from google's results page and we would not get GA events. The PDFs info would somewhat overlap with the landing pages info. Also, if we ever need to move content, we'd now have to redirects the links to the PDFs. What are best practices in this area? To index or not? What do you / your clients do and why? Would a PDF indexed by google and downloaded directly via a link in the SER page pass SEO juice to the domain? What if it's on a subdomain, like when hosted by Pardot? (www1.example.com)
Reporting & Analytics | | hlwebdev1 -
Free Media Site / High Traffic / Low Engagement / Strategies and Questions
Hi, Imagine a site "mediapalooza dot com" where the only thing you do there is view free media. Yet Google Analytics is showing the average view of a media page is about a minute; where the average length of media is 20 - 90 minutes. And imagine that most of this media is "classic" and that it is generally not available elsewhere. Note also that the site ranks terribly in Google, despite having decent Domain Authority (in the high 30's), Page Authority in the mid 40's and a great site and otherwise quite active international user base with page views in the tens of thousands per month. Is it possible that GA is not tracking engagement (time on site) correctly? Even accounting for the imperfect method of GA that measures "next key pressed" as a way to terminate the page as a way to measure time on page, our stats are truly abysmal, in the tenths of a percentage point of time measured when compared with actual time we think the pages are being used. If so, will getting engagement tracking to more accurately measure time on specif pages and site signal Google that this site is actually more important than current ranking indicates? There's lots of discussion about "dwell time" as this relates to ranking, and I'm postulating that if we can show Google that we have extremely good engagement instead of the super low stats that we are reporting now, then we might get a boost in ranking. Am I crazy? Has anyone got any data that proves or disproves this theory? as I write this out, I detect many issues - let's have a discussion on what else might be happening here. We already know that low engagement = low ranking. Will fixing GA to show true engagement have any noticeable impact on ranking? Can't wait to see what the MOZZERS think of this!
Reporting & Analytics | | seo_plus0 -
Best way to block spambots in htaccess
I would like to block Russian Federation, China and Ukraine spam as well as semalt and buttonsforwebsite. I have come up with the following code, what do you think? For the countries: BLOCK COUNTRY DOMAINS RewriteCond %{HTTP_REFERER} .(ru|cn|ua)(/|$) [NC]
Reporting & Analytics | | ijb
RewriteRule .* - [F] And for buttons-for-website.com and semalt-semalt.com: BLOCK REFERERS RewriteCond %{HTTP_REFERER} (semalt|buttons) [NC]
RewriteRule .* - [F] or should it be: BLOCK USER AGENTS RewriteCond %{HTTP_USER_AGENT} (semalt|buttons) [NC]
RewriteRule .* - [F] Could I add (semalt|buttons|o-o-6-o-o|bestwebsitesawards|humanorightswatch) or is that too many?0 -
Universal cookie loss specific traffic
Hi there, A month ago we have installed the Universal Analytics tracking code on our website (www.smartphonehoesjes.nl). Since that time, direct traffic has risen unnaturally and organic traffic dropped unnaturally. I dived into this problem and I discovered there is a cookie loss on when visitors are switching from /cart to the first step checking out. The problem is that not all browsers cause these cookie issues, only Safari and IE older versions (9 and older) and some cases of Chrome. All iPads cause cookie issues (50% of all cases). iPhones cause not a single problem because we are using a separate mobile (m.) site which does not cause cookie problems. We don't have a mobile site for tablets yet, soon it will come. My biggest question is: how can I fix the Safari desktop cookie loss after switching domains from /cart to the first step of the chckout process? Olders versions of IE will slowly fade out and the iPad problem will be fixed when the mobile website is available for tablets. I hope someone can help me.
Reporting & Analytics | | MarcelMoz
Thanks in advance. Marcel0 -
Google Analytics is treating my blog like all the content is just on the home page.
Hello all, I installed Google Analytics on a main website and a blog (blog.travelexinsurance.com) While it appears to be tracking correctly (and when I test it in real time it shows that I'm visiting) but it is treating the entire blog as though it's one page. So I can't see data on blog post X. All I see is that X visitors came to my blog in aggregate. So I see blog.travelex.com has 999 visitors, but it doesn't show that /travel-luggage got 50 visits, while /insurace-tips got 75 and so forth. I assume I screwed up the tracking somehow, but can't figure out where I went wrong. Tracking on the main domain works just fine. It's specific to the blog.
Reporting & Analytics | | Patrick_G0 -
Is there a way to do this sort of keyword research?
I'd like to be able to do some keyword research as follows: Let's say I have two (or maybe more) separate keyword phrases I want to rank, is there a way to find out what combination and/or additional keyword phrases/descriptors are searched for the most? For example, let's say I want to rank well for the following terms. "Green widgets" and "boxes" Is there a tool I can use that will give me ideas/suggestions on how to use those two keywords the best way? Especially showing me long tell keyword suggestions I may not even know about. The tool may show me things people search for like "best boxes for your green widgets" gets more serach results than "best boxes for my green widgets." (Helping me figure out which combo of words in a long tail phrase is better. Or it may tell me a suggestion I didn't think about like "how to build boxes for your green widgets," or "how to put your green widgets into boxes," or "where to find boxes for your green widgets" or other combinations of things people search for using those two phrases? This may be possable using google keyword tools, but if it is, I'm not sure how to do it. Looking forward for any feedback. Thanks!
Reporting & Analytics | | NoahsDad0 -
Increase in 'Googlebot-Image' visits in analytics
Hi, I noticed a substantial increase in 'Googlebot-Image' visits data under Technology>Browser & OS in Google analytics for a few clients. Is this a bug? Are there any known fixes apart from just adding a filter to exclude the data? Regards Niladri
Reporting & Analytics | | neildomain0 -
Setting up Analytics on a Site that Uses Frames For Some Content
I work with a real estate agent and he uses strings from another tool to populate the listings on his site. In an attempt to be able to track traffic to both the framed pages and the non-framed pages he has two sets of analytics code on his site - one inside the frame and one for the regular part of the site. (there's also a third that the company who hosts his site and provides all these other tools put on his site - but I don't think that's really important to this conversation). Not only is it confusing looking at the analytics data, his bounce rate is down right unmanageable. As soon as anyone clicks on any of the listings they've bounced away. Here's a page - all of those listings below " Here are the most recent Toronto Beaches Real Estate Listings" are part of a frame. http://eastendtorontohomes.com/toronto-beach-real-estate-search/ I'm not really sure what to do about it or how to deal with it? Anyone out there got any good advice? And just in case you're wondering there aren't any other options - apart from spending thousands to build his own database thingie. We've thought about that (as other agents in the city have done that), but just aren't sure it's worth it. And, quite frankly he doesn't want to spend the money.
Reporting & Analytics | | annasus0