Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Will having image lightbox with content on a web page SEO friendly?
-
This website is done in CMS. Will having lightbox pop up with content be SEO friendly?
If you go to the web page and click on the images at the bottom of the page. There are lightbox that will display information. Will these lightbox content information be crawl by Google? Will it be consider as content for the url http://jennlee.com/portfolio/bran..
Thanks,
John
-
Hi Dale,
Really stupid question, how do I look at the CSS to identify that? I've viewed source but cant see that information anywhere on the page.
If you wouldn't mind, could you point me in the right direction of some information about this issue, I would be interested in understanding it better, but until you brought it to my attention, I had no idea even to look for it
J
-
Ryan and James,
Take a closer look at the div class of the lightbox (class="contact"). In the CSS for the page in question we find the following:
div.contact {
display: none;
visibility: hidden;}
In my opinion, you're asking the wrong question. This isn't about lightboxes or DA at all; it's about the display:none; and visibility: hidden; elements.
There is no shortage of information about that here on SEOmoz or in the Google Webmaster Forums.
-
Interesting supposition. i've got absolutely no idea if a stronger page changes the specific parts of a page are parsed.
Shouldn't be too difficult to work out though:
If we work on the logic that an exact match search result indicates that the text is being read and used by google, you can then compare javascript parsing across strong and weak pages.
Another way would be to look at the cached text only version across pages and see if there is any difference, although I think I prefer the first suggestion
Seems simple, although it probably isn't
j
-
I agree with your assessment James.
Before I accept this information I would like to ask if you are aware of any other similar examples of lightbox use on a page with better stats? The DA of this page is only 31, and PA is 1. I would like to rule out the idea Google may crawl deeper if the page was deemed more important.
-
James is correct. Your lightbox content is not visible to a Google Bot.
You can see from an exact match search of some text from the page that Google has indexed the visible text: http://bit.ly/nDQLlM
The only place that the exact text from the lightbox appears in the Google index is on this thread: http://bit.ly/mRQICc
-
Sorry for butting in on an old(ish) post, but I have a different opinion on this...
Correct the text used in the example does show up in the source code as HTML, but I dont think that indicates that google is reading that text.
For me there are two ways to check to see if Google is reading text:
1. Do an exact match (quotation marked) search in google.
2. Look at the cached version of the page in google in text only version.
From that information, the lightbox data is not showing up and for me that would indicate that the text is not being read.
Also, an interesting point to note is that 'Fetch as Googlebot' should not be used as a method of identifying what text is being parsed according to searchengineland http://searchengineland.com/see-what-googlebot-sees-on-your-site-27623
Feel free to prove me wrong!
thanks
james
-
I have read that article before. Keep in mind it is from 2008. Technology and Google have advanced substantially in the past 3 years.
100% of the text in all your lightbox is fully viewable by Google presently. William and I both looked and we see the text in your html source code. That means Google can see it as well.
-
Those are not issues on your site.
Your light box images are fully crawlable. Google sees all of the images and the text descriptions. You definitely want to add an ALT description. Otherwise you are in great shape.
-
thanks for all the responses guys.
my thoughts were most of the time it depends upon the script because some script hide data from the viewers while it shows the same data to Search Engine which turns out Clocking issue on website.. this could be proved very dangerous for the website.
Also seems like google does not crawl the images as often than normal web page.. because it hide the contents and creates unauthenticated website.
-
Sure thing brother!
-
Thank you William. Somehow I missed it during my review of the source code.
-
Hi Ryan,
Yes, I just did a search for the text I found in the Lightbox description for the Coco & Max logo. Right there. I've attached a couple images to show what I found.
Is this underneath a Javascript? I'd be interested to learn about the differences between different scripts as I see myself building sites that I would like to use the most SEO beneficial one.
-
Hi William.
Thanks for the feedback. I did look at the HTML and the real text is NOT visible. I am pretty sure that Google can read it even in the javascript, but I am not certain so I did not wish to offer that conclusively. If I knew which version was in use, such as Highslide, I could check and offer a confirmation.
The first image shared is the Coco and Max logo. If you click on that image the Lightbox will appear with a description that says "The Jenn Lee Group developed photography, business cards, expo-banner plus an ecommerce website for Coco and Max using a logo they had already developed. The Jenn Lee Group can pick up the ball at whatever stage you are currently in towards your marketing and advertising initiatives. Call us today! 401-885-3200"
I do not see that text snippet anywhere in the page's source code. Also, there are a total of 7 pictures offered in a group with that first image, each which their own text.
If you have any additional information, I would love to learn as well.
-
Lightbox should have zero negative impact in regards to SEO, providing you have effectively labeled your photos. I love the look of it, and although has a similar effect to flash, they have nothing to do with eachother in regards to negative SEO.
-
Hey Ryan,
The Original Poster is actually talking about the text descriptions of each logo that is listed.
The easy way to figure this out is to look in the HTML. If it's real text, then Google can crawl it. In your case it is.
So the content you have will be indexed.And you can do as Ryan suggested and add Alt Attribute to each image. It will help as well.
-
The biggest gap I see on your site is your images are all missing ALT tags. Search engines don't see images the way people do. By providing an alt tag, you can offer a description of each image. For example your first image alt tag might be "logo Coco & Max Doggie Distinctions".
There are many packages of javascript code which use Lightbox so if you want a more definite answer you would need to take a look at your specific package. Highslide and Suckerfish are two examples of Lightbox javascript coding packages. For additional research you can check out this article.
Another note. I would recommend changing your Meta description to readable text, not a list of key words. Your meta description is what people will see as your listing in search engines. It will not affect your search result ranking.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Will I loose from SEO if I rename my urls to be more keyword friendly?
As a good practice of SEO is to have your keywords in the links. I am thinking of doing some optimization and change my urls to more effective keywords. I am using shopify and there is an option (a tick) that you can check while changing the url (ex. for a category, for a product, for a blog post). This will give a redirection to the old post to the new. Is it good practice? Is it risky for losing SEO or it will help to rank higher because I will have better keywords in my links?
Intermediate & Advanced SEO | | Spiros.im0 -
Will changing category URLs on site hurt SEO?
Hi Moz Community, We're looking to replace some URLs on our Wordpress site and I want to make sure we won't hurt our SEO with the changes. The site is lushpalm.com When we originally launched our site we created pages (which are linked to in our main menu) to essentially display our categories. We did this as a workaround because we didn’t like the URL to have the word “category” in it. Now we would like to make some changes and we want to make sure we’re not going to hurt our SEO in any way by accidentally duplicating content or otherwise. We want to fix our structure and now link to our category pages from our main menu, BUT we want to change the URL of the category page so that it doesn’t have “category” in it, essentially renaming it the name of the page currently linked to in our main menu. So basically, the category lushpalm.com/category/surf-trips, would be renamed with the URL lushpalm.com/surf-trips and the current page that is at lushpalm.com/surf-trips would be therefore replaced. My questions are: If we did this, would that mean that the previous “lushpalm.com/category/surf-trips” would cease to exist? Or is there some imprint of that out on the web? And if it is then would it re-direct to the new page? Would replacing the current page URL with a category hurt our current SEO in any way? Would this change cause any duplicate pages somehow? Thanks so much for your help!
Intermediate & Advanced SEO | | TaraLP1 -
Will 301 Redirects Slow Page Speed?
We have a lot of subdomains that we are switching to subfolders and need to 301 redirect all the pages from those subdomains to the new URL. We have over 1000 that need to be implemented. So, will 301 redirects slow the page speed regardless of which URL the user comes through? Or, as the old urls are dropped from Google's index and bypassed as the new URLs take over in the SERPs, will those redirects then have no effect on page speed? Trying to find a clear answer to this and have yet to find a good answer
Intermediate & Advanced SEO | | MJTrevens0 -
If Robots.txt have blocked an Image (Image URL) but the other page which can be indexed has this image, how is the image treated?
Hi MOZers, This probably is a dumb question but I have a case where the robots.tags has an image url blocked but this image is used on a page (lets call it Page A) which can be indexed. If the image on Page A has an Alt tags, then how is this information digested by crawlers? A) would Google totally ignore the image and the ALT tags information? OR B) Google would consider the ALT tags information? I am asking this because all the images on the website are blocked by robots.txt at the moment but I would really like website crawlers to crawl the alt tags information. Chances are that I will ask the webmaster to allow indexing of images too but I would like to understand what's happening currently. Looking forward to all your responses 🙂 Malika
Intermediate & Advanced SEO | | Malika11 -
Trailing Slashes for Magento CMS pages - 2 URLS - Duplicate content
Hello, Can anyone help me find a solution to Fixing and Creating Magento CMS pages to only use one URL and not two URLS? www.domain.com/testpage www.domain.com/testpage/ I found a previous article that applies to my issue, which is using htaccess to redirect request for pages in magento 301 redirect to slash URL from the non-slash URL. I dont understand the syntax fully in htaccess , but I used this code below. This code below fixed the CMS page redirection but caused issues on other pages, like all my categories and products with this error: "This webpage has a redirect loop ERR_TOO_MANY_REDIRECTS" Assuming you're running at domain root. Change to working directory if needed. RewriteBase / # www check If you're running in a subdirectory, then you'll need to add that in to the redirected url (http://www.mydomain.com/subdirectory/$1 RewriteCond %{HTTP_HOST} !^www. [NC]
Intermediate & Advanced SEO | | iamgreenminded
RewriteRule ^(.*)$ http://www.mydomain.com/$1 [R=301,L] Trailing slash check Don't fix direct file links RewriteCond %{REQUEST_FILENAME} !-f RewriteCond %{REQUEST_URI} !(.)/$
RewriteRule ^(.)$ $1/ [L,R=301] Finally, forward everything to your front-controller (index.php) RewriteCond %{REQUEST_FILENAME} !-f
RewriteCond %{REQUEST_FILENAME} !-d
RewriteRule .* index.php [QSA,L]0 -
Images with a token in the url, in Drupal. How does it affect to SEO?
Hi everyone! I am checking now a website that works with Drupal, and I found that images have urls like this... http://www.brandname.com/sites/default/files/styles/directory_xyz/public/name-of-the-picture.png?itok=T89RpzrK I was wondering how an URL like that with the token at the and, can affect to SEO. I cound't find anything. Anyone knows? Thank you!
Intermediate & Advanced SEO | | teconsite0 -
International SEO - cannibalisation and duplicate content
Hello all, I look after (in house) 3 domains for one niche travel business across three TLDs: .com .com.au and co.uk and a fourth domain on a co.nz TLD which was recently removed from Googles index. Symptoms: For the past 12 months we have been experiencing canibalisation in the SERPs (namely .com.au being rendered in .com) and Panda related ranking devaluations between our .com site and com.au site. Around 12 months ago the .com TLD was hit hard (80% drop in target KWs) by Panda (probably) and we began to action the below changes. Around 6 weeks ago our .com TLD saw big overnight increases in rankings (to date a 70% averaged increase). However, almost to the same percentage we saw in the .com TLD we suffered significant drops in our .com.au rankings. Basically Google seemed to switch its attention from .com TLD to the .com.au TLD. Note: Each TLD is over 6 years old, we've never proactively gone after links (Penguin) and have always aimed for quality in an often spammy industry. **Have done: ** Adding HREF LANG markup to all pages on all domain Each TLD uses local vernacular e.g for the .com site is American Each TLD has pricing in the regional currency Each TLD has details of the respective local offices, the copy references the lacation, we have significant press coverage in each country like The Guardian for our .co.uk site and Sydney Morning Herlad for our Australia site Targeting each site to its respective market in WMT Each TLDs core-pages (within 3 clicks of the primary nav) are 100% unique We're continuing to re-write and publish unique content to each TLD on a weekly basis As the .co.nz site drove such little traffic re-wrting we added no-idex and the TLD has almost compelte dissapread (16% of pages remain) from the SERPs. XML sitemaps Google + profile for each TLD **Have not done: ** Hosted each TLD on a local server Around 600 pages per TLD are duplicated across all TLDs (roughly 50% of all content). These are way down the IA but still duplicated. Images/video sources from local servers Added address and contact details using SCHEMA markup Any help, advice or just validation on this subject would be appreciated! Kian
Intermediate & Advanced SEO | | team_tic1 -
Blocking Pages Via Robots, Can Images On Those Pages Be Included In Image Search
Hi! I have pages within my forum where visitors can upload photos. When they upload photos they provide a simple statement about the photo but no real information about the image,definitely not enough for the page to be deemed worthy of being indexed. The industry however is one that really leans on images and having the images in Google Image search is important to us. The url structure is like such: domain.com/community/photos/~username~/picture111111.aspx I wish to block the whole folder from Googlebot to prevent these low quality pages from being added to Google's main SERP results. This would be something like this: User-agent: googlebot Disallow: /community/photos/ Can I disallow Googlebot specifically rather than just using User-agent: * which would then allow googlebot-image to pick up the photos? I plan on configuring a way to add meaningful alt attributes and image names to assist in visibility, but the actual act of blocking the pages and getting the images picked up... Is this possible? Thanks! Leona
Intermediate & Advanced SEO | | HD_Leona0