Are Google now indexing iFrames?
-
A client is pulling content through an iFrame, and when searching for a snippet of that exact content the page that is pulling the data is being indexed and not the iFrame page. Seen this before?
-
Yeah, I use iframes and if I want to be sure they are NOT indexed, I Just add a "noindex" tag. You may also want to add a "nofollow" tag to avoid spiders to follow links inside the frame. Using iframes may be a good idea to reduce the number of links on a page (Bruce Clay suggestion).
-
I've never seen it before, but like everyone here said, it's not a good idea.
This makes me wonder though:
1. Can you find the original page using a snippet? And if not:
2. Is the page contained in the iframe indexed? (Or better-phrased, is the page that is being framed "noindex"?)
It makes sense to me that if the framed page is noindex, that Google would index the content and attribute it to the page framing it.
One perfect example:
I embed videos using an iframe and then I make the video unlisted in YouTube. My embedded content is indexed and even displayed as a rich snippet....
-
I have noticed content within iFrames being indexed by google and text within those iFrames being attributed to the page/url that is hosting the iFrame. Not sure how often this applies. I avoid iFrames.
Merchant Circle uses them and their pages get credit for content in them.
-
It might have been covered but it does seem that google is ignoring iframes in relation to commets code posted on sites.for instance: our text cached version.: http://webcache.googleusercontent.com/search?q=cache:8IZ95GICp7AJ:gaveltek.com/seoblog/&hl=en&gl=us&strip=1
compare the page title to (use headers it easier)
www.gaveltek.com/seoblog the list "comments" and despite there being some the are not posted. However, I do believe general wordpress comments hold some weight. That is not to sayt that facebook comments do not, its just done via different metrics, like social, and trust, and egngagement.
Cheers
TODD
-
A good way to check is go to google.com and type in your full URL like this:
site:www.domain.com
Then you will be populated with your sites pages of course. Now there is a link there that says: "cache" and you can see what it cached.
I think they may be getting better at knowing what's in a iframe. Look at how many sites use facebook comments on the blog and how do you think thats ran? iframes. Do you remember google and adobe working together at reading .pdf's and flash.
The little magnifying glass has some cool technology behind it that I'm sure helped them know whats really on the site. Without getting to far off track I do feel like they are better at reading iframes. Just my .02c in this thread.
-
last thought... i've only ever used iframes in the aforementioned example. Not an ideal way to display your original content if you want it indexed.
-
It is very typical for Google to ignore iframes. I don't know the precise details of your situation but there are several reasons for iframing that might make sense - this is situational - so no hating!
-
you're an affiliate and using another offer (conversion form) that you have to iframe to generate leads, etc
-
you want to hide duplicate content that appears elsewhere on the site (although there are far more elegant ways to do this)
3)You're pulling video or other syndicated content from a publisher who wants to maintain control (ie not let you outrank them with their own content)
*** Remember that the iframed content can certainly be indexed but usually only from the destination URL's originating source. For example: You are www.insuranceaffilifate.com running an offer from www.insurance.com/form_1011 - you will most likely use insurance.com's form via iframe on your landing page. That form, unless it uses a NOINDEX meta tag, will likely be picked by the search engines from www.insurance.com but will be ignored on your site www.insuranceaffiliate.com.
Hope this helps.
-
-
I have to agree with Julich in that you should move the content to be truly located on www.domain.com instead of iframe.domain.com.
-
I totally agree that they shouldn't be using iFrames and it is part of my recommendations to them, but we need to work with what we have at the moment.
So just to clarify, you would say that www.domain.com which is pulling the data through from iframe.domain.com would rank?
Even though all the content except the navigation, footer, etc is on iframe.domain.com.
-
Normally, it would be www.domain.com (unless it doesn't provide any content outside the iFrame).
But it is not abnormal to also see iframe.domain.com in the SERPS, since it may have some backlinks pointing to it.
Anyway, using iframes is a weird technique and I recommend you merge those into www.domain.com if possible (and don't forget to do some 301 redirections to tell Google your pages have definitely moved to www.domain.com).
-
OK, so if www.domain.com was pulling through content from iframe.domain.com which domain would you expect to rank?
I would personally expect iframe.domain.com to rank as that is actually where the content is and the www.domain.com provides the link to that page. I am currently seeing both domains rank, which has lead me to ask the question.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Site indexed by Google, but (almost) never gets impressions
Hi there, I have a question that I wasn't able to give it a reasonable answer yet, so I'm going to trust on all of you. Basically a site has all its pages indexed by Google (I verified with site:sitename.com) and it also has great and unique content. All on-page grades are A with absolutely no negative factors at all. However its pages do not get impressions almost at all. Of course I didn't expect it to be on page 1 since it has been launched on Dec, 1st, but it looks like Google is ignoring (or giving it bad scores) for some reason. Only things that can contribute to that could be: domain privacy on the domain, redirect from the www to the subdomain we use (we did this because it will be a multi-language site, so we'll assign to each country a subdomain), recency (it has been put online on Dec 1st and the domain is just a couple of months old). Or maybe because we blocked crawlers for a few days before the launch? Exactly a few days before Dec 1st. What do you think? What could be the reason for that? Thanks guys!
Technical SEO | | ruggero0 -
Google Crawling Issues! How Can I Get Google to Crawl My Website Regularly?
Hi Everyone! My website is not being crawled regularly by Google - there are weeks when it's regular but for the past month or so it does not get crawled for seven to eight days. There are some specific pages, that I want to get ranked but they of late are not being crawled AT ALL unless I use the 'Fetch As Google' tool! That's not normal, right? I have checked and re-checked the on-page metrics for these pages (and the website as a whole, backlinking is a regular and ongoing process as well! Sitemap is in place too! Resubmitted it once too! This issue is detrimental to website traffic and rankings! Would really appreciate insights from you guys! Thanks a lot!
Technical SEO | | farhanm1 -
How to Remove /feed URLs from Google's Index
Hey everyone, I have an issue with RSS /feed URLs being indexed by Google for some of our Wordpress sites. Have a look at this Google query, and click to show omitted search results. You'll see we have 500+ /feed URLs indexed by Google, for our many category pages/etc. Here is one of the example URLs: http://www.howdesign.com/design-creativity/fonts-typography/letterforms/attachment/gilhelveticatrade/feed/. Based on this content/code of the XML page, it looks like Wordpress is generating these: <generator>http://wordpress.org/?v=3.5.2</generator> Any idea how to get them out of Google's index without 301 redirecting them? We need the Wordpress-generated RSS feeds to work for various uses. My first two thoughts are trying to work with our Development team to see if we can get a "noindex" meta robots tag on the pages, by they are dynamically-generated pages...so I'm not sure if that will be possible. Or, perhaps we can add a "feed" paramater to GWT "URL Parameters" section...but I don't want to limit Google from crawling these again...I figure I need Google to crawl them and see some code that says to get the pages out of their index...and THEN not crawl the pages anymore. I don't think the "Remove URL" feature in GWT will work, since that tool only removes URLs from the search results, not the actual Google index. FWIW, this site is using the Yoast plugin. We set every page type to "noindex" except for the homepage, Posts, Pages and Categories. We have other sites on Yoast that do not have any /feed URLs indexed by Google at all. Side note, the /robots.txt file was previously blocking crawling of the /feed URLs on this site, which is why you'll see that note in the Google SERPs when you click on the query link given in the first paragraph.
Technical SEO | | M_D_Golden_Peak0 -
Indexing Issue
Hi, I am working on www.stjohnswaydentalpractice.co.uk Google only seems to be indexing two of the pages when i search site:www.stjohnswaydentalpractice.co.uk I have added the site to webmaster tools and created a new sitemap which is showing that it has only submitted two of the pages. Can anyone shed any light for why these pages are not being indexed? Thanks Faye
Technical SEO | | dentaldesign0 -
Best way to handle indexed pages you don't want indexed
We've had a lot of pages indexed by google which we didn't want indexed. They relate to a ajax category filter module that works ok for front end customers but under the bonnet google has been following all of the links. I've put a rule in the robots.txt file to stop google from following any dynamic pages (with a ?) and also any ajax pages but the pages are still indexed on google. At the moment there is over 5000 pages which have been indexed which I don't want on there and I'm worried is causing issues with my rankings. Would a redirect rule work or could someone offer any advice? https://www.google.co.uk/search?q=site:outdoormegastore.co.uk+inurl:default&num=100&hl=en&safe=off&prmd=imvnsl&filter=0&biw=1600&bih=809#hl=en&safe=off&sclient=psy-ab&q=site:outdoormegastore.co.uk+inurl%3Aajax&oq=site:outdoormegastore.co.uk+inurl%3Aajax&gs_l=serp.3...194108.194626.0.194891.4.4.0.0.0.0.100.305.3j1.4.0.les%3B..0.0...1c.1.SDhuslImrLY&pbx=1&bav=on.2,or.r_gc.r_pw.r_qf.&fp=ff301ef4d48490c5&biw=1920&bih=860
Technical SEO | | gavinhoman0 -
Google Page speed
I get the following advice from Google page speed: Suggestions for this page The following resources have identical contents, but are served from different URLs. Serve these resources from a consistent URL to save 1 request(s) and 77.1KiB. http://www.irishnews.com/ http://www.irishnews.com/index.aspx I'm not sure how to fix this the default page is http://www.irishnews.com/index.aspx, anybody know what need to be done please advise. thanks
Technical SEO | | Liammcmullen0 -
Google Index Speed Opinions
Hello Everyone, Under normal circumstances, new posts to my site are indexed almost instantly by Google. I know this because an occasional search with quotation marks surrounding the 1st paragraph of text displays my newly published page. I use this tactic from time to time to ensure contributors aren't syndicating content. My question is this: I've noticed over the last day or so that my newly published articles are not yet indexed. For example, an article that was published over 24 hours ago does not appear to be indexed yet. Is this cause for concern? Is there an average wait time for indexation? XML issue? Thanks in advance for the help/insight.
Technical SEO | | JSOC0 -
Removing a site from Google's index
We have a site we'd like to have pulled from Google's index. Back in late June, we disallowed robot access to the site through the robots.txt file and added a robots meta tag with "no index,no follow" commands. The expectation was that Google would eventually crawl the site and remove it from the index in response to those tags. The problem is that Google hasn't come back to crawl the site since late May. Is there a way to speed up this process and communicate to Google that we want the entire site out of the index, or do we just have to wait until it's eventually crawled again?
Technical SEO | | issuebasedmedia0