Moz Q&A is closed.
After more than 13 years, and tens of thousands of questions, Moz Q&A closed on 12th December 2024. Whilst we’re not completely removing the content - many posts will still be possible to view - we have locked both new posts and new replies. More details here.
Unsolved Why My site pages getting video index viewport issue?
-
Hello,
I have been publishing a good number of blogs on my site Flooring Flow. Though, there's been an error of the video viewport on some of my articles.
I have tried fixing it but the error is still showing in Google Search Console.
Can anyone help me fix it out?
-
@RyanPurkey Dear Sir, My theme, and setting all is good but I am also receiving the same issue. Here are a few links that you can check, please suggest the best solution.
https://theshoy.com/advantages-and-disadvantages-of-motorcycle/
https://theshoy.com/advantages-and-disadvantages-of-smartphones/
Check the attachment below from my search console.
-
this is the link to my webstory https://stayenergies.com/web-stories/benefit-of-walking/ that getting issues with "video outside the viewport"
it is created in WordPress; is there any plugin or setting to solve it.
-
Viewport Configuration: The viewport meta tag is essential for responsive web design, as it tells the browser how to adjust the page's dimensions Ehsaas program and scale to fit different screen sizes. If the viewport is not configured correctly for video elements, search engines may have difficulty indexing and displaying videos properly.
Lazy Loading: Lazy loading is a technique used to improve page load times by deferring the loading of non-essential content, such as images and videos, until the user scrolls to that section of the page. If lazy loading is not correctly implemented for videos, it can lead to indexing issues.
Missing Video Sitemaps: Sitemaps are files that provide information about the content on your website to search engines. If you have videos on your site, it's essential to have a video sitemap that provides details about the videos, such as their titles, descriptions, and URLs. This helps search engines index your videos more effectively.
Accessibility Issues: Video content should be accessible to all users, including those with disabilities. If your videos lack proper captions, transcripts, or other accessibility features, it may lead to indexing issues.
-
If your site pages are experiencing a "video index viewport issue," it likely refers to problems related to how videos are indexed and displayed within the viewport on your website. The viewport is the visible area of a web page that a user can see without scrolling. Without further information, it's challenging to pinpoint the exact issue, but here are some common reasons and potential solutions for video index viewport problems:
Responsive Design: Ensure that your website has a responsive design, meaning it adapts and displays correctly on various screen sizes and devices. If the video container's size isn't adjusting properly to different viewports, the video may be cut off or not visible entirely.
Viewport Meta Tag: Make sure your web pages include the viewport meta tag in the <head> section. This tag helps the browser to correctly scale and render the content to fit different screen sizes.
Example viewport meta tag: <meta name="viewport" content="width=device-width, initial-scale=1">
Video Element Attributes: Check the HTML attributes of the video element. Setting the width and height attributes can help the browser allocate the correct space for the video player.Example video element:<video width="640" height="360" controls>
<source src="video.mp4" type="video/mp4">
Your browser does not support the video tag.
</video>
CSS Issues: Review your CSS styles to ensure that they don't interfere with the video's display. Sometimes, conflicting styles or positioning can lead to improper video rendering.Lazy Loading: If you are using lazy loading for videos, verify that it's implemented correctly. Lazy loading delays the loading of the video until it's about to come into the user's viewport, which might affect how it's indexed and displayed.
JavaScript Conflicts: Check if any JavaScript on your site is causing conflicts with video players or the viewport settings.
Browser Compatibility: Test your site on various browsers to ensure that the issue isn't browser-specific.
Video Format: Ensure that your videos are in a format supported by major browsers, such as MP4 with H.264 video codec and AAC audio codec.
Server Configuration: In some cases, server misconfigurations may lead to video loading issues. Ensure that your server is correctly configured to serve videos.
Testing and Debugging: Use browser developer tools to inspect the video element and check for any errors or warnings related to the video player or viewport.
Click here : Web Development company in Bangladesh -
the same issue is also occurring in my web stories when uploading any video to them. How to solve it. I have also added a video sitemap, also reduces the video size but does not solve the issue.
-
@mitty27 Your iFrame at the top of the page is hard set to 1200px width which could cause problems. Please provide some specific URLs that GSC has identified with the viewport issue for specific answers to those. Thanks and good luck!
-
Your site pages might be encountering Prestige Park Grove video index viewport issues due to improper implementation of the video index or viewport settings. It can lead to videos not displaying correctly within the visible area, affecting user experience. Check and adjust your site's code to ensure proper video indexing and viewport configuration for optimal performance.
Got a burning SEO question?
Subscribe to Moz Pro to gain full access to Q&A, answer questions, and ask your own.
Browse Questions
Explore more categories
-
Moz Tools
Chat with the community about the Moz tools.
-
SEO Tactics
Discuss the SEO process with fellow marketers
-
Community
Discuss industry events, jobs, and news!
-
Digital Marketing
Chat about tactics outside of SEO
-
Research & Trends
Dive into research and trends in the search industry.
-
Support
Connect on product support and feature requests.
Related Questions
-
Unsolved Question about a Screaming Frog crawling issue
Hello, I have a very peculiar question about an issue I'm having when working on a website. It's a WordPress site and I'm using a generic plug in for title and meta updates. When I go to crawl the site through screaming frog, however, there seems to be a hard coded title tag that I can't find anywhere and the plug in updates don't get crawled. If anyone has any suggestions, thatd be great. Thanks!
Technical SEO | | KyleSennikoff0 -
Unsolved Capturing Source Dynamically for UTM Parameters
Does anyone have a tutorial on how to dynamically capture the referring source to be populated in UTM parameters for Google Analytics? We want to syndicate content and be able to see all of the websites that provided referral traffic for this specific objective. We want to set a specific utm_medium and utm_campaign but have the utm_source be dynamic and capture the referring website. If we set a permanent utm_source, it would appear the same for all incoming traffic. Thanks in advance!
Technical SEO | | peteboyd0 -
Solved How to solve orphan pages on a job board
Working on a website that has a job board, and over 4000 active job ads. All of these ads are listed on a single "job board" page, and don’t obviously all load at the same time. They are not linked to from anywhere else, so all tools are listing all of these job ad pages as orphans. How much of a red flag are these orphan pages? Do sites like Indeed have this same issue? Their job ads are completely dynamic, how are these pages then indexed? We use Google’s Search API to handle any expired jobs, so they are not the issue. It’s the active, but orphaned pages we are looking to solve. The site is hosted on WordPress. What is the best way to solve this issue? Just create a job category page and link to each individual job ad from there? Any simpler and perhaps more obvious solutions? What does the website structure need to be like for the problem to be solved? Would appreciate any advice you can share!
Reporting & Analytics | | Michael_M2 -
How to check if an individual page is indexed by Google?
So my understanding is that you can use site: [page url without http] to check if a page is indexed by Google, is this 100% reliable though? Just recently Ive worked on a few pages that have not shown up when Ive checked them using site: but they do show up when using info: and also show their cached versions, also the rest of the site and pages above it (the url I was checking was quite deep) are indexed just fine. What does this mean? thank you p.s I do not have WMT or GA access for these sites
Technical SEO | | linklander0 -
How Does Google's "index" find the location of pages in the "page directory" to return?
This is my understanding of how Google's search works, and I am unsure about one thing in specific: Google continuously crawls websites and stores each page it finds (let's call it "page directory") Google's "page directory" is a cache so it isn't the "live" version of the page Google has separate storage called "the index" which contains all the keywords searched. These keywords in "the index" point to the pages in the "page directory" that contain the same keywords. When someone searches a keyword, that keyword is accessed in the "index" and returns all relevant pages in the "page directory" These returned pages are given ranks based on the algorithm The one part I'm unsure of is how Google's "index" knows the location of relevant pages in the "page directory". The keyword entries in the "index" point to the "page directory" somehow. I'm thinking each page has a url in the "page directory", and the entries in the "index" contain these urls. Since Google's "page directory" is a cache, would the urls be the same as the live website (and would the keywords in the "index" point to these urls)? For example if webpage is found at wwww.website.com/page1, would the "page directory" store this page under that url in Google's cache? The reason I want to discuss this is to know the effects of changing a pages url by understanding how the search process works better.
Technical SEO | | reidsteven750 -
What is the best way to find missing alt tags on my site (site wide - not page by page)?
I am looking to find all the missing alt tags on my site at once. I have a FF extension that use to do it page by page, but my site is huge and that will take forever. Thanks!!
Technical SEO | | franchisesolutions1 -
De-indexing millions of pages - would this work?
Hi all, We run an e-commerce site with a catalogue of around 5 million products. Unfortunately, we have let Googlebot crawl and index tens of millions of search URLs, the majority of which are very thin of content or duplicates of other URLs. In short: we are in deep. Our bloated Google-index is hampering our real content to rank; Googlebot does not bother crawling our real content (product pages specifically) and hammers the life out of our servers. Since having Googlebot crawl and de-index tens of millions of old URLs would probably take years (?), my plan is this: 301 redirect all old SERP URLs to a new SERP URL. If new URL should not be indexed, add meta robots noindex tag on new URL. When it is evident that Google has indexed most "high quality" new URLs, robots.txt disallow crawling of old SERP URLs. Then directory style remove all old SERP URLs in GWT URL Removal Tool This would be an example of an old URL:
Technical SEO | | TalkInThePark
www.site.com/cgi-bin/weirdapplicationname.cgi?word=bmw&what=1.2&how=2 This would be an example of a new URL:
www.site.com/search?q=bmw&category=cars&color=blue I have to specific questions: Would Google both de-index the old URL and not index the new URL after 301 redirecting the old URL to the new URL (which is noindexed) as described in point 2 above? What risks are associated with removing tens of millions of URLs directory style in GWT URL Removal Tool? I have done this before but then I removed "only" some useless 50 000 "add to cart"-URLs.Google says themselves that you should not remove duplicate/thin content this way and that using this tool tools this way "may cause problems for your site". And yes, these tens of millions of SERP URLs is a result of a faceted navigation/search function let loose all to long.
And no, we cannot wait for Googlebot to crawl all these millions of URLs in order to discover the 301. By then we would be out of business. Best regards,
TalkInThePark0