CN106020461A - Video interaction method based on eyeball tracking technology - Google Patents
Video interaction method based on eyeball tracking technology Download PDFInfo
- Publication number
- CN106020461A CN106020461A CN201610322512.9A CN201610322512A CN106020461A CN 106020461 A CN106020461 A CN 106020461A CN 201610322512 A CN201610322512 A CN 201610322512A CN 106020461 A CN106020461 A CN 106020461A
- Authority
- CN
- China
- Prior art keywords
- video
- plug
- pupil
- method based
- label
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
- G06F3/013—Eye tracking input arrangements
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/43—Processing of content or additional data, e.g. demultiplexing additional data from a digital video stream; Elementary client operations, e.g. monitoring of home network or synchronising decoder's clock; Client middleware
- H04N21/431—Generation of visual interfaces for content selection or interaction; Content or additional data rendering
- H04N21/4312—Generation of visual interfaces for content selection or interaction; Content or additional data rendering involving specific graphical features, e.g. screen layout, special fonts or colors, blinking icons, highlights or animations
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/80—Generation or processing of content or additional data by content creator independently of the distribution process; Content per se
- H04N21/81—Monomedia components thereof
- H04N21/812—Monomedia components thereof involving advertisement data
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N5/00—Details of television systems
- H04N5/44—Receiver circuitry for the reception of television signals according to analogue transmission standards
- H04N5/445—Receiver circuitry for the reception of television signals according to analogue transmission standards for displaying additional information
- H04N5/45—Picture in picture, e.g. displaying simultaneously another television channel in a region of the screen
Abstract
The invention relates to a video interaction method based on an eyeball tracking technology. The video interaction method mainly comprises the following steps: A1) when a video picture with a preset plug-in tag is played, collecting the eyeball information of an audience through a camera in real time; A2) extracting the current space coordinate of the pupil from the collected eyeball images, and mapping the current space coordinate of the pupil to the two-dimensional coordinate of a video interface to determine the point of fall of the pupil on the video interface, wherein the center point and the original point of the coordinate of the video interface independently correspond to the space coordinate of the pupil center and the space coordinate of the inner corner of eye of the audience; and A3) after the point of fall of the pupil stays for preset time on the corresponding plug-in tag, analyzing and playing a video file corresponding to the plug-in tag. The video interaction method can cause a common video to become an interactive video, can conveniently realize interaction with audiences, and improves user experience.
Description
Technical field
The invention belongs to video interactive technical field, be specifically related to a kind of video interactive method based on eyeball tracking technology.
Background technology
Traditional Internet video generally plays out by embedding network player in a browser, is substantially in playing process
Unidirectional output, spectators only passively receive, thus lack the interactive of video and spectators both sides.This is for advertisement video,
Especially it is difficult to attract spectators well.
The interactive experience of spectators can be dissolved in linear video to form Novel video by HTML5 technological means, allows the spectators can be real
Existing video is internal to be clicked on, and also realizes the variation of video content simultaneously, but its most clearly disadvantageous be that spectators must manually go
Clicking on video, the lifting for Consumer's Experience is relatively limited.
Even additionally, currently also there is many in Novel video needs improvements, lifting especially needed for advertisement video.
Summary of the invention
Present invention is primarily targeted at, propose a kind of video interactive method based on eyeball tracking technology, real in mode more easily
Existing video is mutual with spectators', improves Consumer's Experience.
In order to realize object above, the present invention proposes a kind of video interactive method based on eyeball tracking technology, mainly includes following
Step:
A1, when being played to the video pictures being preset with plug-in label, by the eyeball image of photographic head Real-time Collection spectators;
A2, from the eyeball image collected, extract the space coordinates that pupil is current, and be mapped to the two-dimensional coordinate of video clip
On, to determine pupil drop point in video clip;The central point of the coordinate of described video clip, initial point respectively with spectators' pupil
The space coordinates at center, the space coordinates correspondence of inner eye corner;
A3, the drop point of pupil stop after reaching Preset Time on corresponding plug-in label, parse and play this plug-in label pair
The video file answered.
In present invention further optimization scheme, described plug-in label shows on corresponding video pictures with picture-in-picture form.
In present invention further optimization scheme, pre-established in described video clip have line of sight model;Described line of sight model is with 3
The form average mark of × 3 is furnished with 9 calibration points, and central campus is on schedule corresponding with the space coordinates of spectators pupil center, the 1st row the 1st
Calibration point on column position is corresponding with the space coordinates of spectators' inner eye corner;Based on the described line of sight model eye to collecting in step A2
In ball image, the pupil of spectators is tracked, the space coordinates current to extract pupil.
In present invention further optimization scheme, step A3 parses the video file that plug-in label is corresponding, at described video
Interface plays out in a form of picture-in-picture.
In present invention further optimization scheme, the video file that described plug-in label is corresponding is object in corresponding video pictures
Three-dimensional image video file.
In present invention further optimization scheme, step 3 parses the video file that plug-in label is corresponding, described video circle
Face directly redirects plays this video file.
In present invention further optimization scheme, this video interactive refers to that the Internet video on browser is mutual;Described video circle
Face refers to the interface of the video player of browser;Described video player forms based on HTML5 exploitation.
In present invention further optimization scheme, in step A1, plug-in label forms based on HTML5 exploitation, and its place mat is at video
On picture, and associate with corresponding video file.
In present invention further optimization scheme, described plug-in label is the JS file being embedded in the DIV that browser page is corresponding
Beneficial effects of the present invention at least includes:
1, solve the unidirectional output of conventional video, make conventional video become interactive video, with the addition of for video circle fresh
Element;
2, the eyeball following the trail of spectators is only needed can to realize alternately, the most convenient, it is greatly improved Consumer's Experience;
3, a greatly place mat has been made on the basis for VR content production, makes the development of VR content production more smooth and easy;
4, the most significant for advertisement video, can relate in this broadcasting tv screen while playing video display is right
The advertisement video of elephant.
Accompanying drawing explanation
Fig. 1 is the video interactive method principle schematic based on eyeball tracking technology that embodiment proposes.
Fig. 2 is the video interactive method schematic flow sheet based on eyeball tracking technology that embodiment proposes.
Fig. 3 is for determining the line of sight model schematic diagram of the drop point of pupil in Fig. 2 step S200.
Detailed description of the invention
For the ease of it will be appreciated by those skilled in the art that the present invention is described further below in conjunction with accompanying drawing and embodiment.
Central scope
Refer to Fig. 1, a kind of based on eyeball tracking technology the video interactive method that embodiment proposes, central scope is:
1, use HTML5 technology at the plug-in label of the upper corresponding video pictures place mat layer of transparent of ordinary video (original video)
(using transparent is to reduce interference during viewing original video) so that it is become interactive video.Can on one video pictures
There to be multiple plug-in label, for different choice;Make in advance simultaneously or prepare one section of video file associated with plug-in label with
Just transfer.This plug-in label refers to following VCE (Video Choice Engine) label example:
Above VCE label be a JavaScript storehouse i.e. JS file the least for showing that interactive video operates, use
JSON describes mutual text, it is provided that multiple branches and reentry point.This VCE label can be embedded in the DIV that browser page is corresponding,
Player is by by automatically identifying that the function of browser enables HTML5 or flash mode playback video.
2, on the basis of traditional player, a top layer video window is increased to realize inserting video or animation broadcasting.
Player mainly with HTML5 technology as platform development, simultaneously compatible flash technology, cross-platform each browser can be realized and hold concurrently
The video playback held, such as at Windows, MacOS, Linux, Windows Phone, the arbitrarily platform such as Android, IOS
On, play video by corresponding browser.
Player is by by automatically identifying that the function of browser enables HTML5 or flash mode playback video.
3, when spectators watch this interactive video, by photographic head, the eyeball of spectators can be tracked, to judge the interest of spectators
Tendency, when on a plug-in label, residence time exceedes default the pupil drop point of spectators (this time is rule of thumb arranged,
Such as 1 second or several seconds, it is not especially limited here), then can select this plug-in label, and by redirecting or with the side of picture-in-picture
The video file that this plug-in label is corresponding is played out by formula, to realize video interactive.
Embodiment flow process
Refer to Fig. 2, as a example by the present embodiment realizes video interactive on the interface of the video player passing through browser on computers
Being introduced, it mainly includes the following steps that S100 to S200:
S100, when being played to the video pictures being preset with plug-in label, by the eyeball image of photographic head Real-time Collection spectators.
Before step S100, the photographic head state on computer can be detected, it may be judged whether under the permission state of spectators, open
Dynamic photographic head;If photographic head does not start, also by manually implemented video interactive, if photographic head has been turned on, then can pass through eye
Ball tracer technique realizes video interactive.
Owing to any picture of the most whole usual video is required for interacting, therefore, plug-in label is only at a part of mesh
It is default in mark picture that (this plug-in label can show on corresponding video pictures with picture-in-picture form, outside not being played to be preset with
When hanging the video pictures of label, video clip will not show plug-in label).In step S100, photographic head can start to broadcast at video
Just the eyeball of spectators is tracked when putting, gathers its image;But in order to save resource occupation, also preferably preset being played to
Just it is acquired when having the video pictures of plug-in label.
S200, from the eyeball image collected, extract the space coordinates that pupil is current, and be mapped to the two dimension seat of video clip
Put on, to determine pupil drop point in video clip;The central point of the coordinate of described video clip, initial point respectively with spectators' pupil
The space coordinates at center, hole, the space coordinates correspondence of inner eye corner.
Referring to Fig. 3, pre-established in video clip can have line of sight model, this line of sight model is furnished with the form average mark of 3 × 3
9 calibration points, numbered 1-9;Central campus is on schedule 5 corresponding with the space coordinates of spectators pupil center O point, the 1st row the 1st
Calibration point 1 on column position is corresponding with the space coordinates of spectators' inner eye corner;Based on this line of sight model to collecting in step S200
In eyeball image, the pupil of spectators is tracked, the space coordinates current to extract pupil.Such as in Fig. 3, the pupil of spectators
Having been moved to N point from O point, corresponding drop point on line of sight model is calibration point 1, namely moves to calibration point 1 from calibration point 5,
Now pupil is β relative to inner eye corner move angle, and video clip lieutenant colonel 5 is L to the distance of calibration point 1 on schedule, deviation angle
Degree is α, so can get pupil drop point in video clip and coordinate.
S300, the drop point of pupil stop after reaching Preset Time on corresponding plug-in label, parse and play this plug-in label
Corresponding video file.
In the present embodiment, the content of the video file that plug-in label is corresponding can be the different play of the movie or television play
Feelings, the displaying of different pictures, or independent displaying of referent such as the articles for daily use etc. in video pictures.
Example 1: the first half story of a play or opera of film A is identical, but latter half has the version of two different story of a play or opera, the present embodiment can be
The last video pictures of first half presets two plug-in labels, each corresponding version;When film A plays last to first half
Video pictures time, video pictures there will be two transparent plug-in labels and select for spectators, each video tab can have letter
Single word makes a distinction and story of a play or opera prompting;Suggesting effect based on this word, the sight line of spectators will certainly fall interested to it
Label on, can determine that the drop point of spectators' pupil based on step S200 and S300, thus judge the interest tendency of spectators,
And jump directly to the picture of corresponding version, this video file is play.
Example 2: a certain section of video pictures in film B, it would appear that scenic spot at, can regard for the different azimuth of this scenic spot
Angle respectively arranges a plug-in label, the video file of the corresponding azimuthal viewing angle of each plug-in label;When film B plays to having
During the video pictures of this scenic spot, video pictures can show "front", "rear", four plug-in labels of "left", "right",
Can determine that the drop point of spectators' pupil based on step S200 and S300, thus judge the interest tendency of spectators, such as spectators
The drop point of pupil stop when reaching Preset Time on " left " this plug-in label, then switch to left view picture and play out,
Make user can watch the landscape on " left side ".
Example 3: a certain section of video pictures in film C it would appear that the cup of a uniqueness, can make in advance one about
The advertisement video of this cup, this advertisement video has the three-dimensional image of this cup to be shown, including outward appearance and internal structure etc.;Should
Advertisement video associates with plug-in label, and this plug-in label is preset on the video pictures occurring this cup in film C;Work as film
B plays to this video pictures, can determine that the drop point of spectators' pupil based on step S200 and S300, it is judged that spectators whether
Interested in this cup, however, it is determined that to be (drop point reaches Preset Time), then transfer this advertisement video and play out.Here it is
Do not interrupt the broadcasting of film C, preferably advertisement video is played out in a form of picture-in-picture.
Certainly, above example 1 to example 3 is only the most several interaction schemes, it is also possible to have other various interaction schemes, the most no longer
Repeat.
Embodiment described above only have expressed the several embodiments of the present invention, and it describes more concrete and in detail, but can not be because of
This and be interpreted as the restriction to the scope of the claims of the present invention.It should be pointed out that, for the person of ordinary skill of the art,
On the premise of present inventive concept, it is also possible to make some deformation and improvement, these broadly fall into protection scope of the present invention.
Therefore, the protection domain of patent of the present invention should be as the criterion with claims.
Claims (9)
1. a video interactive method based on eyeball tracking technology, it is characterised in that comprise the following steps:
A1, when being played to the video pictures being preset with plug-in label, by the eyeball image of photographic head Real-time Collection spectators;
A2, from the eyeball image collected, extract the space coordinates that pupil is current, and be mapped to the two-dimensional coordinate of video clip
On, to determine pupil drop point in video clip;The central point of the coordinate of described video clip, initial point respectively with spectators' pupil
The space coordinates at center, the space coordinates correspondence of inner eye corner;
A3, the drop point of pupil stop after reaching Preset Time on corresponding plug-in label, parse and play this plug-in label pair
The video file answered.
Video interactive method based on eyeball tracking technology the most according to claim 1, it is characterised in that described plug-in mark
Sign and show on corresponding video pictures with picture-in-picture form.
Video interactive method based on eyeball tracking technology the most according to claim 2, it is characterised in that described video circle
Pre-established on face have line of sight model;Described line of sight model is furnished with 9 calibration points with the form average mark of 3 × 3, central campus on schedule with
The space coordinates of spectators pupil center is corresponding, and the calibration point on the 1st row the 1st column position is corresponding with the space coordinates of spectators' inner eye corner;
Step A2 is tracked based on the pupil of spectators in the described line of sight model eyeball image to collecting, current to extract pupil
Space coordinates.
Video interactive method based on eyeball tracking technology the most according to claim 3, it is characterised in that in step A3
Parse the video file that plug-in label is corresponding, play out in a form of picture-in-picture in described video clip.
Video interactive method based on eyeball tracking technology the most according to claim 4, it is characterised in that described plug-in mark
Sign the corresponding three-dimensional image video file that video file is object in corresponding video pictures.
Video interactive method based on eyeball tracking technology the most according to claim 3, it is characterised in that solve in step 3
Separating out the video file that plug-in label is corresponding, described video clip directly redirects to be play this video file.
7. according to the video interactive method based on eyeball tracking technology described in any one of claim 1 to 6, it is characterised in that
This video interactive refers to that the Internet video on browser is mutual;Described video clip refers to the interface of the video player of browser;
Described video player forms based on HTML5 exploitation.
Video interactive method based on eyeball tracking technology the most according to claim 7, it is characterised in that in step A1
Plug-in label forms based on HTML5 exploitation, and its place mat is on video pictures, and associates with corresponding video file.
Video interactive method based on eyeball tracking technology the most according to claim 8, it is characterised in that described plug-in mark
Sign the JS file for being embedded in the DIV that browser page is corresponding.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610322512.9A CN106020461A (en) | 2016-05-13 | 2016-05-13 | Video interaction method based on eyeball tracking technology |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610322512.9A CN106020461A (en) | 2016-05-13 | 2016-05-13 | Video interaction method based on eyeball tracking technology |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106020461A true CN106020461A (en) | 2016-10-12 |
Family
ID=57098242
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610322512.9A Pending CN106020461A (en) | 2016-05-13 | 2016-05-13 | Video interaction method based on eyeball tracking technology |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106020461A (en) |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106651457A (en) * | 2016-12-29 | 2017-05-10 | 深圳艺特珑信息科技有限公司 | Method and system for realizing virtual advertising based on gyro and popularity analysis |
CN106779826A (en) * | 2016-12-05 | 2017-05-31 | 深圳艺特珑信息科技有限公司 | A kind of method and system that advertisement is optimized based on gyroscope and thermal map analysis |
CN106774862A (en) * | 2016-12-03 | 2017-05-31 | 西安科锐盛创新科技有限公司 | VR display methods and VR equipment based on sight line |
CN106990839A (en) * | 2017-03-21 | 2017-07-28 | 张文庆 | A kind of eyeball identification multimedia player and its implementation |
CN107396086A (en) * | 2017-07-28 | 2017-11-24 | 歌尔科技有限公司 | The method and VR helmets of video are played based on VR helmets |
CN107973198A (en) * | 2017-11-30 | 2018-05-01 | 成都科力夫科技有限公司 | The elevator card jettison system of Auto-matching occupant |
CN108235132A (en) * | 2018-03-13 | 2018-06-29 | 哈尔滨市舍科技有限公司 | Panoramic video visual angle regulating method and device based on human eye positioning |
CN109410285A (en) * | 2018-11-06 | 2019-03-01 | 北京七鑫易维信息技术有限公司 | A kind of calibration method, device, terminal device and storage medium |
CN110324694A (en) * | 2019-06-11 | 2019-10-11 | 福建天泉教育科技有限公司 | Video broadcasting method, storage medium |
WO2020042541A1 (en) * | 2018-08-31 | 2020-03-05 | 深圳市沃特沃德股份有限公司 | Eyeball tracking interactive method and device |
CN110956483A (en) * | 2018-09-26 | 2020-04-03 | 上海博泰悦臻网络技术服务有限公司 | Advertisement information pushing method, system, server and vehicle |
CN111198611A (en) * | 2018-11-19 | 2020-05-26 | 中兴通讯股份有限公司 | Method for determining sight line landing point, terminal and computer readable storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102833490A (en) * | 2011-06-15 | 2012-12-19 | 新诺亚舟科技(深圳)有限公司 | Method and system for editing and playing interactive video, and electronic learning device |
CN103108248A (en) * | 2013-01-06 | 2013-05-15 | 王汝迟 | Interactive video implement method and system using the same |
CN103176607A (en) * | 2013-04-16 | 2013-06-26 | 重庆市科学技术研究院 | Eye-controlled mouse realization method and system |
CN103914147A (en) * | 2014-03-29 | 2014-07-09 | 朱定局 | Eye-controlled video interaction method and eye-controlled video interaction system |
CN105425967A (en) * | 2015-12-16 | 2016-03-23 | 中国科学院西安光学精密机械研究所 | Sight tracking and human eye area-of-interest positioning system |
-
2016
- 2016-05-13 CN CN201610322512.9A patent/CN106020461A/en active Pending
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102833490A (en) * | 2011-06-15 | 2012-12-19 | 新诺亚舟科技(深圳)有限公司 | Method and system for editing and playing interactive video, and electronic learning device |
CN103108248A (en) * | 2013-01-06 | 2013-05-15 | 王汝迟 | Interactive video implement method and system using the same |
CN103176607A (en) * | 2013-04-16 | 2013-06-26 | 重庆市科学技术研究院 | Eye-controlled mouse realization method and system |
CN103914147A (en) * | 2014-03-29 | 2014-07-09 | 朱定局 | Eye-controlled video interaction method and eye-controlled video interaction system |
CN105425967A (en) * | 2015-12-16 | 2016-03-23 | 中国科学院西安光学精密机械研究所 | Sight tracking and human eye area-of-interest positioning system |
Non-Patent Citations (2)
Title |
---|
迟健男: "《视线追踪》", 30 June 2011 * |
陆虎敏: "《飞机座舱显示与控制技术》", 31 December 2015 * |
Cited By (17)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106774862A (en) * | 2016-12-03 | 2017-05-31 | 西安科锐盛创新科技有限公司 | VR display methods and VR equipment based on sight line |
CN106779826A (en) * | 2016-12-05 | 2017-05-31 | 深圳艺特珑信息科技有限公司 | A kind of method and system that advertisement is optimized based on gyroscope and thermal map analysis |
CN106779826B (en) * | 2016-12-05 | 2020-09-01 | 深圳艺特珑信息科技有限公司 | Method and system for optimizing advertisement based on gyroscope and heat map analysis |
WO2018103070A1 (en) * | 2016-12-05 | 2018-06-14 | 深圳艺特珑信息科技有限公司 | Gyroscope and heat map analysis-based advertisement optimization method and system |
CN106651457B (en) * | 2016-12-29 | 2019-04-02 | 深圳艺特珑信息科技有限公司 | The method and system that virtual ads are launched is realized based on gyroscope and temperature analysis |
CN106651457A (en) * | 2016-12-29 | 2017-05-10 | 深圳艺特珑信息科技有限公司 | Method and system for realizing virtual advertising based on gyro and popularity analysis |
WO2018120005A1 (en) * | 2016-12-29 | 2018-07-05 | 深圳艺特珑信息科技有限公司 | Method and system for implementing virtual advertisement placement on basis of gyroscope and popularity analysis |
CN106990839A (en) * | 2017-03-21 | 2017-07-28 | 张文庆 | A kind of eyeball identification multimedia player and its implementation |
CN106990839B (en) * | 2017-03-21 | 2020-06-05 | 张文庆 | Eyeball identification multimedia player and implementation method thereof |
CN107396086A (en) * | 2017-07-28 | 2017-11-24 | 歌尔科技有限公司 | The method and VR helmets of video are played based on VR helmets |
CN107973198A (en) * | 2017-11-30 | 2018-05-01 | 成都科力夫科技有限公司 | The elevator card jettison system of Auto-matching occupant |
CN108235132A (en) * | 2018-03-13 | 2018-06-29 | 哈尔滨市舍科技有限公司 | Panoramic video visual angle regulating method and device based on human eye positioning |
WO2020042541A1 (en) * | 2018-08-31 | 2020-03-05 | 深圳市沃特沃德股份有限公司 | Eyeball tracking interactive method and device |
CN110956483A (en) * | 2018-09-26 | 2020-04-03 | 上海博泰悦臻网络技术服务有限公司 | Advertisement information pushing method, system, server and vehicle |
CN109410285A (en) * | 2018-11-06 | 2019-03-01 | 北京七鑫易维信息技术有限公司 | A kind of calibration method, device, terminal device and storage medium |
CN111198611A (en) * | 2018-11-19 | 2020-05-26 | 中兴通讯股份有限公司 | Method for determining sight line landing point, terminal and computer readable storage medium |
CN110324694A (en) * | 2019-06-11 | 2019-10-11 | 福建天泉教育科技有限公司 | Video broadcasting method, storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106020461A (en) | Video interaction method based on eyeball tracking technology | |
CN102290082B (en) | Method and device for processing brilliant video replay clip | |
CN111133365B (en) | Matching content to spatial 3D environment | |
US9851793B1 (en) | Virtual reality system including social graph | |
WO2016095384A1 (en) | Live comment display method and system | |
CN110463195B (en) | Method and apparatus for rendering timed text and graphics in virtual reality video | |
WO2022022036A1 (en) | Display method, apparatus and device, storage medium, and computer program | |
CN105916043A (en) | Barrage control method and device | |
CN105519097A (en) | Systems, devices and methods for displaying pictures in a picture | |
US20130174037A1 (en) | Method and device for adding video information, and method and device for displaying video information | |
CN108021896B (en) | Image pickup method, device, equipment and computer-readable medium based on augmented reality | |
CN106412711B (en) | Barrage control method and device | |
CN102263907A (en) | Play control method of competition video, and generation method and device for clip information of competition video | |
EP3679718A1 (en) | Personalized live media content | |
WO2022213727A1 (en) | Live broadcast interaction method and apparatus, and electronic device and storage medium | |
CN113591524A (en) | Display device and interface display method | |
CN106028078A (en) | Personalized content creating method, personalized content creating device, personalized content play method and personalized content play device | |
CN107635153B (en) | Interaction method and system based on image data | |
TR201901701T4 (en) | THE METHOD, SYSTEM AND COMPUTER PROGRAM FOR THE TRANSFORMATION OF AN IMAGE. | |
CN110458820A (en) | A kind of multimedia messages method for implantation, device, equipment and storage medium | |
CN103384311A (en) | Method for generating interactive videos in batch mode automatically | |
EP2896210A2 (en) | Media content distribution | |
CN108600858B (en) | Video playing method for synchronously displaying AR information | |
CN102074033A (en) | Method and device for animation production | |
ES2434259T3 (en) | Social television service |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20161012 |
|
RJ01 | Rejection of invention patent application after publication |