CN110944222B - Method and system for immersive media content as user moves - Google Patents
Method and system for immersive media content as user moves Download PDFInfo
- Publication number
- CN110944222B CN110944222B CN201811108139.2A CN201811108139A CN110944222B CN 110944222 B CN110944222 B CN 110944222B CN 201811108139 A CN201811108139 A CN 201811108139A CN 110944222 B CN110944222 B CN 110944222B
- Authority
- CN
- China
- Prior art keywords
- viewpoint
- user
- viewpoints
- determining
- content
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/40—Client devices specifically adapted for the reception of or interaction with content, e.g. set-top-box [STB]; Operations thereof
- H04N21/41—Structure of client; Structure of client peripherals
- H04N21/4104—Peripherals receiving signals from specially adapted client devices
- H04N21/4122—Peripherals receiving signals from specially adapted client devices additional display device, e.g. video projector
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04N—PICTORIAL COMMUNICATION, e.g. TELEVISION
- H04N21/00—Selective content distribution, e.g. interactive television or video on demand [VOD]
- H04N21/60—Network structure or processes for video distribution between server and client or between remote clients; Control signalling between clients, server and network components; Transmission of management data between server and client, e.g. sending from server to client commands for recording incoming content stream; Communication details between server and client
- H04N21/65—Transmission of management data between client and server
- H04N21/658—Transmission by the client directed to the server
- H04N21/6587—Control parameters, e.g. trick play commands, viewpoint selection
Abstract
The invention provides a method and a system for immersing media content along with the movement change of a user, comprising the following steps: reading and analyzing a video stream; determining the number of viewpoints, the initial viewpoints, the relative relationship among the viewpoints and the maximum coverage radius value of each viewpoint; feeding back the relative displacement of the user to the initial viewpoint according to the position of the user; selecting a current viewpoint according to the relative relationship among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement; determining a current depth value according to the relative displacement, and determining the view content of the current viewpoint according to the relation between the view and the depth in the viewpoint; content in the video stream corresponding to the viewing field of view content is presented. The invention supports that the watching area of the user changes in a certain relation with the current viewpoint relative to the initial viewpoint, and simultaneously supports the switching of different viewpoints.
Description
Technical Field
The invention relates to the technical field of multimedia, in particular to a method and a system for immersing media content along with the movement change of a user.
Background
With the rapid development of Virtual Reality (VR) technology, the demand of VR systems increases, and the development from three degrees of Freedom (3Degree of Freedom, 3DoF), three degrees of Freedom plus (3Degree of Freedom +, 3DoF +) to six degrees of Freedom (6 Degree of Freedom, 6DoF) is being realized, where 3DoF supports the head of a user to perform three rotations of Yaw, Roll, and Pitch (i.e., Yaw angle, Roll angle, and Pitch angle), 3DoF + supports the head of the user to perform small-range translation motions of X, Y, Z axes in up, down, left, right, front, and back directions on the basis of 3DoF, and 6DoF can not only perform three rotations of Yaw, Roll, and Pitch as 3DoF, but also can track the translation of the user on X, Y, Z axis. Technically, the VR video should add depth information on the basis of the existing three degrees of Freedom (3Degree of Freedom, 3DoF), and support that the scene changes correspondingly with the movement of the user, that is, when the user wears the VR head display to move in any direction, the VR will react correspondingly, so that the user obtains better immersive experience.
Now, detection and identification of immersive media content are important indicators for determining the immersive experience of the user, and in order to meet the requirements of different application scenarios, additional identification of feedback and package information of immersive media content is required to meet the further application of specific information.
Disclosure of Invention
In view of the deficiencies in the prior art, it is an object of the present invention to provide a method and system for immersive media content as a function of user movement.
According to the invention, a method for immersive media content as a function of user movement is provided, comprising:
and (3) analyzing: reading and analyzing a video stream;
a viewpoint determining step: determining the number of viewpoints, the initial viewpoints, the relative relationship among the viewpoints and the maximum coverage radius value of each viewpoint;
and a displacement feedback step: feeding back the relative displacement of the user to the initial viewpoint according to the position of the user;
a viewpoint selection step: selecting a current viewpoint according to the relative relationship among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement;
a viewing field content determination step: determining a current depth value according to the relative displacement, and determining the view content of the current viewpoint according to the relation between the view and the depth in the viewpoint;
a video presenting step: content in the video stream corresponding to the viewing field of view content is presented.
Preferably, the relative relationship between the viewpoints includes: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
Preferably, the user relative displacement includes: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
Preferably, the depth value is a distance of the user's position from the current viewpoint.
According to the invention, a system for immersive media content as a function of user movement is provided, comprising:
an analysis module: reading and analyzing a video stream;
a viewpoint determining module: determining the number of viewpoints, the initial viewpoints, the relative relationship among the viewpoints and the maximum coverage radius value of each viewpoint;
a displacement feedback module: feeding back the relative displacement of the user to the initial viewpoint according to the position of the user;
a viewpoint selection module: selecting a current viewpoint according to the relative relationship among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement;
a viewing field content determination module: determining a current depth value according to the relative displacement, and determining the view content of the current viewpoint according to the relation between the view and the depth in the viewpoint;
a video presentation module: content in the video stream corresponding to the viewing field of view content is presented.
Preferably, the relative relationship between the viewpoints includes: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
Preferably, the user relative displacement includes: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
Preferably, the depth value is a distance of the user's position from the current viewpoint.
Compared with the prior art, the invention has the following beneficial effects:
for immersive media content supporting motion parallax, the method and the system support that the viewing area of the user changes in a certain relation with the current viewpoint relative to the initial viewpoint, and simultaneously support switching of different viewpoints as if the user is completely immersed in a scene, so that the degree of freedom in consumption of the immersive media is fully improved according to the interaction behavior of the user.
Drawings
Other features, objects and advantages of the invention will become more apparent upon reading of the detailed description of non-limiting embodiments with reference to the following drawings:
FIG. 1 is a logic flow diagram of the present invention;
FIG. 2 is a diagram illustrating an exemplary embodiment of the present invention relating to the relationship between the viewing area and the movement information;
fig. 3 is a view point range illustration of a specific application example of the present invention with respect to an immersive virtual museum.
Detailed Description
The present invention will be described in detail with reference to specific examples. The following examples will assist those skilled in the art in further understanding the invention, but are not intended to limit the invention in any way. It should be noted that it would be obvious to those skilled in the art that various changes and modifications can be made without departing from the spirit of the invention. All falling within the scope of the present invention.
As shown in fig. 1, a method of immersive media content as a function of user movement is provided in accordance with the present invention, comprising:
and (3) analyzing: and the server side reads and analyzes the video stream.
A viewpoint determining step: and determining the number of viewpoints, the initial viewpoints, the relative relation among the viewpoints and the maximum coverage radius value of each viewpoint. The relative relationship between viewpoints includes: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
And a displacement feedback step: and feeding back the relative displacement of the user to the initial viewpoint according to the position of the user. The user relative displacement comprises: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
A viewpoint selection step: and selecting the current viewpoint according to the relative relation among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement.
A viewing field content determination step: and determining the current depth value according to the relative displacement, and determining the watching view content of the current view point according to the relation between the view and the depth in the view point. The depth value is the distance of the user's position from the current viewpoint.
A video presenting step: content in the video stream corresponding to the viewing field of view content is presented.
Based on the method for immersing the media content to change along with the movement of the user, the invention also provides a method for immersing the media content to change along with the movement of the user, which comprises the following steps:
and (3) analyzing: and the server side reads and analyzes the video stream.
A viewpoint determining step: and determining the number of viewpoints, the initial viewpoints, the relative relation among the viewpoints and the maximum coverage radius value of each viewpoint. The relative relationship between viewpoints includes: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
And a displacement feedback step: and feeding back the relative displacement of the user to the initial viewpoint according to the position of the user. The user relative displacement comprises: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
A viewpoint selection step: and selecting the current viewpoint according to the relative relation among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement.
A viewing field content determination step: and determining the current depth value according to the relative displacement, and determining the watching view content of the current view point according to the relation between the view and the depth in the view point. The depth value is the distance of the user's position from the current viewpoint.
A video presenting step: content in the video stream corresponding to the viewing field of view content is presented.
The invention aims to provide an identification method for immersive media content along with corresponding changes of user interaction behaviors (such as head movement and body movement), which can indicate the interaction behaviors of a user in consuming the immersive media content and feed back the interaction behaviors to a server side so as to acquire the immersive media content meeting the requirements of the user and different application scenes.
In this embodiment, an immersive virtual museum is taken as an example. Immersive virtual museum, that is, the user can wear three-dimensional head-mounted device, can interact with surrounding environment and nearby object in the exhibition district, and the exhibition area that sees can become certain relation change with the position that the user removed, supports the viewpoint switch between different exhibition districts simultaneously. In immersive media, the position at which a user views media content is called the viewpoint, i.e., the position of the camera. The immersive virtual museum allows a user to switch among a plurality of panoramic videos about an exhibition area based on position information of different viewpoints, and simultaneously supports motion parallax, namely, viewing content in the current viewpoint is allowed to change along with user interaction behaviors, for example, the user can observe a more detailed part in a scene through behaviors close to exhibits, and a better immersive experience is obtained. Assuming that different exhibition area scenes are shot by independent and fixed panoramic cameras, a user can freely change a moving route so as to realize switching between different viewpoints and immerse the scenes in the corresponding exhibition areas; meanwhile, the head position can be changed to feel the motion parallax, so that the effect of watching objects close to and far away from the current viewpoint coverage range is realized, namely, the adaptive adjustment of the interactive behavior (recorded as real-time relative displacement) of the actual watching area of the user along with the viewpoint range is supported.
In order to achieve the above purpose, the following technical solutions are adopted in this embodiment:
for immersive media content that supports motion parallax, the user viewing region is enabled to change in a relationship to the current viewpoint relative to the initial viewpoint, while different viewpoint switching is enabled, as if fully immersed in the scene. And attaches necessary indication information (e.g., package information, transport information, consumption information) uniquely associated therewith to substantially increase the freedom in consumption of the immersive media in accordance with user interaction behavior.
In the invention, the necessary indication information of the immersive media content to be added can be realized by taking the following information as an example:
information one: for indicating the total number of viewpoints within the complete scene;
and information II: for indicating an initial viewpoint;
and (3) information three: for indicating relative position information between different viewpoints to distinguish viewpoint contents;
and information four: on the basis of the information three, indicating constraint information for switching among different viewpoints;
and information five: the coverage range information is used for indicating different viewpoint contents so as to adjust the actual watching area of the user according to the real-time relative displacement of the user;
information six: and on the basis of the fifth information, adding the corresponding relation between the actual watching area of the user and the real-time relative displacement.
And information seven: and on the basis of the information five, indicating the displacement information of the user relative to the current viewpoint.
And carrying out information identification on the variation of the immersive media content along with the movement, wherein the identification information indicates the movement information of the user, the packaging information of the relation between the area range watched by the user and the movement information and indicates the current viewpoint information of the user.
For immersive media content that supports motion parallax, the user viewing region is enabled to change in a relationship to the current viewpoint relative to the initial viewpoint, while different viewpoint switching is enabled, as if fully immersed in the scene. And necessary indication information uniquely associated with the video media is added to fully feed back the interactive behavior of the immersive media content, so that the specific application requirements of the video media are further realized.
For the above problem, as shown in fig. 2, the following fields can be added as needed:
hmovement _ x: x-coordinate information indicating viewer movement;
hmovement _ y: y-coordinate information indicating viewer movement;
hmovement _ z: z-coordinate information indicating viewer movement;
parallelx _ flag: indicating a video supporting motion parallax;
viewpoint _ ID: indicating the current viewpoint information of the user;
move _ depth: defining a depth value of the viewing area, specifying the relative distance of the real-time position of the user relative to the initial viewpoint, and obtaining the depth value according to the feedback coordinates hmovement _ x, hmovement _ y and hmovement _ z;
behavior _ coefficient: defining as a magnification behavior coefficient;
sphere _ radius: a radius representing a spherical viewing area;
viewing _ range _ field: the method comprises the steps of representing an area range which can be watched by a user at a real-time position, and determining according to a watching depth, a behavior coefficient and a spherical area radius;
num _ viewpoint: a preset total number of viewpoints;
viewpoint _ x (0), viewpoint _ y (0), viewpoint _ z (0): x, y, z coordinate information indicating an initial viewpoint position, set to (0, 0, 0);
viewpoint _ x (i): indicating x coordinate information of the position of the viewpoint i relative to the initial viewpoint;
viewpoint _ y (i): y coordinate information indicating the position of the viewpoint i relative to the initial viewpoint;
viewpoint _ z (i): z coordinate information indicating a position of the viewpoint i with respect to the initial viewpoint;
rmax (i): indicating that the viewpoint i covers the maximum radius information.
Based on the above information, taking vrviewpoint changefeed and Depth and viewing range information box as examples, an organization structure of these information is given below.
1.VRViewpointChangeFeedback
Depth and viewing range information box
2.1 definition
■ data Box Type (Box Type): dvri'
Contained in (Container) project Omni VideoBox
Mandatory (Mandatory) not Mandatory
Number (Quantity) zero or one
The data box provides a relative relationship between user viewing range and viewpoint depth in a spherical area.
2.2 grammar
View point Change Struct view point Change Structure syntax
For a better understanding of the meaning of the above fields, reference is made to the application examples mentioned below.
Based on the above expression, specific application examples are given below:
in the application of panoramic video, due to the technical limitation, when a user watches the video, the user is limited to the included 360-degree visual angle range, and the watching content of the user cannot change along with the step movement of the user. For the panoramic video with the depth information, a user can freely change the action route to realize the switching between different viewpoints and immerse the panoramic video in the corresponding exhibition area scene; meanwhile, the head position can be changed to feel the motion parallax, so that the effect of watching objects close to and far away from the current viewpoint coverage range is realized, namely, the adjustment of the adaptability of the actual watching area of the user along with the real-time relative displacement is supported.
Specifically, as shown in fig. 3, when the user moves within the view coverage of the exhibition area 1 in the museum, the client directly feeds back the user's real-time relative displacement to the server by locating the corresponding information of hmovement _ x, hmovement _ y, and hmovement _ z, the server may obtain the depth value of the viewing area according to the feedback coordinates hmovement _ x, hmovement _ y, and hmovement _ z, determine the area range that the user can view at the current viewpoint according to the analyzed viewing depth, behavior coefficient, and spherical area radius, and then present the corresponding viewing area at the viewpoint 1 to the user, generally speaking, the user's movement within the current viewpoint coverage can achieve the effect of "approaching" and "departing" the viewing object. When the user moves from the exhibition area 1 to the exhibition area 2, the real-time position is within the maximum radius coverage of the exhibition area 2, accordingly, the viewing content is obtained from the scene provided by the viewpoint 2, and when the user moves freely within the coverage of the exhibition area 2, the viewing area changes as described above.
Those skilled in the art will appreciate that, in addition to implementing the system and its various devices, modules, units provided by the present invention as pure computer readable program code, the system and its various devices, modules, units provided by the present invention can be fully implemented by logically programming method steps in the form of logic gates, switches, application specific integrated circuits, programmable logic controllers, embedded microcontrollers and the like. Therefore, the system and various devices, modules and units thereof provided by the invention can be regarded as a hardware component, and the devices, modules and units included in the system for realizing various functions can also be regarded as structures in the hardware component; means, modules, units for performing the various functions may also be regarded as structures within both software modules and hardware components for performing the method.
The foregoing description of specific embodiments of the present invention has been presented. It is to be understood that the present invention is not limited to the specific embodiments described above, and that various changes or modifications may be made by one skilled in the art within the scope of the appended claims without departing from the spirit of the invention. The embodiments and features of the embodiments of the present application may be combined with each other arbitrarily without conflict.
Claims (6)
1. A method of immersive media content as a function of user movement, comprising:
and (3) analyzing: reading and analyzing a video stream;
a viewpoint determining step: determining the initial viewpoints, the relative relationship among the viewpoints and the maximum coverage radius value of each viewpoint;
and a displacement feedback step: feeding back the relative displacement of the user to the initial viewpoint according to the position of the user;
a viewpoint selection step: selecting a current viewpoint according to the relative relationship among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement;
a viewing field content determination step: determining a current depth value according to the relative displacement, and determining the view content of the current viewpoint according to the relation between the view and the depth in the viewpoint;
a video presenting step: presenting content in the video stream corresponding to the viewing horizon content;
the depth value is the distance of the user's position from the current viewpoint.
2. The method of immersive media content of claim 1 as a function of user movement, wherein said relative relationship between viewpoints comprises: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
3. The method of immersive media content of claim 1 as a function of user movement, wherein said user relative displacement comprises: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
4. A system for immersive media content as a function of user movement, comprising:
an analysis module: reading and analyzing a video stream;
a viewpoint determining module: determining the initial viewpoints, the relative relationship among the viewpoints and the maximum coverage radius value of each viewpoint;
a displacement feedback module: feeding back the relative displacement of the user to the initial viewpoint according to the position of the user;
a viewpoint selection module: selecting a current viewpoint according to the relative relationship among the viewpoints, the maximum coverage radius value of each viewpoint and the relative displacement;
a viewing field content determination module: determining a current depth value according to the relative displacement, and determining the view content of the current viewpoint according to the relation between the view and the depth in the viewpoint;
a video presentation module: presenting content in the video stream corresponding to the viewing horizon content;
the depth value is the distance of the user's position from the current viewpoint.
5. The system of immersive media content as recited in claim 4, wherein the relative relationship between the viewpoints comprises: and establishing a coordinate system by taking the initial viewpoint as a coordinate origin, and determining coordinates of viewpoints other than the initial viewpoint.
6. The system of immersive media content as recited in claim 4, wherein the user relative displacement comprises: x-coordinate information, y-coordinate information, z-coordinate information of the user's movement.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811108139.2A CN110944222B (en) | 2018-09-21 | 2018-09-21 | Method and system for immersive media content as user moves |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201811108139.2A CN110944222B (en) | 2018-09-21 | 2018-09-21 | Method and system for immersive media content as user moves |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110944222A CN110944222A (en) | 2020-03-31 |
CN110944222B true CN110944222B (en) | 2021-02-12 |
Family
ID=69904623
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201811108139.2A Active CN110944222B (en) | 2018-09-21 | 2018-09-21 | Method and system for immersive media content as user moves |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110944222B (en) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112150603B (en) * | 2019-06-28 | 2023-03-28 | 上海交通大学 | Initial visual angle control and presentation method and system based on three-dimensional point cloud |
CN113542907B (en) * | 2020-04-16 | 2022-09-23 | 上海交通大学 | Multimedia data transceiving method, system, processor and player |
CN112492289A (en) * | 2020-06-23 | 2021-03-12 | 中兴通讯股份有限公司 | Immersion media data processing method and device, storage medium and electronic device |
CN114374675B (en) * | 2020-10-14 | 2023-02-28 | 腾讯科技(深圳)有限公司 | Media file encapsulation method, media file decapsulation method and related equipment |
Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104685858A (en) * | 2012-09-28 | 2015-06-03 | 阿尔卡特朗讯 | Immersive videoconference method and system |
CN105763790A (en) * | 2014-11-26 | 2016-07-13 | 鹦鹉股份有限公司 | Video System For Piloting Drone In Immersive Mode |
CN105959665A (en) * | 2016-05-05 | 2016-09-21 | 清华大学深圳研究生院 | Panoramic 3D video generation method for virtual reality equipment |
CN105988578A (en) * | 2015-03-04 | 2016-10-05 | 华为技术有限公司 | Interactive video display method, device and system |
CN106291930A (en) * | 2015-06-24 | 2017-01-04 | 联发科技股份有限公司 | Head mounted display |
CN106502501A (en) * | 2016-10-31 | 2017-03-15 | 宁波视睿迪光电有限公司 | Index localization method and device |
CN106576158A (en) * | 2014-08-13 | 2017-04-19 | 瑞典爱立信有限公司 | Immersive video |
CN107197318A (en) * | 2017-06-19 | 2017-09-22 | 深圳市望尘科技有限公司 | A kind of real-time, freedom viewpoint live broadcasting method shot based on multi-cam light field |
CN107396085A (en) * | 2017-08-24 | 2017-11-24 | 三星电子(中国)研发中心 | A kind of processing method and system of full multi-view video image |
CN108235113A (en) * | 2016-12-14 | 2018-06-29 | 上海交通大学 | A kind of panoramic video renders and presents attribute indicating means and system |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6573912B1 (en) * | 2000-11-07 | 2003-06-03 | Zaxel Systems, Inc. | Internet system for virtual telepresence |
US9756316B2 (en) * | 2013-11-04 | 2017-09-05 | Massachusetts Institute Of Technology | Joint view expansion and filtering for automultiscopic 3D displays |
US11184599B2 (en) * | 2017-03-15 | 2021-11-23 | Pcms Holdings, Inc. | Enabling motion parallax with multilayer 360-degree video |
-
2018
- 2018-09-21 CN CN201811108139.2A patent/CN110944222B/en active Active
Patent Citations (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104685858A (en) * | 2012-09-28 | 2015-06-03 | 阿尔卡特朗讯 | Immersive videoconference method and system |
CN106576158A (en) * | 2014-08-13 | 2017-04-19 | 瑞典爱立信有限公司 | Immersive video |
CN105763790A (en) * | 2014-11-26 | 2016-07-13 | 鹦鹉股份有限公司 | Video System For Piloting Drone In Immersive Mode |
CN105988578A (en) * | 2015-03-04 | 2016-10-05 | 华为技术有限公司 | Interactive video display method, device and system |
CN106291930A (en) * | 2015-06-24 | 2017-01-04 | 联发科技股份有限公司 | Head mounted display |
CN105959665A (en) * | 2016-05-05 | 2016-09-21 | 清华大学深圳研究生院 | Panoramic 3D video generation method for virtual reality equipment |
CN106502501A (en) * | 2016-10-31 | 2017-03-15 | 宁波视睿迪光电有限公司 | Index localization method and device |
CN108235113A (en) * | 2016-12-14 | 2018-06-29 | 上海交通大学 | A kind of panoramic video renders and presents attribute indicating means and system |
CN107197318A (en) * | 2017-06-19 | 2017-09-22 | 深圳市望尘科技有限公司 | A kind of real-time, freedom viewpoint live broadcasting method shot based on multi-cam light field |
CN107396085A (en) * | 2017-08-24 | 2017-11-24 | 三星电子(中国)研发中心 | A kind of processing method and system of full multi-view video image |
Also Published As
Publication number | Publication date |
---|---|
CN110944222A (en) | 2020-03-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110944222B (en) | Method and system for immersive media content as user moves | |
US11037601B2 (en) | Spherical video editing | |
KR102138536B1 (en) | Method for processing overlay in 360-degree video system | |
US11257233B2 (en) | Volumetric depth video recording and playback | |
CN111837383A (en) | Method and apparatus for transmitting and receiving metadata about coordinate system of dynamic viewpoint | |
US20180329482A1 (en) | Method for providing content and apparatus therefor | |
KR20200038170A (en) | Method for processing overlay in 360-degree video system and apparatus for the same | |
EP3422149B1 (en) | Methods, apparatus, systems, computer programs for enabling consumption of virtual content for mediated reality | |
CN112514398A (en) | Method and apparatus for marking user interactions on overlays for omnidirectional content and grouping overlays to a background | |
US10768711B2 (en) | Mediated reality | |
CN112150603B (en) | Initial visual angle control and presentation method and system based on three-dimensional point cloud | |
CN113709093B (en) | Packaging method, device and medium of three-dimensional point cloud | |
US11587200B2 (en) | Method and apparatus for enabling multiple timeline support for omnidirectional content playback | |
US11438731B2 (en) | Method and apparatus for incorporating location awareness in media content | |
US20190295324A1 (en) | Optimized content sharing interaction using a mixed reality environment | |
CN114830676A (en) | Video processing device and manifest file for video streaming | |
CN113178019B (en) | Indication information identification method, system and storage medium based on video content | |
US10692286B2 (en) | Device and method of creating an augmented interactive virtual reality system | |
US20200160600A1 (en) | Methods, Apparatus, Systems, Computer Programs for Enabling Consumption of Virtual Content for Mediated Reality | |
CN113453083B (en) | Immersion type media obtaining method and device under multi-degree-of-freedom scene and storage medium | |
US20200336720A1 (en) | Device and method of creating an augmented interactive virtual reality system | |
US11875472B1 (en) | Method of rendering dynamic labels in an extended reality environment | |
CN110704673A (en) | Feedback information identification method, system and storage medium based on video content consumption | |
CN110574369B (en) | Method and apparatus for providing contents | |
CN117218319A (en) | Augmented reality processing method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |