CN108257219B - Method for realizing panoramic multipoint roaming - Google Patents

Method for realizing panoramic multipoint roaming Download PDF

Info

Publication number
CN108257219B
CN108257219B CN201810096799.7A CN201810096799A CN108257219B CN 108257219 B CN108257219 B CN 108257219B CN 201810096799 A CN201810096799 A CN 201810096799A CN 108257219 B CN108257219 B CN 108257219B
Authority
CN
China
Prior art keywords
roaming
scene
panoramic
camera
current
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810096799.7A
Other languages
Chinese (zh)
Other versions
CN108257219A (en
Inventor
卢健智
龚德明
何为
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong 3vjia Information Technology Co Ltd
Original Assignee
Guangdong 3vjia Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong 3vjia Information Technology Co Ltd filed Critical Guangdong 3vjia Information Technology Co Ltd
Priority to CN201810096799.7A priority Critical patent/CN108257219B/en
Publication of CN108257219A publication Critical patent/CN108257219A/en
Application granted granted Critical
Publication of CN108257219B publication Critical patent/CN108257219B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T15/003D [Three Dimensional] image rendering
    • G06T15/005General purpose rendering architectures

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Graphics (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Processing Or Creating Images (AREA)
  • Studio Devices (AREA)
  • Two-Way Televisions, Distribution Of Moving Picture Or The Like (AREA)

Abstract

The invention discloses a method for realizing panoramic multipoint roaming, which comprises the following steps: s1: drawing a roaming path, S2: rendering a panorama, S3: acquisition coordinates, S4: generating a panoramic file, S5: acquiring the position of the spatial point, S6: the invention selects the cubic panorama as a research target, is beneficial to image storage and compression due to simple structure of the cubic panorama, is more suitable for multi-space free switching, is beneficial to virtual roaming of freely-designated roaming routes between the panoramas, realizes seamless multi-scene immersive experience, converts switching of a plurality of scenes into visual drawing operation in the expression of an actual three-dimensional roaming scene, and adopts different user-defined roaming route strategies aiming at different roaming scenes to realize high-quality virtual panorama roaming drawing.

Description

Method for realizing panoramic multipoint roaming
Technical Field
The invention relates to the technical field of tile multi-brick paving, in particular to a method for realizing panoramic multipoint roaming.
Background
The virtual reality technology is a new technology developed in the last three decades, integrates and synthesizes information technologies such as machine vision, computer graphic drawing, sensors, three-dimensional simulation, artificial intelligence and the like, constructs a virtual scene, and supports user interaction, thereby achieving the purpose that a user personally observes and even controls scene information. The virtual reality system needs to satisfy three characteristics: conception, immersion, interaction. Due to limitations in computer performance (computing power, storage power), computer graphics drawing three-dimensional scenes is often rough and does not provide sufficient immersion and sufficient interactivity. At present, the image drawing technology is developed rapidly, and an implementation means which can be borne by the current computer level is provided for scene vivid construction. A close-range small-area scene virtual roaming system constructed by the technology appears in the market, such as a virtual museum, a virtual landscape and the like. Currently, a roaming system for a three-dimensional virtual scene mostly adopts a mode similar to single scene switching, and a user can only jump to a certain scene by clicking a scene thumbnail. In the virtual roaming system established by combining fixed-point collected shot pictures into a panoramic image and then fusing a specific algorithm, the seamless virtual roaming of a user-defined route cannot be met by single jump type browsing. Namely, the user-defined multipoint roaming is realized in the virtual scene roaming. The panoramic multipoint roaming method is invented on the basis of the existing scene roaming technology by combining the actual situation.
Disclosure of Invention
The invention aims to provide a method for realizing panoramic multipoint roaming, which comprises the following specific steps:
s1: drawing a roaming path, drawing virtual position points roaming in each space on a planar design drawing of the panoramic scheme, wherein the points represent positions of the camera in the space, one space can have a plurality of points and can also draw points across scenes, the roaming points are drawn in sequence from a starting point to an end point to form a complete roaming path, and scenes can be freely switched according to the drawn paths in the actual roaming process;
s2: rendering a panoramic picture, rendering a cross panoramic picture according to a designed panoramic scheme, wherein each space corresponds to a segmentation picture of 6 surfaces and a preview picture of the current panoramic scheme, rendering a finished picture storage server, and simultaneously recording a corresponding path for a panoramic display engine to obtain a corresponding panoramic picture;
s3: collecting coordinates, recording space point coordinates in S1 in the rendering process of S2, storing the coordinate points in a database, and simultaneously recording scene information and a browsing path in S1;
s4: generating a panoramic file, and generating a panoramic roaming file to be displayed according to the data of S2 and S3, wherein the panoramic roaming file comprises panoramic roaming scene information, path information, hot spot plane coordinates and the like;
s5: acquiring the position of a space point, acquiring the coordinate position of the space point in each scene by using a specific algorithm and combining scene information parameters according to the panoramic roaming file of S4, acquiring the positions of multiple points, and extracting the position of the space point from the panoramic display file, wherein the specific method comprises the following steps:
(1) the coordinates P (x, y) of the plane points in the current scene, the camera height H,
(2) plane coordinates P1(x1, y1) of the next scene are obtained,
(3) taking the difference value of P (x, y) and P1(x1, y1) to obtain P2(x1-x, y1-y), namely P2(m, n),
(4) the azimuth angle P2 to the camera is calculated using the arctan function. The formula is as follows:
F=Math.atan2(n,m)*180.0/Math.PI+90;
where n is the difference between the next scene and the current scene with respect to the y-axis of the camera coordinates, m is the difference between the next scene and the current scene with respect to the x-axis of the camera coordinates,
(5) the distance from P2(m, n) to the camera is calculated as follows:
D=Math.pow((m2+n2),0.5);
(6) the azimuth angle from the current scene camera to P2 is calculated by using an arctangent function, and the formula is as follows:
T=Math.atan(H/D)*180/Math.PI;
h is the height value of the camera in the current scene, the height value is fixed relative to the same scene, and D is the distance from a point P2 to the center coordinate of the camera;
(7) the hot spot scaling is calculated and,
S=T/90*1.5;
wherein T is an azimuth angle from the current scene to a P2 coordinate, the hot spot scaling is determined by dividing the azimuth angle by 90 degrees and multiplying the azimuth angle by a factor of 1.5 times, and the closer the hot spot is to the center point of the camera, the farther the hot spot is, the smaller the hot spot is;
(8) calculating the self rotation degree of the current hot spot
R=T-90;
Wherein T is an azimuth angle from a current scene camera to P2, T-90 degrees obtains an angle position of a current hot spot rotating by +/-90 degrees on the same plane, and the larger the distance of a camera point deviating from the center is, the more the shape of the hot spot tends to be an ellipse;
(9) determining the position of the current hotspot in space according to the calculated P2(m, n) coordinate, the distance D of P2 from the center point of the camera, and the azimuth angle of P2 relative to the center point of the camera;
the relative positions of the current space and other hot spots are changed after the scene is refreshed, so that the current space is the next scene relative to other hot spots, and the drawing of the hot spot roaming path is synchronously refreshed according to the real-time scene conversion;
s6: and drawing a multipoint path, acquiring coordinates of path points in the virtual scene according to the steps, and drawing the multipoint roaming path in the corresponding scene by using an image drawing technology.
Compared with the prior art, the invention has the beneficial effects that:
(1) the cubic panorama is selected as a research target, and the cubic panorama has a simple structure, is beneficial to image storage and compression, is more suitable for multi-space free switching, is beneficial to virtual roaming of freely-specified roaming routes among panoramas, and realizes seamless multi-scene immersive experience;
(2) in the actual three-dimensional roaming scene expression, the switching of a plurality of scenes is converted into visual drawing operation, and different user-defined roaming route strategies are adopted for different roaming scenes, so that the roaming drawing of the high-quality virtual panoramic image is realized;
(3) the method adopts the drawing technology based on the image to avoid the harsh requirement of huge data volume on system hardware during the reconstruction of the three-dimensional scene, ensures that the scene roaming which is as coherent as possible is obtained by using the path coordinate data as little as possible through reasonably drawing the panoramic path points, and further completes the seamless immersive virtual roaming of any specified path without restriction;
(4) the invention realizes the operation of switching the visual panoramic roaming scene. The method for switching the roaming scene with the self-defined path in the multipoint and multi-scene is simple, efficient and practical. The actual interactive experience in the panoramic roaming use process is improved.
Drawings
FIG. 1 is a schematic view of a 2D mode panoramic path rendering according to the present invention;
FIG. 2 is a schematic view of a 3D mode panoramic path of the present invention;
FIG. 3 is a schematic view of the panoramic multipoint calculation of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail with reference to the following embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
Example 1
Referring to fig. 1-3, the present invention provides a technical solution: a method for realizing panoramic multipoint roaming comprises the following steps:
s1: drawing a roaming path, drawing virtual position points roaming in each space on a planar design drawing of the panoramic scheme, wherein the points represent positions of the camera in the space, one space can have a plurality of points and can also draw points across scenes, the roaming points are drawn in sequence from a starting point to an end point to form a complete roaming path, and scenes can be freely switched according to the drawn paths in the actual roaming process;
s2: rendering a panoramic picture, rendering a cross panoramic picture according to a designed panoramic scheme, wherein each space corresponds to a segmentation picture of 6 surfaces and a preview picture of the current panoramic scheme, rendering a finished picture storage server, and simultaneously recording a corresponding path for a panoramic display engine to obtain a corresponding panoramic picture;
s3: collecting coordinates, recording space point coordinates in S1 in the rendering process of S2, storing the coordinate points in a database, and simultaneously recording scene information and a browsing path in S1;
s4: generating a panoramic file, and generating a panoramic roaming file to be displayed according to the data of S2 and S3, wherein the panoramic roaming file comprises panoramic roaming scene information, path information, hot spot plane coordinates and the like;
s5: acquiring the position of a space point, acquiring the coordinate position of the space point in each scene by using a specific algorithm and combining scene information parameters according to the panoramic roaming file of S4, acquiring the positions of multiple points, and extracting the position of the space point from the panoramic display file, wherein the specific method comprises the following steps:
(1) the coordinates P (x, y) of the plane points in the current scene, the camera height H,
(2) plane coordinates P1(x1, y1) of the next scene are obtained,
(3) taking the difference value of P (x, y) and P1(x1, y1) to obtain P2(x1-x, y1-y), namely P2(m, n),
(4) the azimuth angle P2 to the camera is calculated using the arctan function. The formula is as follows:
F=Math.atan2(n,m)*180.0/Math.PI+90;
where n is the difference between the next scene and the current scene with respect to the y-axis of the camera coordinates, m is the difference between the next scene and the current scene with respect to the x-axis of the camera coordinates,
(5) the distance from P2(m, n) to the camera is calculated as follows:
D=Math.pow((m2+n2),0.5);
(6) the azimuth angle from the current scene camera to P2 is calculated by using an arctangent function, and the formula is as follows:
T=Math.atan(H/D)*180/Math.PI;
h is the height value of the camera in the current scene, the height value is fixed relative to the same scene, and D is the distance from a point P2 to the center coordinate of the camera;
(7) the hot spot scaling is calculated and,
S=T/90*1.5;
wherein T is an azimuth angle from the current scene to a P2 coordinate, the hot spot scaling is determined by dividing the azimuth angle by 90 degrees and multiplying the azimuth angle by a factor of 1.5 times, and the closer the hot spot is to the center point of the camera, the farther the hot spot is, the smaller the hot spot is;
(8) calculating the self rotation degree of the current hot spot
R=T-90;
Wherein T is an azimuth angle from a current scene camera to P2, T-90 degrees obtains an angle position of a current hot spot rotating by +/-90 degrees on the same plane, and the larger the distance of a camera point deviating from the center is, the more the shape of the hot spot tends to be an ellipse;
(9) determining the position of the current hotspot in space according to the calculated P2(m, n) coordinate, the distance D of P2 from the center point of the camera, and the azimuth angle of P2 relative to the center point of the camera;
the relative positions of the current space and other hot spots are changed after the scene is refreshed, so that the current space is the next scene relative to other hot spots, and the drawing of the hot spot roaming path is synchronously refreshed according to the real-time scene conversion;
s6: and drawing a multipoint path, acquiring coordinates of path points in the virtual scene according to the steps, and drawing the multipoint roaming path in the corresponding scene by using an image drawing technology.
Once the panorama and the related scene information are dynamically loaded from the panorama file, a panorama multipoint roaming path can be generated in real time. The whole process is processed on line, at present, the ultra-high-definition panoramic image can be completely displayed within a plurality of seconds, and near real-time multipoint path real-time drawing is basically realized.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any person skilled in the art should be considered to be within the technical scope of the present invention, and the technical solutions and the inventive concepts thereof according to the present invention should be equivalent or changed within the scope of the present invention.

Claims (1)

1. A method for realizing panoramic multipoint roaming is characterized in that: the method for realizing the panoramic multipoint roaming specifically comprises the following steps:
s1: drawing a roaming path, drawing virtual position points roaming in each space on a planar design drawing of the panoramic scheme, wherein the points represent positions of the camera in the space, one space can have a plurality of points and can also draw points across scenes, the roaming points are drawn in sequence from a starting point to an end point to form a complete roaming path, and scenes can be freely switched according to the drawn paths in the actual roaming process;
s2: rendering a panoramic picture, rendering a cross panoramic picture according to a designed panoramic scheme, wherein each space corresponds to a segmentation picture of 6 surfaces and a preview picture of the current panoramic scheme, rendering a finished picture storage server, and simultaneously recording a corresponding path for a panoramic display engine to obtain a corresponding panoramic picture;
s3: collecting coordinates, recording space point coordinates in S1 in the rendering process of S2, storing the coordinate points in a database, and simultaneously recording scene information and a browsing path in S1;
s4: generating a panoramic file, and generating a panoramic roaming file to be displayed according to the data of S2 and S3, wherein the panoramic roaming file comprises panoramic roaming scene information, path information and hot spot plane coordinates;
s5: acquiring the position of a space point, acquiring the coordinate position of the space point in each scene by using a specific algorithm and combining scene information parameters according to the panoramic roaming file of S4, acquiring the positions of multiple points, and extracting the position of the space point from the panoramic display file, wherein the specific method comprises the following steps:
(1) the coordinates P (x, y) of the plane points in the current scene, the camera height H,
(2) plane coordinates P1(x1, y1) of the next scene are obtained,
(3) taking the difference value of P (x, y) and P1(x1, y1) to obtain P2(x1-x, y1-y), namely P2(m, n),
(4) the azimuth angle P2 to the camera is calculated using the arctangent function, as follows:
F=Math.atan2(n,m)*180.0/Math.PI+90;
where n is the difference between the next scene and the current scene with respect to the y-axis of the camera coordinates, m is the difference between the next scene and the current scene with respect to the x-axis of the camera coordinates,
(5) the distance from P2(m, n) to the camera is calculated as follows:
D=Math.pow((m2+n2),0.5);
(6) the azimuth angle from the current scene camera to P2 is calculated by using an arctangent function, and the formula is as follows:
T=Math.atan(H/D)*180/Math.PI;
h is the height value of the camera in the current scene, the height value is fixed relative to the same scene, and D is the distance from a point P2 to the center coordinate of the camera;
(7) the hot spot scaling is calculated and,
S=T/90*1.5;
wherein T is an azimuth angle from the current scene to a P2 coordinate, the hot spot scaling is determined by dividing the azimuth angle by 90 degrees and multiplying the azimuth angle by a factor of 1.5 times, and the closer the hot spot is to the center point of the camera, the farther the hot spot is, the smaller the hot spot is;
(8) calculating the self rotation degree of the current hot spot
R=T-90;
Wherein T is an azimuth angle from a current scene camera to P2, T-90 degrees obtains an angle position of a current hot spot rotating by +/-90 degrees on the same plane, and the larger the distance of a camera point deviating from the center is, the more the shape of the hot spot tends to be an ellipse;
(9) determining the position of the current hotspot in space according to the calculated P2(m, n) coordinate, the distance D of P2 from the center point of the camera, and the azimuth angle of P2 relative to the center point of the camera;
the relative positions of the current space and other hot spots are changed after the scene is refreshed, so that the current space is the next scene relative to other hot spots, and the drawing of the hot spot roaming path is synchronously refreshed according to the real-time scene conversion;
s6: and drawing a multipoint path, acquiring coordinates of path points in the virtual scene according to the steps, and drawing the multipoint roaming path in the corresponding scene by using an image drawing technology.
CN201810096799.7A 2018-01-31 2018-01-31 Method for realizing panoramic multipoint roaming Active CN108257219B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810096799.7A CN108257219B (en) 2018-01-31 2018-01-31 Method for realizing panoramic multipoint roaming

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810096799.7A CN108257219B (en) 2018-01-31 2018-01-31 Method for realizing panoramic multipoint roaming

Publications (2)

Publication Number Publication Date
CN108257219A CN108257219A (en) 2018-07-06
CN108257219B true CN108257219B (en) 2021-05-11

Family

ID=62743409

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810096799.7A Active CN108257219B (en) 2018-01-31 2018-01-31 Method for realizing panoramic multipoint roaming

Country Status (1)

Country Link
CN (1) CN108257219B (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110728755B (en) * 2018-07-16 2022-09-27 阿里巴巴集团控股有限公司 Method and system for roaming among scenes, model topology creation and scene switching
CN109461210B (en) * 2018-10-15 2023-03-21 杭州群核信息技术有限公司 Panoramic roaming method for online home decoration
CN110134478B (en) * 2019-04-28 2022-04-05 深圳市思为软件技术有限公司 Scene conversion method and device of panoramic scene and terminal equipment
CN111340686B (en) * 2020-02-19 2023-05-23 华南理工大学 Virtual reality scene assessment method, system and medium with crowd bias
CN111524230B (en) * 2020-04-01 2021-01-26 中南大学 Linkage browsing method for three-dimensional model and unfolded panoramic image and computer system
CN111698424A (en) * 2020-06-22 2020-09-22 四川易热科技有限公司 Method for complementing live-action roaming 3D information through common camera
CN111698425B (en) * 2020-06-22 2021-11-23 四川可易世界科技有限公司 Method for realizing consistency of real scene roaming technology
CN111949122A (en) * 2020-06-23 2020-11-17 微幻科技(北京)有限公司 Method and device for generating virtual roaming data
CN113660528B (en) * 2021-05-24 2023-08-25 杭州群核信息技术有限公司 Video synthesis method and device, electronic equipment and storage medium
CN113747138A (en) * 2021-07-30 2021-12-03 杭州群核信息技术有限公司 Video generation method and device for virtual scene, storage medium and electronic equipment
CN113870429B (en) * 2021-09-29 2022-04-26 生态环境部卫星环境应用中心 Web end three-dimensional scene posture self-adaptive path roaming method considering terrain fluctuation
CN114357348B (en) * 2021-12-29 2023-05-09 北京有竹居网络技术有限公司 Display method and device and electronic equipment
CN115857702B (en) * 2023-02-28 2024-02-02 北京国星创图科技有限公司 Scene roaming and visual angle conversion method under space scene
CN116188702B (en) * 2023-04-28 2023-07-25 北京德塔精要信息技术有限公司 Method, system, equipment and storage medium for constructing camera roaming route
CN116452718B (en) * 2023-06-15 2023-09-12 山东捷瑞数字科技股份有限公司 Path planning method, system, device and storage medium for scene roaming

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101055494A (en) * 2006-04-13 2007-10-17 上海虚拟谷数码科技有限公司 Dummy scene roaming method and system based on spatial index cube panoramic video
CN102054121A (en) * 2009-11-04 2011-05-11 沈阳隆惠科技有限公司 Method for building 3D (three-dimensional) panoramic live-action online game platform
CN102096938A (en) * 2011-01-27 2011-06-15 中国科学院遥感应用研究所 Construction method capable of measuring panoramic picture
CN104219584A (en) * 2014-09-25 2014-12-17 广州市联文信息科技有限公司 Reality augmenting based panoramic video interaction method and system
EP3029989A1 (en) * 2014-12-01 2016-06-08 Huawei Technologies Co., Ltd. Wlan user roaming

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101055494A (en) * 2006-04-13 2007-10-17 上海虚拟谷数码科技有限公司 Dummy scene roaming method and system based on spatial index cube panoramic video
CN102054121A (en) * 2009-11-04 2011-05-11 沈阳隆惠科技有限公司 Method for building 3D (three-dimensional) panoramic live-action online game platform
CN102096938A (en) * 2011-01-27 2011-06-15 中国科学院遥感应用研究所 Construction method capable of measuring panoramic picture
CN104219584A (en) * 2014-09-25 2014-12-17 广州市联文信息科技有限公司 Reality augmenting based panoramic video interaction method and system
EP3029989A1 (en) * 2014-12-01 2016-06-08 Huawei Technologies Co., Ltd. Wlan user roaming

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
A WebGL-based 3D virtual home roaming by seamlessly connecting videos to panoramas;Lingjun Tao等;《IEEE Xplore》;20160218;第498-503页 *
多视点虚拟场景平滑漫游技术研究;李婷婷;《中国优秀硕士学位论文全文数据库.信息科技辑》;20160415(第04期);第I138-1458页 *

Also Published As

Publication number Publication date
CN108257219A (en) 2018-07-06

Similar Documents

Publication Publication Date Title
CN108257219B (en) Method for realizing panoramic multipoint roaming
US10846913B2 (en) System and method for infinite synthetic image generation from multi-directional structured image array
US10818029B2 (en) Multi-directional structured image array capture on a 2D graph
US11636637B2 (en) Artificially rendering images using viewpoint interpolation and extrapolation
CN106296783B (en) A kind of space representation method of combination space overall situation 3D view and panoramic pictures
US10650574B2 (en) Generating stereoscopic pairs of images from a single lens camera
US10262426B2 (en) System and method for infinite smoothing of image sequences
US10719939B2 (en) Real-time mobile device capture and generation of AR/VR content
EP2643820B1 (en) Rendering and navigating photographic panoramas with depth information in a geographic information system
WO2020181125A1 (en) Method, apparatus, medium, and device for processing multi-angle free-perspective video data
Thatte et al. Depth augmented stereo panorama for cinematic virtual reality with head-motion parallax
WO2013074561A1 (en) Modifying the viewpoint of a digital image
US20210225065A1 (en) Creating action shot video from multi-view capture data
WO2022002181A1 (en) Free viewpoint video reconstruction method and playing processing method, and device and storage medium
EP3857499A1 (en) Panoramic light field capture, processing and display
WO2024077897A1 (en) Virtual scene display control method and apparatus, storage medium and electronic device
CN105301789A (en) Stereoscopic display device following human eye positions
Peer et al. Panoramic depth imaging: Single standard camera approach
CN205193366U (en) Follow people eye position's stereoscopic display device
CN106445282A (en) Interaction method based on augmented reality
CN115512089A (en) Rapid browsing method of BIM (building information modeling) model
CN109961395A (en) The generation of depth image and display methods, device, system, readable medium
WO2022022548A1 (en) Free viewpoint video reconstruction and playing processing method, device, and storage medium
Thatte et al. Real-World Virtual Reality With Head-Motion Parallax
Zhu et al. Multi-view Generation Method Based on the Sparse Feature Points Matching

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant