CN109427095B - Method and system for displaying mixed reality scene - Google Patents

Method and system for displaying mixed reality scene Download PDF

Info

Publication number
CN109427095B
CN109427095B CN201710748651.2A CN201710748651A CN109427095B CN 109427095 B CN109427095 B CN 109427095B CN 201710748651 A CN201710748651 A CN 201710748651A CN 109427095 B CN109427095 B CN 109427095B
Authority
CN
China
Prior art keywords
equipment
positioning
scene
player
virtual
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710748651.2A
Other languages
Chinese (zh)
Other versions
CN109427095A (en
Inventor
刘德建
苏文瑛
潘海珲
江卓鸿
李学科
吕鹏飞
王熠中
陈宏展
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Fujian TQ Digital Co Ltd
Original Assignee
Fujian TQ Digital Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Fujian TQ Digital Co Ltd filed Critical Fujian TQ Digital Co Ltd
Priority to CN201710748651.2A priority Critical patent/CN109427095B/en
Publication of CN109427095A publication Critical patent/CN109427095A/en
Application granted granted Critical
Publication of CN109427095B publication Critical patent/CN109427095B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/006Mixed reality
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/20Input arrangements for video game devices
    • A63F13/21Input arrangements for video game devices characterised by their sensors, purposes or types
    • A63F13/213Input arrangements for video game devices characterised by their sensors, purposes or types comprising photodetecting means, e.g. cameras, photodiodes or infrared cells
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/25Output arrangements for video game devices
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/25Output arrangements for video game devices
    • A63F13/26Output arrangements for video game devices having at least one additional display device, e.g. on the game controller or outside a game booth
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/50Controlling the output signals based on the game progress
    • A63F13/52Controlling the output signals based on the game progress involving aspects of the displayed game scene
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F13/00Video games, i.e. games using an electronically generated display having two or more dimensions
    • A63F13/60Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor
    • A63F13/65Generating or modifying game content before or while executing the game program, e.g. authoring tools specially adapted for game development or game-integrated level editor automatically by game devices or servers from real world data, e.g. measurement in live racing competition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/10Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals
    • A63F2300/1087Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by input arrangements for converting player-generated signals into game device control signals comprising photodetecting means, e.g. a camera
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/30Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game characterized by output arrangements for receiving control signals generated by the game device
    • AHUMAN NECESSITIES
    • A63SPORTS; GAMES; AMUSEMENTS
    • A63FCARD, BOARD, OR ROULETTE GAMES; INDOOR GAMES USING SMALL MOVING PLAYING BODIES; VIDEO GAMES; GAMES NOT OTHERWISE PROVIDED FOR
    • A63F2300/00Features of games using an electronically generated display having two or more dimensions, e.g. on a television screen, showing representations related to the game
    • A63F2300/60Methods for processing data by generating or executing the game program
    • A63F2300/69Involving elements of the real world in the game world, e.g. measurement in live races, real video

Abstract

The invention relates to the field of mixed reality, in particular to a method and a system for displaying a mixed reality scene. The invention configures the authority for more than three MR devices; the permissions comprise player permissions, on-site observer permissions and off-site observer permissions; the player MR equipment with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the player MR equipment; the spectator MR equipment and the server with the permission of the observers in the field respectively update the virtual scenes corresponding to the spectator MR equipment and the server according to the attribute values; the entity camera acquires a real scene; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera; the server superposes the real scene and the virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene; and displaying the mixed reality scene by the display equipment. The mixed reality scene observed by the player wearing the MR equipment is displayed in a third person perspective.

Description

Method and system for displaying mixed reality scene
Technical Field
The invention relates to the field of mixed reality, in particular to a method and a system for displaying a mixed reality scene.
Background
VR (virtual reality) technology is known today to simulate an environment by computer-generated real-time dynamic three-dimensional stereo images and to use sensing devices to enable interaction. VR technology creates a new virtual world, which cannot be linked with the real world.
MR (Mix Reality mixed Reality) technology has been proposed in recent years by Microsoft corporation to merge virtual and Reality together. The general process is as follows:
1) MR-enabled devices model in real time by constantly scanning the real-world environment around the experiencer.
2) The physical coordinates of the device in real space are calculated using a Spatial Understanding (Spatial Understanding) technique.
3) The real world is mapped into a virtual coordinate system using a Spatial Mapping (Spatial Mapping) technique with the MR device as the origin of coordinates.
4) The virtual environment generated by calculation is superposed on the real world, and the interaction with virtual things is realized through gestures, so that the virtual environment has more natural user experience compared with a VR device.
The current MR game on Windows Store can be divided into a stand-alone version and a networked version. The networking version is that a PC is appointed as a server, then a plurality of MR devices are connected together through a local area network, and the server is responsible for managing and forwarding scene data, so that the consistent virtual scene seen on different terminals is realized.
However, only the players who actually participate in the existing halftone MR game can see the virtual and real combined pictures, and other people cannot obtain the same experience as the players through the spectator.
Disclosure of Invention
The technical problem to be solved by the invention is as follows: how to display a mixed reality scene viewed by a player wearing the MR device from a third person perspective.
In order to solve the technical problems, the invention adopts the technical scheme that:
the invention provides a method for displaying a mixed reality scene, which comprises the following steps:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, the MR equipment of the player with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the MR equipment of the player;
s3, the onlooker MR equipment with the in-field observer authority and the server respectively update the virtual scenes corresponding to the server according to the attribute values;
s4, the entity camera acquires a real scene; a positioning MR device with off-site observer authority is physically fixed with the physical camera;
s5, the server superimposes the real scene and a virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
and S6, displaying the mixed reality scene by the display equipment.
The invention also provides a system for displaying the mixed reality scene, which comprises more than three MR devices, an entity camera, a display device and a server;
the server includes one or more processors and memory, the memory storing a program and configured to perform the following steps by the one or more processors:
s1, configuring authority for more than three MR devices; the permissions comprise player permissions, on-site observer permissions and off-site observer permissions;
s2, when the MR equipment of the player with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the MR equipment of the player, acquiring the attribute value;
s3, updating a virtual scene corresponding to the server according to the attribute value; triggering the onlooker MR equipment with the in-field observer authority to update the virtual scene corresponding to the onlooker MR equipment according to the attribute value;
s4, acquiring a real scene from the entity camera; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera;
s5, superposing the real scene and the virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
and S6, triggering display equipment to display the mixed reality scene.
The invention has the beneficial effects that: the physical camera and the MR equipment are physically fixed, so that a server can indirectly acquire the position information of the physical camera according to the position information of the MR equipment, a virtual scene matched with a real scene can be acquired when the real camera acquires the real scene at any position, the real scene and the virtual scene can be superposed, and a mixed reality scene which can be observed by a player wearing the MR equipment is displayed at the view angle of the current position of the physical camera; viewers who do not wear MR devices can also experience mixed reality scenes viewed by players through live display devices. Meanwhile, the user operation data of the player can be synchronously updated to the MR device with the permission of the observer in the field, and the commentator can better experience the mixed reality scene watched by the player through the MR device with the permission of the observer in the field.
Drawings
FIG. 1 is a block flow diagram of an embodiment of a method for displaying a mixed reality scene;
fig. 2 is a block diagram of a specific implementation of a terminal for displaying a mixed reality scene according to the present invention;
FIG. 3 is a diagram of a hardware deployment according to a second embodiment;
description of the reference symbols:
1. an MR device; 2. a physical camera; 3. a display device; 4. a server; 41. a processor; 42. a memory.
Detailed Description
In order to explain technical contents, achieved objects, and effects of the present invention in detail, the following description is made with reference to the accompanying drawings in combination with the embodiments.
The noun interpretation:
Figure BDA0001390612700000031
referring to fig. 1 to 3 of the drawings,
as shown in fig. 1, the present invention provides a method for displaying a mixed reality scene, including:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, the MR equipment of the player with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the MR equipment of the player;
s3, the onlooker MR equipment and the server with the permission of the observers in the field respectively update the virtual scenes corresponding to the onlooker MR equipment and the server according to the attribute values;
s4, the entity camera acquires a real scene; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera;
s5, the server superimposes the real scene and a virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
and S6, displaying the mixed reality scene by the display equipment.
Further, the method also comprises the following steps:
acquiring a first space model corresponding to the player MR device;
acquiring a second space model corresponding to the onlooker MR device;
acquiring a third spatial model corresponding to the positioning MR device;
presetting a virtual object with a constant position;
and mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object.
As can be seen from the above description, a fixed virtual object is preset at a position in a real space, and the virtual object is used as a reference object to map a third space model corresponding to a positioning MR device physically fixed by a physical camera, a first space model corresponding to a player MR device worn by a player, and a second space model corresponding to a spectator MR device worn by an in-field commentator to the same coordinate system, so as to synchronize space models corresponding to different MR devices, and enable other MR devices to accurately update their own virtual scenes when data of the player MR devices is updated.
Further, the S5 specifically is:
acquiring the coordinates of the positioning MR equipment in the coordinate system in real time to obtain positioning coordinates;
acquiring a field angle of an entity camera;
converting a virtual scene corresponding to the server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene; rendering the current virtual scene to the real scene.
Further, still include:
horizontally placing a physical camera;
modifying the initial horizontal rotation offset of the positioning MR device to be zero;
calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset;
the initial vertical rotational offset of the positioning MR device is modified depending on the relative vertical rotational offset.
As can be seen from the above description, since the physical fixation of the physical camera and the positioning MR device is difficult to ensure that the positioning MR device is parallel to the physical camera, the real-time rotational offset of the positioning MR device is difficult to accurately reflect the real-time rotational offset of the physical camera. However, the amount of rotational displacement has a great influence on the final effect, and for example, a rotational deviation of the subject from the camera 10m,1 ° causes a translational misalignment of at least 17 cm. According to the method, the solid camera is horizontally placed by means of the level meter, according to the principle that the XOZ plane is consistent with the real horizontal plane when the MR equipment establishes a self coordinate system, the initial horizontal rotation offset of the positioning MR equipment is modified to be zero no matter how much the current horizontal rotation offset of the positioning MR equipment is, and then the initial vertical rotation offset is modified, namely the initial orientation of the virtual camera corresponding to the positioning MR equipment is modified to be consistent with the initial orientation of the solid camera, so that the real-time rotation offset of the positioning MR equipment is consistent with the real-time rotation offset of the solid camera.
Further, the S3 specifically is:
when the server receives an attribute value from one MR device, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
the server updates a virtual scene corresponding to the server according to the attribute value;
the server sends the attribute value to the onlooker MR equipment;
and the onlooker MR device updates the virtual scene corresponding to the onlooker MR device according to the attribute value.
According to the description, when the server receives the updating data corresponding to the virtual scene, the authority of the MR device of the updating data source is verified, and the disorder of the game process caused by the misoperation of other MR devices except the player MR device is avoided.
As shown in fig. 2, the present invention further provides a system for displaying a mixed reality scene, which includes more than three MR devices 1, an entity camera 2, a display device 3 and a server 4;
the server 4 comprises one or more processors 41 and a memory 42, the memory 42 storing programs and being configured to perform the following steps by the one or more processors 41:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, when the MR equipment of the player with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the MR equipment of the player, acquiring the attribute value;
s3, updating a virtual scene corresponding to the server according to the attribute value; triggering the onlooker MR equipment with the in-field observer authority to update the virtual scene corresponding to the onlooker MR equipment according to the attribute value;
s4, acquiring a real scene from the entity camera; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera;
s5, superposing the real scene and the virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
and S6, triggering display equipment to display the mixed reality scene.
Further, still include:
acquiring a first space model corresponding to the player MR device;
acquiring a second space model corresponding to the bystander MR equipment;
acquiring a third spatial model corresponding to the positioning MR device;
presetting a virtual object with a constant position;
and mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object.
Further, the S5 specifically is:
acquiring coordinates of the positioning MR equipment in the coordinate system in real time to obtain positioning coordinates;
acquiring the field angle of the entity camera;
converting a virtual scene corresponding to the server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene; rendering the current virtual scene to the real scene.
Further, still include:
horizontally placing a solid camera;
modifying the initial horizontal rotational offset of the positioning MR device to zero;
calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset;
the initial vertical rotational offset of the positioning MR device is modified depending on the relative vertical rotational offset.
Further, the S3 specifically is:
when an attribute value from one MR device is received, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
updating a virtual scene corresponding to the server according to the attribute value;
sending the attribute value to a spectator MR device;
and triggering the onlooker MR device to update the virtual scene corresponding to the onlooker MR device according to the attribute value.
The first embodiment of the invention is as follows:
the embodiment provides a method for displaying a mixed reality scene, which comprises the following steps:
s1, configuring authority for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, acquiring a first space model corresponding to player MR equipment; acquiring a second space model corresponding to the onlooker MR device; acquiring a third spatial model corresponding to the positioning MR device;
the player MR equipment is configured with player authority, directly participates in game interaction, and can control part of game elements in the game; the onlooker MR equipment is configured with the authority of an observer in the field, can freely select an angle for watching a virtual-real combined scene, personally watches the game process of a player in the scene, but cannot operate game elements; the positioning MR equipment is configured with the authority of an off-site observer, and the position information of the positioning MR equipment is sent to the server in real time, so that the server can indirectly obtain the position information of the entity camera physically fixed with the positioning MR equipment according to the position information of the positioning MR equipment.
S3, presetting a virtual object with a constant position; mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object;
optionally, when running an application, the MR device, such as a Hololens developed by Microsoft corporation, may establish a left-hand coordinate system with its own position as an origin of coordinates, a projection of a right front side of the MR device on a horizontal plane as a positive Z-axis direction, and a vertically upward direction as a positive Y-axis direction; successively scanning the same space by the player MR equipment, the spectator MR equipment and the positioning MR equipment, respectively modeling the same space in respective coordinate systems, and respectively obtaining a first space model, a second space model and a third space model; a common world coordinate system is established through a world anchor point (namely a virtual object with a fixed position in a real space) which is placed in a space in advance, a conversion matrix from different MR equipment coordinate systems to the world coordinate system is calculated, and virtual scenes constructed by different MR equipment are superposed;
s4, horizontally placing an entity camera; modifying the initial horizontal rotation offset of the positioning MR device to be zero; calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset; modifying an initial vertical rotational offset of the positioning MR device in accordance with the relative vertical rotational offset;
optionally, a vertical rotation offset of the positioning MR device with respect to the physical camera is calculated according to the real scene and the virtual scene corresponding to the positioning MR device, so as to obtain a relative vertical rotation offset, specifically:
placing a virtual vertical slender model in a virtual scene corresponding to the positioning MR equipment, wherein the virtual vertical slender model is positioned in the middle of the visual field of a virtual camera of the positioning MR equipment; a vertical rod is arranged at a position far away from the entity camera in a real scene and is positioned in the middle of the visual field of the entity camera;
superposing a virtual scene and a real scene corresponding to the positioning MR equipment, if the virtual vertical slender model and the vertical rod are not superposed, then:
keeping the position of the virtual camera for positioning the MR device still, and adjusting the rotation amount (namely, rotating around a vertical Y axis) of the virtual camera for positioning the MR device along the horizontal direction, wherein the movement direction of the virtual vertical slender model in the virtual image is opposite to the rotation direction of the virtual camera for positioning the MR device;
and rotating the virtual camera of the positioning MR equipment until the virtual vertical slender model is superposed with the vertical rod, and taking a negative value of the rotation amount of the virtual camera of the positioning MR equipment around the Y axis as the vertical rotation offset.
S5, the player MR equipment with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the player MR equipment;
optionally, the attributes of a virtual object include coordinates, a rotation angle, and a scaling; when a player operates a virtual object, the attribute value of the virtual object changes correspondingly.
S6, the onlooker MR equipment with the in-field observer authority and the server respectively update the virtual scenes corresponding to the server according to the attribute values; the method specifically comprises the following steps:
when the server receives an attribute value from one MR device, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
the server updates a virtual scene corresponding to the server according to the attribute value;
the server sends the attribute value to the onlooker MR equipment;
and the onlooker MR device updates the virtual scene corresponding to the onlooker MR device according to the attribute value.
Optionally, if there is more than one player MR device with player authority, the attribute value may be sent to other player MR devices, and the other player MR devices update the virtual scene corresponding to themselves according to the attribute value.
S7, the entity camera acquires a real scene; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera;
s8, the server superimposes the real scene and a virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene; the method specifically comprises the following steps:
acquiring coordinates of the positioning MR equipment in the coordinate system in real time to obtain positioning coordinates;
acquiring the field angle of the entity camera;
converting a virtual scene corresponding to a server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene; rendering the current virtual scene to the real scene;
optionally, the server includes a renderer for rendering a virtual scene corresponding to the server onto a real scene acquired by the physical camera, so as to obtain a mixed reality scene.
Converting a virtual scene corresponding to the server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene, which specifically comprises the following steps:
setting the field angle parameter, the coordinate and the rotation offset parameter of the virtual camera corresponding to the renderer to be the same as the field angle, the positioning coordinate and the rotation offset of the entity camera; the parameters of the virtual camera and the parameters of the physical camera are completely consistent, so that the virtual scene shooting from the angle of the physical camera is simulated.
S9, displaying the mixed reality scene by display equipment;
alternatively, the reality device is a large screen connected to the server, and the spectators who do not wear the MR device can see the mixed reality scene viewed by the player through the player MR device through the large screen.
The second embodiment of the invention is as follows:
the embodiment provides a system for displaying a mixed reality scene, which comprises more than three MR devices, an entity camera, a display device and a server;
alternatively, the hardware deployment diagram is shown in fig. 3, wherein the icon of the smartphone represents a Hololens device. The dotted line connection refers to connection through a wireless network, the solid line connection refers to connection through a wired network, and the connection between the user and the equipment is physical wearing or holding;
the server includes one or more processors and memory, the memory storing a program and configured to perform the following steps by the one or more processors:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions, and off-site observer permissions.
S2, acquiring a first space model corresponding to player MR equipment; acquiring a second space model corresponding to the bystander MR equipment; a third spatial model corresponding to the positioning of the MR device is acquired.
S3, presetting a virtual object with a constant position; and mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object.
S4, horizontally placing an entity camera; modifying the initial horizontal rotational offset of the positioning MR device to zero; calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset; the initial vertical rotational offset of the positioning MR device is modified depending on the relative vertical rotational offset.
And S5, when the player MR equipment with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the player MR equipment, acquiring the attribute value.
S6, updating the virtual scene corresponding to the server according to the attribute value; triggering the onlooker MR equipment with the in-field observer authority to update the virtual scene corresponding to the onlooker MR equipment according to the attribute value; the method comprises the following specific steps:
when an attribute value from one MR device is received, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
updating a virtual scene corresponding to the server according to the attribute value;
sending the attribute value to a spectator MR device;
and triggering the onlooker MR device to update the virtual scene corresponding to the onlooker MR device according to the attribute value.
S7, acquiring a real scene from the entity camera; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera.
S8, superposing the real scene and the virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene; the method specifically comprises the following steps:
acquiring coordinates of the positioning MR equipment in the coordinate system in real time to obtain positioning coordinates;
acquiring the field angle of the entity camera;
converting a virtual scene corresponding to the server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR device 1 to obtain a current virtual scene; rendering the current virtual scene to the real scene.
And S9, triggering display equipment to display the mixed reality scene.
The third embodiment of the invention is as follows:
the embodiment of the invention provides a system for displaying a mixed reality scene to carry out an MR interactive game of Chaojie character.
The hardware required for this embodiment is shown in table 1:
TABLE 1
Figure BDA0001390612700000121
The background director clicks 'question making' on the server, and disordered virtual scenes of Chinese character fragments appear on all the MR equipment and a large screen connected with the program recording system;
the player 1 operates one of the shards, the data of the shard is uploaded to the server, and then the server broadcasts the data to the MR games running on the MR devices worn by all guests and all Mac Pro through a wireless route. In this embodiment, since the game of "chang jie" character building according to this embodiment is a competitive game, player 1 and player 2 cannot check the progress of each other, and data is not shared. However, the data of the fragment operated by the player 1 and the player 2 is broadcasted to the MR device and the server worn by the guest;
and outputting a virtual-real combined picture after the virtual scene on the virtual scene renderer of the server is mixed with the real scene shot by the camera, wherein the picture is presented in front of each scene appearance through a program recording system.
In summary, according to the method and system for displaying a mixed reality scene provided by the invention, the physical camera and the MR device are physically fixed, so that the server can indirectly obtain the position information of the physical camera according to the position information of the MR device, and thus when the physical camera obtains a real scene at any position, a virtual scene matched with the real scene can be obtained, the real scene and the virtual scene can be superimposed, and the mixed reality scene viewable by a player wearing the MR device can be displayed at the view angle of the current position of the physical camera; viewers who do not wear MR devices can also experience mixed reality scenes viewed by players through live display devices. Meanwhile, the user operation data of the player can be synchronously updated to the MR device with the permission of the observer in the field, and the commentator can better experience the mixed reality scene watched by the player through the MR device with the permission of the observer in the field.
The above description is only an embodiment of the present invention, and is not intended to limit the scope of the present invention, and all equivalent modifications made by the present invention and the contents of the accompanying drawings, which are directly or indirectly applied to the related technical fields, are included in the scope of the present invention.

Claims (4)

1. A method of displaying a mixed reality scene, comprising:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, the MR equipment of the player with the player authority modifies the attribute value of a virtual object in the virtual scene corresponding to the MR equipment of the player;
s3, the onlooker MR equipment and the server with the permission of the observers in the field respectively update the virtual scenes corresponding to the onlooker MR equipment and the server according to the attribute values;
the S3 specifically comprises the following steps:
when the server receives an attribute value from one MR device, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
the server updates a virtual scene corresponding to the server according to the attribute value;
the server sends the attribute value to the onlooker MR equipment;
updating a virtual scene corresponding to the onlooker MR equipment by the onlooker MR equipment according to the attribute value;
s4, the entity camera acquires a real scene; a positioning MR device with off-site observer authority is physically fixed with the physical camera;
s5, the server superimposes the real scene and a virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
the S5 specifically comprises the following steps:
acquiring the coordinates of the positioning MR equipment in a coordinate system in real time to obtain positioning coordinates;
acquiring a field angle of an entity camera;
converting a virtual scene corresponding to a server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene; rendering the current virtual scene to the real scene;
s6, displaying the mixed reality scene by display equipment;
further comprising:
acquiring a first space model corresponding to the player MR device;
acquiring a second space model corresponding to the bystander MR equipment;
acquiring a third space model corresponding to the positioning MR equipment;
presetting a virtual object with a constant position;
and mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object.
2. The method of displaying a mixed reality scene of claim 1, further comprising:
horizontally placing a physical camera;
modifying the initial horizontal rotational offset of the positioning MR device to zero;
calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset;
the initial vertical rotational offset of the positioning MR device is modified depending on the relative vertical rotational offset.
3. A system for displaying a mixed reality scene is characterized by comprising more than three MR devices, an entity camera, a display device and a server;
the server includes one or more processors and memory, the memory storing a program and configured to perform the following steps by the one or more processors:
s1, configuring authorities for more than three MR devices; the permissions include player permissions, on-site observer permissions and off-site observer permissions;
s2, when a player MR device with the player permission modifies the attribute value of a virtual object in a virtual scene corresponding to the player MR device, acquiring the attribute value;
s3, updating a virtual scene corresponding to the server according to the attribute value; triggering the onlooker MR equipment with the in-field observer authority to update the virtual scene corresponding to the onlooker MR equipment according to the attribute value;
the S3 specifically comprises the following steps:
when the server receives an attribute value from one MR device, verifying the authority of the MR device to obtain a verification result; the verification result comprises player authority, on-site observer authority and off-site observer authority;
if the verification result is the player authority, then:
the server updates a virtual scene corresponding to the server according to the attribute value;
the server sends the attribute value to the onlooker MR equipment;
updating a virtual scene corresponding to the onlooker MR equipment by the onlooker MR equipment according to the attribute value;
s4, acquiring a real scene from the entity camera; the positioning MR device with the authority of the off-site observer is physically fixed with the physical camera;
s5, superposing the real scene and the virtual scene corresponding to the server according to the real-time position information of the positioning MR equipment to obtain a mixed reality scene;
the S5 specifically comprises the following steps:
acquiring the coordinates of the positioning MR equipment in a coordinate system in real time to obtain positioning coordinates;
acquiring a field angle of an entity camera;
converting a virtual scene corresponding to a server according to the field angle, the positioning coordinates and the current rotation offset of the positioning MR equipment to obtain a current virtual scene; rendering the current virtual scene to the real scene;
s6, triggering display equipment to display the mixed reality scene;
further comprising:
acquiring a first space model corresponding to the player MR equipment;
acquiring a second space model corresponding to the onlooker MR device;
acquiring a third spatial model corresponding to the positioning MR device;
presetting a virtual object with a constant position;
and mapping the first space model, the second space model and the third space model to the same coordinate system by taking the virtual object with the constant position as a reference object.
4. The system for displaying a mixed reality scene of claim 3, further comprising:
horizontally placing a physical camera;
modifying the initial horizontal rotational offset of the positioning MR device to zero;
calculating the vertical rotation offset of the positioning MR equipment relative to the entity camera according to the real scene and the virtual scene corresponding to the positioning MR equipment to obtain the relative vertical rotation offset;
the initial vertical rotational offset of the positioning MR device is modified depending on the relative vertical rotational offset.
CN201710748651.2A 2017-08-28 2017-08-28 Method and system for displaying mixed reality scene Active CN109427095B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710748651.2A CN109427095B (en) 2017-08-28 2017-08-28 Method and system for displaying mixed reality scene

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710748651.2A CN109427095B (en) 2017-08-28 2017-08-28 Method and system for displaying mixed reality scene

Publications (2)

Publication Number Publication Date
CN109427095A CN109427095A (en) 2019-03-05
CN109427095B true CN109427095B (en) 2022-10-21

Family

ID=65502431

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710748651.2A Active CN109427095B (en) 2017-08-28 2017-08-28 Method and system for displaying mixed reality scene

Country Status (1)

Country Link
CN (1) CN109427095B (en)

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112148122A (en) * 2020-08-25 2020-12-29 中国电子科技集团公司第三十八研究所 Third-party visual angle implementation method for wearable augmented/mixed reality equipment
CN114489342A (en) * 2022-01-29 2022-05-13 联想(北京)有限公司 Image processing method and device and electronic equipment

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104715479A (en) * 2015-03-06 2015-06-17 上海交通大学 Scene reproduction detection method based on augmented virtuality
CN106898049A (en) * 2017-01-18 2017-06-27 北京商询科技有限公司 A kind of spatial match method and system for mixed reality equipment

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20160267720A1 (en) * 2004-01-30 2016-09-15 Electronic Scripting Products, Inc. Pleasant and Realistic Virtual/Augmented/Mixed Reality Experience

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104715479A (en) * 2015-03-06 2015-06-17 上海交通大学 Scene reproduction detection method based on augmented virtuality
CN106898049A (en) * 2017-01-18 2017-06-27 北京商询科技有限公司 A kind of spatial match method and system for mixed reality equipment

Also Published As

Publication number Publication date
CN109427095A (en) 2019-03-05

Similar Documents

Publication Publication Date Title
CN110382066B (en) Mixed reality observer system and method
CN107976811B (en) Virtual reality mixing-based method simulation laboratory simulation method of simulation method
US10692288B1 (en) Compositing images for augmented reality
CN105137705B (en) A kind of creation method and device of virtual ball curtain
US20050219695A1 (en) Horizontal perspective display
WO2018000609A1 (en) Method for sharing 3d image in virtual reality system, and electronic device
US20210038975A1 (en) Calibration to be used in an augmented reality method and system
JPWO2017094543A1 (en) Information processing apparatus, information processing system, information processing apparatus control method, and parameter setting method
CN106210856A (en) Internet video live broadcasting platform is watched the method and system of 3D panoramic video
CN109246404A (en) The more optical projection systems and method of spectators' seat including direction-agile
CN109427095B (en) Method and system for displaying mixed reality scene
CN104317546A (en) Situational interactive experience simulation system
CN106530408A (en) Museum temporary exhibition planning and design system
WO2017062730A1 (en) Presentation of a virtual reality scene from a series of images
Constantine et al. Project esky: Enabling high fidelity augmented reality on an open source platform
Marner et al. Exploring interactivity and augmented reality in theater: A case study of Half Real
EP3542877A1 (en) Optimized content sharing interaction using a mixed reality environment
CN103871094A (en) Swept-volume-based three-dimensional display system data source generating method
CN109427094B (en) Method and system for acquiring mixed reality scene
IJsselsteijn et al. A room with a cue: The efficacy of movement parallax, occlusion, and blur in creating a virtual window
CN115423916A (en) XR (X-ray diffraction) technology-based immersive interactive live broadcast construction method, system and medium
CN114708407A (en) Virtual three-dimensional space information display method, device and program product
CN110197524A (en) Stereo display method, unit and computer readable storage medium
KR102492985B1 (en) A digital theme park implementation system that provides a digital twin-based metaverse
Khedwala et al. Analysis of Auto Generation of 3D Model Using Multiple 2D Graphics to Manifest Through Augmented Reality

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant