CN108805985A - Virtual Space method and apparatus - Google Patents
Virtual Space method and apparatus Download PDFInfo
- Publication number
- CN108805985A CN108805985A CN201810244761.XA CN201810244761A CN108805985A CN 108805985 A CN108805985 A CN 108805985A CN 201810244761 A CN201810244761 A CN 201810244761A CN 108805985 A CN108805985 A CN 108805985A
- Authority
- CN
- China
- Prior art keywords
- virtual space
- virtual
- entrance
- ambient
- state value
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T19/00—Manipulating 3D models or images for computer graphics
- G06T19/006—Mixed reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/20—Perspective computation
- G06T15/205—Image-based rendering
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T15/00—3D [Three Dimensional] image rendering
- G06T15/10—Geometric effects
- G06T15/40—Hidden part removal
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Graphics (AREA)
- General Engineering & Computer Science (AREA)
- Geometry (AREA)
- Computer Hardware Design (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Computing Systems (AREA)
- Processing Or Creating Images (AREA)
Abstract
Inventor discloses a kind of Virtual Space method:By the model carry in virtual scene to Virtual Space origin;Along Virtual Space periphery, ambient is set;The first collision object of setting simultaneously places it in Virtual Space entrance, and the size of the first collision object is identical as Virtual Space entrance size;Setting the second collision object is in virtual camera;Setting Virtual Space entrance initial state value is false, and when virtual camera enters Virtual Space from the Virtual Space entrance, it is true to change its state value;It is fictitious time when virtual camera enters Virtual Space and Virtual Space entrance initial value, triggers collision accident, hide other models in addition to entrance in Virtual Space;When virtual camera leaves Virtual Space, event is left in triggering, shows all models being hidden, and Virtual Space entrance initial state value is set to false as, to realize the function for switching of shuttling naturally back and forth in Virtual Space and realistic space.
Description
Technical field
The present invention relates to computer software fields, more particularly to a kind of Virtual Space method and apparatus.
Background technology
Augmented reality (Augmented Reality, abbreviation AR), be it is a kind of in real time calculate camera image position and
Angle and the technology for adding respective image, the target of this technology is that virtual world is sleeved on real world on the screen and is carried out
It is interactive.
Augmented reality skill is originally existing by " seamless " the integrated new technology of real world information and virtual world information
It is difficult the entity information (visual information, sound, taste, tactile etc.) experienced in the certain time spatial dimension in the real world, leads to
The science and technology such as computer are crossed, are superimposed again after analog simulation, virtual Information application to real world is felt by human sensory
Know, to reach the sensory experience of exceeding reality.True environment and virtual object have been added to the same picture in real time
Or space exists simultaneously.It not only presents the information of real world, but also virtual information is shown simultaneously, two kinds of letters
Breath is complementary to one another, is superimposed.In the augmented reality of visualization, user utilizes Helmet Mounted Display, real world and computer graphic
It is multiple to be synthesized together, the true world can be seen around it.
Augmented reality contain multimedia, three-dimensional modeling, real-time video show and control, Multi-sensor Fusion, reality
When track and registration, scene fusion etc. new technologies and new tool.Augmented reality provides under normal circumstances, can different from the mankind
With the information of perception.This technology is most proposed earlier than nineteen ninety.With the promotion of accompanied electronic product operational capability, augmented reality
Purposes it is more and more wider.
The current Virtual Spaces AR interaction schemes are mostly object to be superimposed in realistic space, and traditional AR experience is mostly
In order to show some object and generate interaction with it.
Invention content
For this reason, it may be necessary to which providing a kind of new AR interactive modes can naturally be worn back and forth with realizing in Virtual Space and realistic space
The function of shuttle switching.
To achieve the above object, a kind of Virtual Space method is inventor provided, is included the following steps:
By the model carry in virtual scene to Virtual Space origin;
Ambient is set along Virtual Space periphery, the ambient does not overlap with Virtual Space and surrounds virtual scene
Region except entrance;The rendering prior to the ambient is rendered to the entrance of the Virtual Space;
The first collision object of setting simultaneously places it in Virtual Space entrance, and the size of the first collision object enters with Virtual Space
Mouth size is identical;Setting the second collision object is in virtual camera;
Setting Virtual Space entrance initial state value is false, and when virtual camera enters void from the Virtual Space entrance
When quasi- space, it is true to change its state value;
It is fictitious time when virtual camera enters Virtual Space and Virtual Space entrance initial value, triggers collision accident, hide
Other models in Virtual Space in addition to entrance;When virtual camera leaves Virtual Space, event is left in triggering, shows quilt
Hiding all models, and Virtual Space entrance initial state value is set to false as.
Further, in the Virtual Space method, in step " by the model carry in virtual scene to Virtual Space
Further include step before origin ":The threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space origin, and with institute
It is virtual scene father node to state Virtual Space origin.
Further, in the Virtual Space method, the ambient is by following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
Further, in the Virtual Space method, the second collision object is preset as ensureing itself and other collision objects
The size of crash response is triggered when colliding enough.
Inventor proposes a kind of Virtual Space device, including model carry unit, ambient setting unit, rendering simultaneously
Unit, collision object setting unit, state value setting unit and hit-treatment unit;
The model carry unit is used for the model carry in virtual scene to Virtual Space origin;
The ambient setting unit be used for along Virtual Space periphery be arranged ambient, the ambient not with Virtual Space
It overlaps and surrounds the region except virtual scene entrance;The rendering unit is for rendering default object, and rendering unit
The rendering prior to the ambient is rendered to the entrance of the Virtual Space;
For the first collision object to be arranged and places it in Virtual Space entrance, described first touches the collision object setting unit
The size for hitting object is identical as Virtual Space entrance size;The collision object setting unit is additionally operable to setting the second collision object in virtual
Video camera;
The state value setting unit is false for Virtual Space entrance initial state value to be arranged, and when virtual camera from
When the Virtual Space entrance enters Virtual Space, it is true that the state value setting unit, which changes its state value,;
The hit-treatment unit is used to when virtual camera enters Virtual Space and Virtual Space entrance initial value be false
When triggering collision accident and hide other models in addition to entrance in Virtual Space;The hit-treatment unit is additionally operable to when void
Event is left in triggering when quasi- video camera leaves Virtual Space, shows that all models being hidden, state value setting unit will be virtual
Space entry initial state value is set to false as.
Further, in the Virtual Space device, the model carry unit is by the model carry in virtual scene
To before the origin of Virtual Space, the threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space origin, and with
The Virtual Space origin is virtual scene father node.
Further, further include ambient processing unit in the Virtual Space device, for being done to the ambient
Following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
Further, in the Virtual Space device, the second collision object is preset as ensureing itself and other collision objects
The size of crash response is triggered when colliding enough.
It is different from the prior art, above-mentioned technical proposal can be used by being superimposed a Virtual Space in realistic space
Family can only be shuttled switching by mobile device in Virtual Space and realistic space, and transition effect is natural.Skill of the present invention
The AR interactions of art scheme can be generated with Virtual Space and be interacted, and be roamed in Virtual Space using the space orientation of AR.In void
In quasi- space, realistic space can just be seen by only penetrating the space of transmission gate, and be needed in realistic space through transmission gate
It can see Virtual Space, the realization for effect of not exposing the false is ensure that with this.
Description of the drawings
Fig. 1 is the flow chart of the Virtual Space method described in an embodiment of the present invention;
Fig. 2 is the structural schematic diagram of the Virtual Space device described in an embodiment of the present invention.
Reference sign:
1- model carry units
2- ambient setting units
3- rendering units
4- collides object setting unit
5- state value setting units
6- hit-treatment units
7- ambient processing units
Specific implementation mode
For the technology contents of technical solution, construction feature, the objects and the effects are described in detail, below in conjunction with specific reality
It applies example and attached drawing is coordinated to be explained in detail.
Referring to Fig. 1, for the flow chart of the method described in an embodiment of the present invention;Described method includes following steps:
S1, the threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space origin, and with the virtual sky
Between origin be virtual scene father node.
S2, by the model carry in virtual scene to Virtual Space origin.
S3, ambient is set along Virtual Space periphery, the ambient does not overlap with Virtual Space and surrounds virtual
Region except scene entrance;The rendering prior to the ambient is rendered to the entrance of the Virtual Space;It is described in this step
Ambient is by following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
S4, setting the first collision object simultaneously place it in Virtual Space entrance, the size of the first collision object with it is virtual empty
Between entrance size it is identical;The second collision object of setting is in virtual camera, and the second collision object is preset as ensureing itself and other
The size of crash response is triggered when collision object collides enough.
S5, setting Virtual Space entrance initial state value are false, and when virtual camera from the Virtual Space entrance into
When entering Virtual Space, it is true to change its state value.
S6, when virtual camera enters Virtual Space and Virtual Space entrance initial value it is fictitious time, triggers collision accident, it is hidden
Hide other models in addition to entrance in Virtual Space;When virtual camera leaves Virtual Space, event is left in triggering, shows
All models being hidden, and Virtual Space entrance initial state value is set to false as.
It is different from the prior art, the Virtual Space method that above-mentioned technical proposal is provided in realistic space by being superimposed one
A Virtual Space can make user that can only shuttle switching, and mistake in Virtual Space and realistic space by mobile device
Cross effect nature.The AR interactions of technical solution of the present invention can be generated with Virtual Space and be interacted, and be existed using the space orientation of AR
It is roamed in Virtual Space.In Virtual Space, realistic space can just be seen by only penetrating the space of transmission gate, and in realistic space
In need just see Virtual Space through transmission gate, the realization for effect of not exposing the false is ensure that with this.
Referring to Fig. 2, the knot of the Virtual Space device described in another embodiment of the present invention proposed simultaneously for inventor
Structure schematic diagram;The Virtual Space device include model carry unit 1, ambient setting unit 2, rendering unit 3, collision object set
Set unit 4, state value setting unit 5, hit-treatment unit 6 and ambient processing unit 7;
The model carry unit 1 is used for the model carry in virtual scene to Virtual Space origin;In addition, the mould
Type carry unit is by before the model carry to Virtual Space origin in virtual scene, also by the three-dimensional mould of Virtual Space entrance
Type bottom centre point is set as Virtual Space origin, and using the Virtual Space origin as virtual scene father node.
The ambient setting unit 2 be used for along Virtual Space periphery be arranged ambient, the ambient not with virtual sky
Between overlap and surround the region except virtual scene entrance;The rendering unit 3 renders single for rendering default object
Member renders the entrance of the Virtual Space rendering prior to the ambient;
The collision object setting unit 4 is for being arranged the first collision object and placing it in Virtual Space entrance, and described first
The size for colliding object is identical as Virtual Space entrance size;The collision object setting unit is additionally operable to setting the second collision object in void
Quasi- video camera, the second collision object are preset as the ruler for ensureing to trigger crash response when it collides with other collision objects enough
It is very little.
The state value setting unit 5 is false for Virtual Space entrance initial state value to be arranged, and works as virtual camera
When entering Virtual Space from the Virtual Space entrance, it is true that the state value setting unit 5, which changes its state value,;
The hit-treatment unit 6 is used to when virtual camera enters Virtual Space and Virtual Space entrance initial value be false
When triggering collision accident and hide other models in addition to entrance in Virtual Space;The hit-treatment unit 6 is additionally operable to work as
Event is left in triggering when virtual camera leaves Virtual Space, shows that all models being hidden, state value setting unit 5 will be empty
Quasi- space entry initial state value is set to false as.
Further, further include ambient processing unit 7 in the Virtual Space device, for the ambient
Do following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
The Virtual Space device realize the flow of Virtual Space method approximately as:
The threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space origin by S1, model carry unit 1, and
Using the Virtual Space origin as virtual scene father node.
S2, model carry unit 1 are by the model carry in virtual scene to Virtual Space origin.
Ambient is arranged along Virtual Space periphery in S3, ambient setting unit 2, and the ambient does not occur with Virtual Space
It is overlapped and surrounds the region except virtual scene entrance;Rendering unit 3 renders prior to the packet entrance of the Virtual Space
Enclose the rendering of object;In this step, the ambient is by following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
S4, collision object setting unit 4 are arranged the first collision object and place it in Virtual Space entrance, the first collision object
Size it is identical as Virtual Space entrance size;It collides object setting unit 4 and second collision object is set in virtual camera, and is described
Second collision object is preset as the size for ensureing to trigger crash response when it collides with other collision objects enough.
It is false that Virtual Space entrance initial state value, which is arranged, in S5 state values setting unit 5, and when virtual camera is from described
When Virtual Space entrance enters Virtual Space, it is true that state value setting unit 5, which changes its state value,.
S6, when virtual camera enters Virtual Space and Virtual Space entrance initial value it is fictitious time, triggers collision accident, touch
It hits processing unit 6 and hides other models in Virtual Space in addition to entrance;When virtual camera leaves Virtual Space, triggering
Event is left, hit-treatment unit 6 shows that all models being hidden, state value setting unit 5 are initial by Virtual Space entrance
State value is set to false as.
It is different from the prior art, the Virtual Space device that above-mentioned technical proposal is provided in realistic space by being superimposed one
A Virtual Space can make user that can only shuttle switching, and mistake in Virtual Space and realistic space by mobile device
Cross effect nature.The AR interactions of technical solution of the present invention can be generated with Virtual Space and be interacted, and be existed using the space orientation of AR
It is roamed in Virtual Space.In Virtual Space, realistic space can just be seen by only penetrating the space of transmission gate, and in realistic space
In need just see Virtual Space through transmission gate, the realization for effect of not exposing the false is ensure that with this.
It should be noted that herein, relational terms such as first and second and the like are used merely to a reality
Body or operation are distinguished with another entity or operation, are deposited without necessarily requiring or implying between these entities or operation
In any actual relationship or order or sequence.Moreover, the terms "include", "comprise" or its any other variant are intended to
Non-exclusive inclusion, so that process, method, article or terminal device including a series of elements include not only those
Element, but also include other elements that are not explicitly listed, or further include for this process, method, article or end
The intrinsic element of end equipment.In the absence of more restrictions, being limited by sentence " including ... " or " including ... "
Element, it is not excluded that there is also other elements in process, method, article or the terminal device including the element.This
Outside, herein, " being more than ", " being less than ", " being more than " etc. are interpreted as not including this number;" more than ", " following ", " within " etc. understandings
It includes this number to be.
It should be understood by those skilled in the art that, the various embodiments described above can be provided as method, apparatus or computer program production
Product.Complete hardware embodiment, complete software embodiment or embodiment combining software and hardware aspects can be used in these embodiments
Form.All or part of step in the method that the various embodiments described above are related to can be instructed by program relevant hardware come
It completes, the program can be stored in the storage medium that computer equipment can be read, for executing the various embodiments described above side
All or part of step described in method.The computer equipment, including but not limited to:Personal computer, server, general-purpose computations
It is machine, special purpose computer, the network equipment, embedded device, programmable device, intelligent mobile terminal, smart home device, wearable
Smart machine, vehicle intelligent equipment etc.;The storage medium, including but not limited to:RAM, ROM, magnetic disc, tape, CD, sudden strain of a muscle
It deposits, USB flash disk, mobile hard disk, storage card, memory stick, webserver storage, network cloud storage etc..
The various embodiments described above are with reference to method, equipment (system) and the computer program product according to embodiment
Flowchart and/or the block diagram describes.It should be understood that can be realized by computer program instructions every in flowchart and/or the block diagram
The combination of flow and/or box in one flow and/or box and flowchart and/or the block diagram.These computers can be provided
Program instruction is to the processor of computer equipment to generate a machine so that the finger executed by the processor of computer equipment
It enables and generates to specify in one flow of flow chart or multiple flows and/or one box of block diagram or multiple boxes
The device of function.
These computer program instructions, which may also be stored in, can guide computer equipment computer operate in a specific manner to set
In standby readable memory so that the instruction generation being stored in the computer equipment readable memory includes the manufacture of command device
Product, command device realization refer in one flow of flow chart or multiple flows and/or one box of block diagram or multiple boxes
Fixed function.
These computer program instructions can be also loaded on computer equipment so that be executed on a computing device a series of
To generate computer implemented processing, the instruction to execute on a computing device is provided for realizing in flow operating procedure
The step of function of being specified in one flow of figure or multiple flows and/or one box of block diagram or multiple boxes.
Although the various embodiments described above are described, once a person skilled in the art knows basic wounds
The property made concept, then additional changes and modifications can be made to these embodiments, so example the above is only the implementation of the present invention,
It is not intended to limit the scope of patent protection of the present invention, it is every to utilize equivalent structure made by description of the invention and accompanying drawing content
Or equivalent process transformation, it is applied directly or indirectly in other relevant technical fields, the patent for being similarly included in the present invention
Within protection domain.
Claims (8)
1. a kind of Virtual Space method, which is characterized in that include the following steps:
By the model carry in virtual scene to Virtual Space origin;
Ambient is set along Virtual Space periphery, the ambient does not overlap with Virtual Space and surrounds virtual scene entrance
Except region;The rendering prior to the ambient is rendered to the entrance of the Virtual Space;
The first collision object of setting simultaneously places it in Virtual Space entrance, and the size and Virtual Space entrance of the first collision object are big
It is small identical;Setting the second collision object is in virtual camera;
It is false that Virtual Space entrance initial state value, which is arranged, and when virtual camera enters virtual sky from the Virtual Space entrance
Between when, it is true to change its state value;
It is fictitious time when virtual camera enters Virtual Space and Virtual Space entrance initial value, triggers collision accident, hide virtual
Other models in space in addition to entrance;When virtual camera leaves Virtual Space, event is left in triggering, and display is hidden
All models, and Virtual Space entrance initial state value is set to false as.
2. Virtual Space method as described in claim 1, which is characterized in that in step " by the model carry in virtual scene
To Virtual Space origin " further include before step:The threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space
Origin, and using the Virtual Space origin as virtual scene father node.
3. Virtual Space method as claimed in claim 1 or 2, which is characterized in that the ambient is by following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
4. Virtual Space method as claimed in claim 1 or 2, which is characterized in that the second collision object is preset as ensureing it
The size of crash response is triggered when colliding with other collision objects enough.
5. a kind of Virtual Space device, which is characterized in that including model carry unit, ambient setting unit, rendering unit, touch
Hit object setting unit, state value setting unit and hit-treatment unit;
The model carry unit is used for the model carry in virtual scene to Virtual Space origin;
The ambient setting unit is used to that ambient to be arranged along Virtual Space periphery, and the ambient does not occur with Virtual Space
It is overlapped and surrounds the region except virtual scene entrance;The rendering unit is for rendering default object, and rendering unit is to institute
State rendering of the entrance rendering prior to the ambient of Virtual Space;
The collision object setting unit is for being arranged the first collision object and placing it in Virtual Space entrance, the first collision object
Size it is identical as Virtual Space entrance size;The collision object setting unit is additionally operable to the second collision object of setting in virtual camera shooting
Machine;
The state value setting unit is false for Virtual Space entrance initial state value to be arranged, and when virtual camera is from described
When Virtual Space entrance enters Virtual Space, it is true that the state value setting unit, which changes its state value,;
The hit-treatment unit is used to enter Virtual Space when virtual camera and Virtual Space entrance initial value touches for fictitious time
It sends out collision accident and hides other models in Virtual Space in addition to entrance;The hit-treatment unit is additionally operable to virtually to take the photograph
Event is left in triggering when camera leaves Virtual Space, shows all models being hidden, and state value setting unit is by Virtual Space
Entrance initial state value is set to false as.
6. Virtual Space device as claimed in claim 5, which is characterized in that the model carry unit will be in virtual scene
Before model carry to Virtual Space origin, the threedimensional model bottom centre point of Virtual Space entrance is set as Virtual Space original
Point, and using the Virtual Space origin as virtual scene father node.
7. such as Virtual Space device described in claim 5 or 6, which is characterized in that further include ambient processing unit, for pair
The ambient does following processing:
Channel mask is set as 0, and Color Channel is not written into any result;
Only object is rendered when the depth of object is less than or equal to the depth value of itself;
To the rendering of the ambient prior to other opaque articles in addition to the entrance of Virtual Space.
8. such as Virtual Space method described in claim 5 or 6, which is characterized in that the second collision object is preset as ensureing it
The size of crash response is triggered when colliding with other collision objects enough.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810244761.XA CN108805985B (en) | 2018-03-23 | 2018-03-23 | Virtual space method and device |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810244761.XA CN108805985B (en) | 2018-03-23 | 2018-03-23 | Virtual space method and device |
Publications (2)
Publication Number | Publication Date |
---|---|
CN108805985A true CN108805985A (en) | 2018-11-13 |
CN108805985B CN108805985B (en) | 2022-02-15 |
Family
ID=64095320
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810244761.XA Active CN108805985B (en) | 2018-03-23 | 2018-03-23 | Virtual space method and device |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN108805985B (en) |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8405680B1 (en) * | 2010-04-19 | 2013-03-26 | YDreams S.A., A Public Limited Liability Company | Various methods and apparatuses for achieving augmented reality |
US20130155106A1 (en) * | 2011-12-20 | 2013-06-20 | Xerox Corporation | Method and system for coordinating collisions between augmented reality and real reality |
CN103489214A (en) * | 2013-09-10 | 2014-01-01 | 北京邮电大学 | Virtual reality occlusion handling method, based on virtual model pretreatment, in augmented reality system |
CN105389848A (en) * | 2015-11-06 | 2016-03-09 | 网易(杭州)网络有限公司 | Drawing system and method of 3D scene, and terminal |
CN106056663A (en) * | 2016-05-19 | 2016-10-26 | 京东方科技集团股份有限公司 | Rendering method for enhancing reality scene, processing module and reality enhancement glasses |
CN106157359A (en) * | 2015-04-23 | 2016-11-23 | 中国科学院宁波材料技术与工程研究所 | A kind of method for designing of virtual scene experiencing system |
CN106548519A (en) * | 2016-11-04 | 2017-03-29 | 上海玄彩美科网络科技有限公司 | Augmented reality method based on ORB SLAM and the sense of reality of depth camera |
CN106598229A (en) * | 2016-11-11 | 2017-04-26 | 歌尔科技有限公司 | Virtual reality scene generation method and equipment, and virtual reality system |
CN107368188A (en) * | 2017-07-13 | 2017-11-21 | 河北中科恒运软件科技股份有限公司 | The prospect abstracting method and system based on spatial multiplex positioning in mediation reality |
-
2018
- 2018-03-23 CN CN201810244761.XA patent/CN108805985B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8405680B1 (en) * | 2010-04-19 | 2013-03-26 | YDreams S.A., A Public Limited Liability Company | Various methods and apparatuses for achieving augmented reality |
US20130155106A1 (en) * | 2011-12-20 | 2013-06-20 | Xerox Corporation | Method and system for coordinating collisions between augmented reality and real reality |
CN103489214A (en) * | 2013-09-10 | 2014-01-01 | 北京邮电大学 | Virtual reality occlusion handling method, based on virtual model pretreatment, in augmented reality system |
CN106157359A (en) * | 2015-04-23 | 2016-11-23 | 中国科学院宁波材料技术与工程研究所 | A kind of method for designing of virtual scene experiencing system |
CN105389848A (en) * | 2015-11-06 | 2016-03-09 | 网易(杭州)网络有限公司 | Drawing system and method of 3D scene, and terminal |
CN106056663A (en) * | 2016-05-19 | 2016-10-26 | 京东方科技集团股份有限公司 | Rendering method for enhancing reality scene, processing module and reality enhancement glasses |
CN106548519A (en) * | 2016-11-04 | 2017-03-29 | 上海玄彩美科网络科技有限公司 | Augmented reality method based on ORB SLAM and the sense of reality of depth camera |
CN106598229A (en) * | 2016-11-11 | 2017-04-26 | 歌尔科技有限公司 | Virtual reality scene generation method and equipment, and virtual reality system |
CN107368188A (en) * | 2017-07-13 | 2017-11-21 | 河北中科恒运软件科技股份有限公司 | The prospect abstracting method and system based on spatial multiplex positioning in mediation reality |
Non-Patent Citations (2)
Title |
---|
高宇等: ""基于增强现实的虚拟实景空间的研究与实现"", 《小型微型计算机系统》 * |
高超等: ""基于增强现实的虚拟实景空间漫游机制研究与实现"", 《计算机工程与设计》 * |
Also Published As
Publication number | Publication date |
---|---|
CN108805985B (en) | 2022-02-15 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11282264B2 (en) | Virtual reality content display method and apparatus | |
CN102270275B (en) | The method of selecting object and multimedia terminal in virtual environment | |
JP7337104B2 (en) | Model animation multi-plane interaction method, apparatus, device and storage medium by augmented reality | |
WO2016150292A1 (en) | Virtuality-and-reality-combined interactive method and system for merging real environment | |
CN106780421A (en) | Finishing effect methods of exhibiting based on panoramic platform | |
GB2376397A (en) | Virtual or augmented reality | |
KR20160080064A (en) | Virtual sensor in a virtual environment | |
CN106600709A (en) | Decoration information model-based VR virtual decoration method | |
Montero et al. | Designing and implementing interactive and realistic augmented reality experiences | |
JP2022505998A (en) | Augmented reality data presentation methods, devices, electronic devices and storage media | |
CN106296778A (en) | Virtual objects motion control method and device | |
CN109035415B (en) | Virtual model processing method, device, equipment and computer readable storage medium | |
CN108536288A (en) | The treating method and apparatus of interactive operation in virtual reality | |
CN108694073A (en) | Control method, device, equipment and the storage medium of virtual scene | |
CN105184858A (en) | Method for augmented reality mobile terminal | |
CN107291244A (en) | A kind of VR visits system | |
CN114332374A (en) | Virtual display method, equipment and storage medium | |
CN106683193A (en) | Three-dimensional model design method and design device | |
CN114092670A (en) | Virtual reality display method, equipment and storage medium | |
US10272337B2 (en) | Non transitory computer-readable storage medium and method of controlling a computer | |
Chae et al. | Introduction of physics simulation in augmented reality | |
CN108958568A (en) | A kind of display, exchange method and the device of three dimentional graph display mean camber UI | |
CN109407825A (en) | Interactive approach and device based on virtual objects | |
CN108805985A (en) | Virtual Space method and apparatus | |
Cai et al. | An interactive augmented reality system based on LeapMotion and Metaio |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |