CN107145227A - The exchange method and device of virtual reality scenario - Google Patents
The exchange method and device of virtual reality scenario Download PDFInfo
- Publication number
- CN107145227A CN107145227A CN201710263759.2A CN201710263759A CN107145227A CN 107145227 A CN107145227 A CN 107145227A CN 201710263759 A CN201710263759 A CN 201710263759A CN 107145227 A CN107145227 A CN 107145227A
- Authority
- CN
- China
- Prior art keywords
- ray
- virtual reality
- reality scenario
- target object
- final goal
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 51
- 238000001514 detection method Methods 0.000 claims description 74
- 230000002452 interceptive effect Effects 0.000 claims description 53
- 238000012545 processing Methods 0.000 claims description 7
- 230000003993 interaction Effects 0.000 description 42
- 230000000694 effects Effects 0.000 description 17
- 230000006870 function Effects 0.000 description 13
- 238000003860 storage Methods 0.000 description 12
- 210000001699 lower leg Anatomy 0.000 description 10
- 230000008569 process Effects 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 239000011521 glass Substances 0.000 description 7
- 238000005516 engineering process Methods 0.000 description 6
- 239000007787 solid Substances 0.000 description 6
- 238000004891 communication Methods 0.000 description 5
- 230000009471 action Effects 0.000 description 4
- 230000008859 change Effects 0.000 description 4
- 238000003825 pressing Methods 0.000 description 4
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000013473 artificial intelligence Methods 0.000 description 2
- 230000006399 behavior Effects 0.000 description 2
- 230000008878 coupling Effects 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000013507 mapping Methods 0.000 description 2
- 230000007246 mechanism Effects 0.000 description 2
- 239000000203 mixture Substances 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000000149 penetrating effect Effects 0.000 description 2
- 238000002360 preparation method Methods 0.000 description 2
- 230000004044 response Effects 0.000 description 2
- 238000005094 computer simulation Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 210000000697 sensory organ Anatomy 0.000 description 1
- 239000011800 void material Substances 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Processing Or Creating Images (AREA)
Abstract
The invention discloses a kind of exchange method of virtual reality scenario and device.Wherein, this method includes:The initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, wherein, the ray is shown in the virtual reality scenario;The final goal object collided with the ray is detected in the initial target object;The control that ray is chosen described in the final goal object is operated.The present invention is solved can not also realize the technical problem interacted with the object in virtual reality scenario in the prior art.
Description
Technical field
The present invention relates to computer realm, in particular to the exchange method and device of a kind of virtual reality scenario.
Background technology
Nowadays, in order that Consumer's Experience is on the spot in person, the things in three dimensions is really experienced.Increasing application is opened
Hair business begins to focus on the application under virtual reality scenario.Wherein, virtual reality (Virtual Reality, abbreviation VR) technology,
Be using computer simulation produce a three dimensions virtual world there is provided the various sense organ pseudo-experiences of user, held in user
When row is mobile, computer is by complicated three-dimensional computing, to strengthen true telepresenc of the user in three-dimensional world.
At present, interactive mode commonly used in the prior art includes:Mouse is interacted.Wherein, mouse interaction is for any one
The selection mode of program includes:
1) for two-dimensional world, shown object is respectively provided with corresponding position on two-dimentional (abbreviation 2D) screen, passes through
Mouse clicks on above-mentioned object position on 2D screens, it is possible to realize the interaction with the object.
2) three-dimensional world shown for flat-panel screens, a first point A selected on two-dimentional (abbreviation 2D) screen, then
In three-dimensional (abbreviation 3D) space corresponding with 2D screens, the position B that the point A selected is corresponded in 3D Virtual Space is obtained,
Pass through point A and position B, it is possible to it is determined that the object to be interacted, is completed in flat-panel screens to realize by mouse interaction
The interaction of the three-dimensional world of display.That is, participate in interaction in the three-dimensional world that flat-panel screens is shown is still
The object in two dimensional surface space is embodied, in interaction, mouse is still to be handed over the object implementatio8 in two dimensional surface space
Mutually.
However, the mouse interactive mode provided in currently available technology, be mainly used in shown by two dimensional surface space
Object interact, and for the three dimensional object in Virtual Reality scene, do not provide also and a kind of operate simplified interaction
Mode.
For it is above-mentioned the problem of, effective solution is not yet proposed at present.
The content of the invention
The embodiments of the invention provide a kind of exchange method of virtual reality scenario and device, at least to solve prior art
In can not also realize the technical problem interacted with the object in virtual reality scenario.
One side according to embodiments of the present invention there is provided a kind of exchange method of virtual reality scenario, including:In void
The initial target object for allowing to collide with ray is determined in the object shown in plan reality scene, wherein, it is above-mentioned virtual
Above-mentioned ray is shown in reality scene;The final mesh collided with above-mentioned ray is detected in above-mentioned initial target object
Mark object;The control that above-mentioned ray in above-mentioned final goal object is chosen is operated.
Another aspect according to embodiments of the present invention, additionally provides a kind of interactive device of virtual reality scenario, including:Really
Order member, for determining the initial target pair for allowing to collide with ray in the object that is shown in virtual reality scenario
As, wherein, above-mentioned ray is shown in above-mentioned virtual reality scenario;Detection unit, for being examined in above-mentioned initial target object
Measure the final goal object collided with above-mentioned ray;Operating unit, for being penetrated to above-mentioned in above-mentioned final goal object
The control that line is chosen is operated.
In embodiments of the present invention, determine that permission collides with ray in the object shown in virtual reality scenario
Initial target object after, wherein, ray is shown in virtual reality scenario, then detect in above-mentioned initial target object with
The final goal object that ray collides, is operated to the control that ray in final goal object is chosen.So as to realize
By ray collision detection from initial target object, the final goal object to be interacted in virtual reality scenario is obtained,
The purpose that control in the final goal object chosen with reaching to ray is operated, and then overcome can not in correlation technique
The problem of to the object in virtual reality scenario precisely interact.Further, realized and virtual existing by ray collision detection
Object in real field scape is interacted, and also simplify interactive operation, without carrying out complicated friendship using hardware devices such as mouses
Mutual computing, and then reach in the case where ensureing interaction accuracy, improve the effect of interactive efficiency.
Brief description of the drawings
Accompanying drawing described herein is used for providing a further understanding of the present invention, constitutes the part of the application, this hair
Bright schematic description and description is used to explain the present invention, does not constitute inappropriate limitation of the present invention.In the accompanying drawings:
Fig. 1 is the schematic diagram of the hardware environment of the exchange method of virtual reality scenario according to embodiments of the present invention;
Fig. 2 is a kind of flow chart of the exchange method of optional virtual reality scenario according to embodiments of the present invention;
Fig. 3 is a kind of schematic diagram of the exchange method of optional virtual reality scenario according to embodiments of the present invention;
Fig. 4 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 5 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 6 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 7 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 8 is a kind of schematic diagram of the interactive device of optional virtual reality scenario according to embodiments of the present invention;
Fig. 9 is a kind of structured flowchart of the interactive terminal of virtual reality scenario according to embodiments of the present invention.
Embodiment
In order that those skilled in the art more fully understand the present invention program, below in conjunction with the embodiment of the present invention
Accompanying drawing, the technical scheme in the embodiment of the present invention is clearly and completely described, it is clear that described embodiment is only
The embodiment of a part of the invention, rather than whole embodiments.Based on the embodiment in the present invention, ordinary skill people
The every other embodiment that member is obtained under the premise of creative work is not made, should all belong to the model that the present invention is protected
Enclose.
It should be noted that term " first " in description and claims of this specification and above-mentioned accompanying drawing, "
Two " etc. be for distinguishing similar object, without for describing specific order or precedence.It should be appreciated that so using
Data can exchange in the appropriate case, so as to embodiments of the invention described herein can with except illustrating herein or
Order beyond those of description is implemented.In addition, term " comprising " and " having " and their any deformation, it is intended that cover
Lid is non-exclusive to be included, for example, the process, method, system, product or the equipment that contain series of steps or unit are not necessarily limited to
Those steps or unit clearly listed, but may include not list clearly or for these processes, method, product
Or the intrinsic other steps of equipment or unit.
Embodiment 1
There is provided a kind of embodiment of the exchange method of above-mentioned virtual reality scenario in embodiments of the present invention.It is used as one
Optional embodiment is planted, the exchange method of the virtual reality scenario can be, but not limited to be applied to application ring as shown in Figure 1
In border, display needs to be interacted in the equipment 102 for realizing virtual reality (Virtual Reality, abbreviation VR) scene interactivity
Object, the initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, its
In, ray is shown in virtual reality scenario;The final goal pair collided with ray is detected in initial target object
As;Then the control that ray in final goal object is chosen is operated.
In the present embodiment, determine that permission and ray collide in the object shown in virtual reality scenario first
After beginning destination object, wherein, ray is shown in virtual reality scenario, then detected and ray in above-mentioned initial target object
The final goal object collided, is operated to the control that ray in final goal object is chosen.Pass through so as to realize
Ray collision detection obtains the final goal object to be interacted in virtual reality scenario, and ray is chosen most with reaching
The purpose that control in whole destination object is operated, and then overcome can not be to the object in virtual reality scenario in correlation technique
The problem of precisely interact.
Alternatively, in the present embodiment, it is above-mentioned to be used to realize that the equipment of virtual reality scenario interaction include but not limit
In at least one of:VR glasses, and other hardware devices for realizing virtual reality scenario interaction.Above-mentioned simply one kind is shown
Example, the present embodiment does not do any restriction to this.
According to embodiments of the present invention there is provided a kind of exchange method of virtual reality scenario, as shown in Fig. 2 this method bag
Include:
The initial target pair for allowing to collide with ray is determined in S202, the object shown in virtual reality scenario
As, wherein, ray is shown in virtual reality scenario;
S204, detects the final goal object collided with ray in initial target object;
S206, is operated to the control that ray in final goal object is chosen.
Alternatively, in the present embodiment, the exchange method of above-mentioned virtual reality scenario can be, but not limited to be applied to use
Virtual reality device carry out virtual reality interactive application during, for example, above-mentioned virtual reality interactive application can include but
It is not limited to reality-virtualizing game, the broadcasting of virtual reality video display, virtual reality shopping etc..Above-mentioned is only a kind of example, this implementation
Any restriction is not done to this in example.Wherein, the virtual reality device can include but is not limited to VR glasses.Further, above-mentioned friendship
Mutual process can be, but not limited to interact with the object in the plane in virtual reality scenario, for example, as shown in figure 3, dotted line
Virtual reality scenario content shown in VR glasses (equipment 102 of virtual reality scenario interaction) is shown in square frame:Two
Display needs interactive object A to object E in dimensional plane, wherein, the final goal object collided with ray is object E;
In addition, in the present embodiment, above-mentioned interaction can also but be not limited to enter with the three-dimensional character object in virtual reality scenario
Row interaction, for example, showing as shown in figure 4, being shown in dashed rectangle in VR glasses (equipment 102 of virtual reality scenario interaction)
The virtual reality scenario content shown:Display needs interactive object P, object X, object Y and object Z in three dimensions, wherein,
The final goal object collided with ray is object Y.Above-mentioned is only not do any limit to this in a kind of example, the present embodiment
It is fixed.
It should be noted that in the present embodiment, determining permission with penetrating in the object shown in virtual reality scenario
After the initial target object that line collides, wherein, ray is shown in virtual reality scenario, then in above-mentioned initial target object
In detect the final goal object that is collided with ray, the control that ray in final goal object is chosen is grasped
Make.So as to realize by ray collision detection from initial target object, acquisition to be interacted most in virtual reality scenario
Whole destination object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes phase
The problem of to the object in virtual reality scenario can not precisely interact in the technology of pass.Further, ray collision detection is passed through
Realization is interacted with the object in virtual reality scenario, also simplify interactive operation, without using the hardware devices such as mouse
Complicated interacting operation is carried out, and then is reached in the case where ensureing interaction accuracy, the effect of interactive efficiency is improved.
In addition, in the present embodiment, above-mentioned collision detection (also referred to as collision detection (Collision detection)),
Typically refer to a kind of judge the method whether two or more objects produce common factor.It can apply to electronic game and other calculated
In physics, it can also be applied in artificial intelligence.Except being determined for whether two objects have collided, it can also use
In the time (TOI) for calculating impact, and obtain the position of object intersection-type collision.
Alternatively, in the present embodiment, determine that permission occurs with ray in the object shown in virtual reality scenario
The initial target object of collision includes:Determine that the ray type with ray has corresponding relation according to the object type of object
Destination object type, the object of destination object type is defined as the initial target object for allowing to collide with ray.From
And realize using the mapping relations between ray type and object type, determine the initial target for allowing to collide with ray
Object, to avoid object unrelated with ray in virtual reality scenario from participating in interaction deterministic process, reduces the judgement time, Jin Erda
To the effect for simplifying interaction judgment step.
Alternatively, in the present embodiment, the final goal pair collided with ray is detected in initial target object
As including:Each initial target object is traveled through, detects that each initial target object space shared in virtual reality scenario is
It is no to there is the part overlapped with ray;The initial target object that there will be the part overlapped with ray is defined as final goal pair
As.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to
Amount, above-mentioned detection current target object space shared in virtual reality scenario whether there is the part that is overlapped with ray can be with
But it is not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.That is, by virtual reality
Each point position in scape on shared space and ray carries out collision detection respectively, is accurately judged to collide with ray to reach
Final goal object effect.If, in the present embodiment, can be with addition, the multiframe picture shown by virtual reality scenario
But it is not limited to all do collision detection to each frame, to determine whether include above-mentioned ray in space shared in virtual reality scenario
Point in the vector of composition, so as to ensure the accuracy of collision detection.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting
It is dynamic to operate, choose operation etc..Execution event corresponding with aforesaid operations can include but is not limited at least one of:
1) entry event (Enter):Ray starts to point to some object, by the ray for comparing this frame and previous frame
The change for hitting list object, see with the presence or absence of the object that newly adds, the event of the new addition object of triggering;
2) event (Leave) is left:Ray, which leaves, points to some object, by the ray for comparing this frame and previous frame
The change for hitting list object, see with the presence or absence of the object newly deleted, the new event for deleting object of triggering;
3) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list
The event of all objects;
4) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list
There is the event of object;
5) moving event (Move):Ray is moved on each control in final goal object, in triggering interaction list
The event of all objects.
6) other:For some special events, such as drag events of drag operation triggering etc..
It should be noted that above-mentioned control can be, but not limited to the function for a series of pre-productions for constituting interface
Control, for example, button, check box, sliding block, progress bar etc..Above-mentioned is only not do any limit to this in a kind of example, the present embodiment
It is fixed.
Determine that permission is sent out with ray in the embodiment provided by the application, the object shown in virtual reality scenario
After the initial target object of raw collision, wherein, ray is shown in virtual reality scenario, then examined in above-mentioned initial target object
The final goal object collided with ray is measured, the control that ray in final goal object is chosen is operated.From
And realize by ray collision detection from initial target object, obtain the final goal to be interacted in virtual reality scenario
Object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes correlation technique
In can not in virtual reality scenario object carry out precisely interact the problem of.
As a kind of optional scheme, determine that permission is touched with ray in the object shown in virtual reality scenario
The initial target object hit includes:
S1, obtains the object type of object;
S2, determines that the ray type with ray has the destination object type of corresponding relation in object type;
S3, initial target object is defined as in object by type for the object of destination object type.
Alternatively, in the present embodiment, the type of the object to allowing to collide with ray is set to destination object class
Type, is represented with object collision objects, by for the ray type of collision detection laser collision Object tables
Show.Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.
It should be noted that in the present embodiment, the above-mentioned collision mode for being used to carrying out ray collision detection can with but not
It is limited to use the collision bounding box in correlation technique, by the way that the individual being detected is surrounded with collision bounding box, detects that each is touched
Hit whether bounding box has intersection, when producing intersection, detect to produce collision.Above-mentioned is only a kind of example, this reality
Apply in example and the mode of collision detection is not limited, other collision detection modes can also be used.
The embodiment provided by the application, by classifying by type to the object in virtual reality scenario, with reality
Now pair destination object type for having corresponding relation with ray type carries out collision detection, so as to realize for not being destination object
The object of type directly skips over detection, reaches that ray reduces the quantity of the object of collision detection, realization greatlys save collision detection
Time loss effect, and then reach simplify interaction judgment step effect.
As a kind of optional scheme, the final goal object collided with ray is detected in initial target object
Including:
S1, following steps are performed to each object in initial target object, wherein, initial target object is below performing
It is considered as current target object during step:
S11, detection current target object space shared in virtual reality scenario whether there is the portion overlapped with ray
Point;
S12, if space has the part overlapped with ray, final goal object is defined as by current target object.
Alternatively, in the present embodiment, it can be, but not limited to hold the initial target object for being defined as destination object type
Row Ergodic judgement process, wherein, it regard each initial target object as current target object respectively:Detection current target object exists
Shared space whether there is the part overlapped with ray in virtual reality scenario;There will be the initial of the part that is overlapped with ray
Destination object is defined as final goal object.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to
Amount.In the present embodiment, above-mentioned detection current target object space shared in virtual reality scenario whether there is and ray
The part of coincidence can be, but not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.Namely
Say, each point position on space and ray shared in virtual reality scenario is subjected to collision detection respectively, it is accurate to reach
Judge the effect of final goal object collided with ray.Wherein, above-mentioned current target object institute in virtual reality scenario
The space accounted for can be, but not limited to as shown in figure 3, by taking object E as an example, space shared in virtual reality scenario object E is one
Individual rectangle plane region;As shown in figure 4, by taking object Y as an example, space shared in virtual reality scenario object Y is a people
Object angle color Y shared solid spaces in three dimensions.
It is specific to be illustrated with reference to the example below, for example, by taking object E shown in Fig. 3 as an example, if detection emergent ray and object E
Corresponding rectangle plane region, which exists, to intersect, then can detect that object E has the part overlapped with ray, can determine object E
For final goal object.In another example, by taking object Y shown in Fig. 4 as an example, if detection emergent ray character Y institutes corresponding with object Y
The solid space accounted for, which exists, to intersect, then can detect that object Y has the part overlapped with ray, can be defined as object Y finally
Destination object.
The embodiment provided by the application, by traveling through each initial target object, judges that each is initial successively respectively
Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, to realize according to judged result
Accurately detect for interactive final goal object, it is ensured that interactive accuracy.
As a kind of optional scheme, detection current target object space shared in virtual reality scenario whether there is
The part overlapped with ray includes:
Whether S1, detection space includes the point on ray, wherein, ray includes:Starting point and terminal, the direction of ray by
Starting point points to terminal;
S2, if detecting, space includes the point on ray, it is determined that goes out space and there is the part overlapped with ray.
Alternatively, in the present embodiment, above-mentioned ray can be, but not limited to be to include the Directed Vector of starting point and terminal.
Wherein, above-mentioned starting point can be, but not limited to send the position of ray in virtual reality scenario, and terminal is according to ray institute
The position reached after finger direction extension certain distance, each position on this ray is required for carrying out collision detection.
For example, as shown in fig. 5, it is assumed that starting point (such as uses Hand with position of the finger in virtual reality scenario
Position is represented) exemplified by, according to the extended distance (such as being represented with Max Distance) of the ray set in advance, acquisition pair
The terminal (such as being represented with Max Range Position) answered, whether being taken up space according to finger orientation detection object Y, it is upper to include
State the point on ray.For example, as shown in fig. 6, collision detection result is represented:Solid shared by the corresponding character Y of object Y is empty
Between with above-mentioned ray (starting point be Hand Position, terminal be Max Range Position) there is intersection, crosspoint is
Position shown in Fig. 6 rhombuses, then can detect that object Y has the part overlapped with ray, object Y can be defined as into final goal pair
As.
Alternatively, in the present embodiment, terminal is determined by following steps:According to starting point in virtual reality scenario
Position, the direction vector for direction indication, the predetermined distance range value determined, determine terminal in virtual reality scenario
Position.
For example, as shown in fig. 7, the formula of the terminal for determining above-mentioned ray is as follows:
Max Range Position=Hand Position+Hand Forward Vector*Max Distance
Wherein, above-mentioned Hand Forward Vector are used for the unit vector for representing directions of rays.That is in starting
On the three-dimensional coordinate of point, the three-dimensional coordinate along the extended distance in directions of rays is superimposed, so as to obtain the three-dimensional coordinate of terminal.
The embodiment provided by the application, by whether detecting space comprising the point on ray, to realize that detection is current
Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, that is to say, that will virtually show
Each point position in real field scape on shared space and ray carries out collision detection respectively, is accurately judged to and ray with reaching
The effect of the final goal object of collision, so as to ensure the accuracy of collision detection, and then realizes the effect for improving interaction accuracy
Really.
As a kind of optional scheme, the control chosen to ray in final goal object, which carries out operation, to be included:
S1, chooses the control on the position that ray overlaps with final goal object in final goal object;
S2, is operated to control.
Alternatively, in the present embodiment, carrying out operation to control includes:
S21, when detecting interactive button corresponding with virtual reality scenario and being pressed, determines to detect and starts to click on
Event;
S22, when detecting interactive button corresponding with virtual reality scenario and being lifted, determines that detecting end clicks on
Event;
S23, detect start click event and terminate click event when, to control carry out clicking operation.
Alternatively, in the present embodiment, object can include but is not limited in above-mentioned virtual reality scenario:Illusory Dynamic Graph
Object in object (being referred to as UMG objects) designed by shape (Unreal Mobile) UI designers, destination object type.
It should be noted that in the present embodiment, if what is collided with ray is UMG objects, need to call UMG phases
The control logic of pass, for example, in UE4 environment, for UMG objects, by calling Widget On series of functions to trigger
Widget event response, so as to reach purpose of the compatible general game engine for UMG support.That is, virtual
Real content developer can use general UI preparation method to make user interface (User for virtual reality scenario
Interface, abbreviation UI) menu, the menu may be directly applied to virtual reality scenario and interacts.In addition, in this implementation
In example, if what is collided with ray is not UMG objects, but the object in destination object type, then it can refer to this implementation
Interactive mode described in example is interacted.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting
It is dynamic to operate, choose operation etc..Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.For example, to click on behaviour
As example, it can be, but not limited to determine by following event:
1) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list
The event of all objects.The readjustment machine that the On series of functions triggering that widget has been present has been present is directly invoked for UMG
System, for the call back function of other then self-defined events;
2) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list
There is the event of object.The callback mechanism that the On series of functions triggering that widget has been present has been present is directly invoked for UMG,
For the call back function of other then self-defined events.
That is, when detecting interactive button corresponding with virtual reality scenario and being pressed, determining and detecting out
Initial point hits event;When detecting interactive button corresponding with virtual reality scenario and being lifted, determine that detecting end clicks on
Event;Then detect start click event and terminate click event when, confirm to control carry out clicking operation.
For example, by taking Fig. 5 as an example, when selected object Y is final goal object, choosing ray to be overlapped with final goal object
Position (such as object Y right crus of diaphragm), then on the position (such as object Y right crus of diaphragm) of coincidence control perform clicking operation, example
Such as detect that interactive button is pressed start click event and event is clicked in end when interaction button is lifted, then can confirm
Clicking operation is performed to above-mentioned control.So as to realize the interactive controlling of character Y corresponding to object Y right crus of diaphragm, for example, lead to
Cross click right crus of diaphragm and realize that control lifts right crus of diaphragm and performs striking action.Above-mentioned is only this not to be appointed in a kind of example, the present embodiment
What is limited.
The embodiment provided by the application, the position for choosing ray to be overlapped with final goal object in final goal object
The control put, is operated to the control.Completed so as to realize in virtual reality scenario by ray between object
Interaction, to ensure to interact accuracy.
It should be noted that for foregoing each method embodiment, in order to be briefly described, therefore it is all expressed as a series of
Combination of actions, but those skilled in the art should know, the present invention is not limited by described sequence of movement because
According to the present invention, some steps can be carried out sequentially or simultaneously using other.Secondly, those skilled in the art should also know
Know, embodiment described in this description belongs to preferred embodiment, involved action and module is not necessarily of the invention
It is necessary.
Through the above description of the embodiments, those skilled in the art can be understood that according to above-mentioned implementation
The method of example can add the mode of required general hardware platform to realize by software, naturally it is also possible to by hardware, but a lot
In the case of the former be more preferably embodiment.Understood based on such, technical scheme is substantially in other words to existing
The part that technology contributes can be embodied in the form of software product, and the computer software product is stored in a storage
In medium (such as ROM/RAM, magnetic disc, CD), including some instructions are to cause a station terminal equipment (can be mobile phone, calculate
Machine, server, or network equipment etc.) perform method described in each of the invention embodiment.
Embodiment 2
According to embodiments of the present invention, a kind of the virtual of exchange method for being used to implement above-mentioned virtual reality scenario is additionally provided
The interactive device of reality scene.As shown in figure 8, the device can include:
1) determining unit 802, for determining that permission is touched with ray in the object that is shown in virtual reality scenario
The initial target object hit, wherein, ray is shown in virtual reality scenario;
2) detection unit 804, for detecting the final goal object collided with ray in initial target object;
3) operating unit 806, for being operated to the control that ray in final goal object is chosen.
Alternatively, in the present embodiment, the interactive device of above-mentioned virtual reality scenario can be, but not limited to be applied to use
Virtual reality device carry out virtual reality interactive application during, for example, above-mentioned virtual reality interactive application can include but
It is not limited to reality-virtualizing game, the broadcasting of virtual reality video display, virtual reality shopping etc..Above-mentioned is only a kind of example, this implementation
Any restriction is not done to this in example.Wherein, the virtual reality device can include but is not limited to VR glasses.Further, above-mentioned friendship
Mutual process can be, but not limited to interact with the object in the plane in virtual reality scenario, for example, as shown in figure 3, dotted line
Virtual reality scenario content shown in VR glasses (equipment 102 of virtual reality scenario interaction) is shown in square frame:Two
Display needs interactive object A to object E in dimensional plane, wherein, the final goal object collided with ray is object E;
In addition, in the present embodiment, above-mentioned interaction can also but be not limited to enter with the three-dimensional character object in virtual reality scenario
Row interaction, for example, showing as shown in figure 4, being shown in dashed rectangle in VR glasses (equipment 102 of virtual reality scenario interaction)
The virtual reality scenario content shown:Display needs interactive object P, object X, object Y and object Z in three dimensions, wherein,
The final goal object collided with ray is object Y.Above-mentioned is only not do any limit to this in a kind of example, the present embodiment
It is fixed.
It should be noted that in the present embodiment, determining permission with penetrating in the object shown in virtual reality scenario
After the initial target object that line collides, wherein, ray is shown in virtual reality scenario, then in above-mentioned initial target object
In detect the final goal object that is collided with ray, the control that ray in final goal object is chosen is grasped
Make.So as to realize by ray collision detection from initial target object, acquisition to be interacted most in virtual reality scenario
Whole destination object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes phase
The problem of to the object in virtual reality scenario can not precisely interact in the technology of pass.Further, ray collision detection is passed through
Realization is interacted with the object in virtual reality scenario, also simplify interactive operation, without using the hardware devices such as mouse
Complicated interacting operation is carried out, and then is reached in the case where ensureing interaction accuracy, the effect of interactive efficiency is improved.
In addition, in the present embodiment, above-mentioned collision detection (also referred to as collision detection (Collision detection)),
Typically refer to a kind of judge the method whether two or more objects produce common factor.It can apply to electronic game and other calculated
In physics, it can also be applied in artificial intelligence.Except being determined for whether two objects have collided, it can also use
In the time (TOI) for calculating impact, and obtain the position of object intersection-type collision.
Alternatively, in the present embodiment, determine that permission occurs with ray in the object shown in virtual reality scenario
The initial target object of collision includes:Determine that the ray type with ray has corresponding relation according to the object type of object
Destination object type, the object of destination object type is defined as the initial target object for allowing to collide with ray.From
And realize using the mapping relations between ray type and object type, determine the initial target for allowing to collide with ray
Object, to avoid object unrelated with ray in virtual reality scenario from participating in interaction deterministic process, reduces the judgement time, Jin Erda
To the effect for simplifying interaction judgment step.
Alternatively, in the present embodiment, the final goal pair collided with ray is detected in initial target object
As including:Each initial target object is traveled through, detects that each initial target object space shared in virtual reality scenario is
It is no to there is the part overlapped with ray;The initial target object that there will be the part overlapped with ray is defined as final goal pair
As.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to
Amount, above-mentioned detection current target object space shared in virtual reality scenario whether there is the part that is overlapped with ray can be with
But it is not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.That is, by virtual reality
Each point position in scape on shared space and ray carries out collision detection respectively, is accurately judged to collide with ray to reach
Final goal object effect.If, in the present embodiment, can be with addition, the multiframe picture shown by virtual reality scenario
But it is not limited to all do collision detection to each frame, to determine whether include above-mentioned ray in space shared in virtual reality scenario
Point in the vector of composition, so as to ensure the accuracy of collision detection.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting
It is dynamic to operate, choose operation etc..Execution event corresponding with aforesaid operations can include but is not limited at least one of:
1) entry event (Enter):Ray starts to point to some object, by the ray for comparing this frame and previous frame
The change for hitting list object, see with the presence or absence of the object that newly adds, the event of the new addition object of triggering;
2) event (Leave) is left:Ray, which leaves, points to some object, by the ray for comparing this frame and previous frame
The change for hitting list object, see with the presence or absence of the object newly deleted, the new event for deleting object of triggering;
3) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list
The event of all objects;
4) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list
There is the event of object;
5) moving event (Move):Ray is moved on each control in final goal object, in triggering interaction list
The event of all objects.
6) other:For some special events, such as drag events of drag operation triggering etc..
It should be noted that above-mentioned control can be, but not limited to the function for a series of pre-productions for constituting interface
Control, for example, button, check box, sliding block, progress bar etc..Above-mentioned is only not do any limit to this in a kind of example, the present embodiment
It is fixed.
Determine that permission is sent out with ray in the embodiment provided by the application, the object shown in virtual reality scenario
After the initial target object of raw collision, wherein, ray is shown in virtual reality scenario, then examined in above-mentioned initial target object
The final goal object collided with ray is measured, the control that ray in final goal object is chosen is operated.From
And realize by ray collision detection from initial target object, obtain the final goal to be interacted in virtual reality scenario
Object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes correlation technique
In can not in virtual reality scenario object carry out precisely interact the problem of.
As a kind of optional scheme, determining unit 802 includes:
1) acquisition module, the object type for obtaining object;
2) the first determining module, the ray type for being determined in object type with ray has the mesh of corresponding relation
Mark object type;
3) the second determining module, initial target object is defined as in object by type for the object of destination object type.
Alternatively, in the present embodiment, the type of the object to allowing to collide with ray is set to destination object class
Type, is represented with object collision objects, by for the ray type of collision detection laser collision Object tables
Show.Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.
It should be noted that in the present embodiment, the above-mentioned collision mode for being used to carrying out ray collision detection can with but not
It is limited to use the collision bounding box in correlation technique, by the way that the individual being detected is surrounded with collision bounding box, detects that each is touched
Hit whether bounding box has intersection, when producing intersection, detect to produce collision.Above-mentioned is only a kind of example, this reality
Apply in example and the mode of collision detection is not limited, other collision detection modes can also be used.
The embodiment provided by the application, by classifying by type to the object in virtual reality scenario, with reality
Now pair destination object type for having corresponding relation with ray type carries out collision detection, so as to realize for not being destination object
The object of type directly skips over detection, reaches that ray reduces the quantity of the object of collision detection, realization greatlys save collision detection
Time loss effect, and then reach simplify interaction judgment step effect.
As a kind of optional scheme, detection unit 804 includes:
1) processing module, for performing following steps to each object in initial target object, wherein, initial target pair
As being considered as current target object when performing following steps:
S1, detection current target object space shared in virtual reality scenario whether there is the portion overlapped with ray
Point;
S2, if space has the part overlapped with ray, final goal object is defined as by current target object.
Alternatively, in the present embodiment, it can be, but not limited to hold the initial target object for being defined as destination object type
Row Ergodic judgement process, wherein, it regard each initial target object as current target object respectively:Detection current target object exists
Shared space whether there is the part overlapped with ray in virtual reality scenario;There will be the initial of the part that is overlapped with ray
Destination object is defined as final goal object.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to
Amount.In the present embodiment, above-mentioned detection current target object space shared in virtual reality scenario whether there is and ray
The part of coincidence can be, but not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.Namely
Say, each point position on space and ray shared in virtual reality scenario is subjected to collision detection respectively, it is accurate to reach
Judge the effect of final goal object collided with ray.Wherein, above-mentioned current target object institute in virtual reality scenario
The space accounted for can be, but not limited to as shown in figure 3, by taking object E as an example, space shared in virtual reality scenario object E is one
Individual rectangle plane region;As shown in figure 4, by taking object Y as an example, space shared in virtual reality scenario object Y is a people
Object angle color Y shared solid spaces in three dimensions.
It is specific to be illustrated with reference to the example below, for example, by taking object E shown in Fig. 3 as an example, if detection emergent ray and object E
Corresponding rectangle plane region, which exists, to intersect, then can detect that object E has the part overlapped with ray, can determine object E
For final goal object.In another example, by taking object Y shown in Fig. 4 as an example, if detection emergent ray character Y institutes corresponding with object Y
The solid space accounted for, which exists, to intersect, then can detect that object Y has the part overlapped with ray, can be defined as object Y finally
Destination object.
The embodiment provided by the application, by traveling through each initial target object, judges that each is initial successively respectively
Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, to realize according to judged result
Accurately detect for interactive final goal object, it is ensured that interactive accuracy.
As a kind of optional scheme, processing module realizes detection current target object in virtual reality by following steps
Shared space whether there is the part overlapped with ray in scene:
Whether S1, detection space includes the point on ray, wherein, ray includes:Starting point and terminal, the direction of ray by
Starting point points to terminal;
S2, if detecting, space includes the point on ray, it is determined that goes out space and there is the part overlapped with ray.
Alternatively, in the present embodiment, above-mentioned ray can be, but not limited to be to include the Directed Vector of starting point and terminal.
Wherein, above-mentioned starting point can be, but not limited to send the position of ray in virtual reality scenario, and terminal is according to ray institute
The position reached after finger direction extension certain distance, each position on this ray is required for carrying out collision detection.
For example, as shown in fig. 5, it is assumed that starting point (such as uses Hand with position of the finger in virtual reality scenario
Position is represented) exemplified by, according to the extended distance (such as being represented with Max Distance) of the ray set in advance, acquisition pair
The terminal (such as being represented with Max Range Position) answered, whether being taken up space according to finger orientation detection object Y, it is upper to include
State the point on ray.For example, as shown in fig. 6, collision detection result is represented:Solid shared by the corresponding character Y of object Y is empty
Between with above-mentioned ray (starting point be Hand Position, terminal be Max Range Position) there is intersection, crosspoint is
Position shown in Fig. 6 rhombuses, then can detect that object Y has the part overlapped with ray, object Y can be defined as into final goal pair
As.
Alternatively, in the present embodiment, processing module determines terminal by following steps by following steps:According to starting
Position of the point in virtual reality scenario, the direction vector for direction indication, the predetermined distance range value determined, determine end
Position of the point in virtual reality scenario.
For example, as shown in fig. 7, the formula of the terminal for determining above-mentioned ray is as follows:
Max Range Position=Hand Position+Hand Forward Vector*Max Distance
Wherein, above-mentioned Hand Forward Vector are used for the unit vector for representing directions of rays.That is in starting
On the three-dimensional coordinate of point, the three-dimensional coordinate along the extended distance in directions of rays is superimposed, so as to obtain the three-dimensional coordinate of terminal.
The embodiment provided by the application, by whether detecting space comprising the point on ray, to realize that detection is current
Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, that is to say, that will virtually show
Each point position in real field scape on shared space and ray carries out collision detection respectively, is accurately judged to and ray with reaching
The effect of the final goal object of collision, so as to ensure the accuracy of collision detection, and then realizes the effect for improving interaction accuracy
Really.
As a kind of optional scheme, operating unit 806 includes:
1) module is chosen, for choosing the control on the position that ray overlaps with final goal object in final goal object
Part;
2) operation module, for being operated to control.
Alternatively, in the present embodiment, operation module includes:
(1) first determination sub-module, for when detecting interactive button corresponding with virtual reality scenario and being pressed, really
Make to detect and start click event;
(2) second determination sub-modules, for when detecting interactive button corresponding with virtual reality scenario and being lifted, really
Make and detect end click event;
(3) the 3rd determination sub-modules, for detect start click event and terminate click event when, to control carry out
Clicking operation.
Alternatively, in the present embodiment, object can include but is not limited in above-mentioned virtual reality scenario:Illusory Dynamic Graph
Object in object (being referred to as UMG objects) designed by shape (Unreal Mobile) UI designers, destination object type.
It should be noted that in the present embodiment, if what is collided with ray is UMG objects, need to call UMG phases
The control logic of pass, for example, in UE4 environment, for UMG objects, by calling Widget On series of functions to trigger
Widget event response, so as to reach purpose of the compatible general game engine for UMG support.That is, virtual
Real content developer can use general UI preparation method to make user interface (User for virtual reality scenario
Interface, abbreviation UI) menu, the menu may be directly applied to virtual reality scenario and interacts.In addition, in this implementation
In example, if what is collided with ray is not UMG objects, but the object in destination object type, then it can refer to this implementation
Interactive mode described in example is interacted.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting
It is dynamic to operate, choose operation etc..Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.For example, to click on behaviour
As example, it can be, but not limited to determine by following event:
1) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list
The event of all objects.The readjustment machine that the On series of functions triggering that widget has been present has been present is directly invoked for UMG
System, for the call back function of other then self-defined events;
2) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list
There is the event of object.The callback mechanism that the On series of functions triggering that widget has been present has been present is directly invoked for UMG,
For the call back function of other then self-defined events.
That is, when detecting interactive button corresponding with virtual reality scenario and being pressed, determining and detecting out
Initial point hits event;When detecting interactive button corresponding with virtual reality scenario and being lifted, determine that detecting end clicks on
Event;Then detect start click event and terminate click event when, confirm to control carry out clicking operation.
For example, by taking Fig. 5 as an example, when selected object Y is final goal object, choosing ray to be overlapped with final goal object
Position (such as object Y right crus of diaphragm), then on the position (such as object Y right crus of diaphragm) of coincidence control perform clicking operation, example
Such as detect that interactive button is pressed start click event and event is clicked in end when interaction button is lifted, then can confirm
Clicking operation is performed to above-mentioned control.So as to realize the interactive controlling of character Y corresponding to object Y right crus of diaphragm, for example, lead to
Cross click right crus of diaphragm and realize that control lifts right crus of diaphragm and performs striking action.Above-mentioned is only this not to be appointed in a kind of example, the present embodiment
What is limited.
The embodiment provided by the application, the position for choosing ray to be overlapped with final goal object in final goal object
The control put, is operated to the control.Completed so as to realize in virtual reality scenario by ray between object
Interaction, to ensure to interact accuracy.
Herein it should be noted that above-mentioned module is identical with example and application scenarios that the step of correspondence is realized, but not
It is limited to the disclosure of that of above-described embodiment 1.It should be noted that above-mentioned module as a part for device may operate in as
It in hardware environment shown in Fig. 1, can be realized, can also be realized by hardware by software, wherein, hardware environment includes network
Environment.
Embodiment 3
According to embodiments of the present invention, a kind of the virtual of exchange method for being used to implement above-mentioned virtual reality scenario is additionally provided
The interactive terminal of reality scene, as shown in figure 9, the terminal includes:
1) communication interface 902, are set to obtain ray shown in virtual reality scenario;
2) processor 904, are connected with communication interface 902, are set to determine in the object that shows in virtual reality scenario
Go out to allow the initial target object collided with ray, wherein, ray is shown in virtual reality scenario;It is also configured to first
The final goal object collided with ray is detected in beginning destination object;It is also configured to ray institute in final goal object
The control chosen is operated;
3) display 906, are connected with processor 904, are set to real ray, initial target object and final goal pair
As;
4) memory 908, are connected with communication interface 902, processor 904 and display 906, are set to storage above-mentioned initial
Destination object, final goal object, and ray position information.
Alternatively, the specific example in the present embodiment may be referred to showing described in above-described embodiment 1 and embodiment 2
Example, the present embodiment will not be repeated here.
One of ordinary skill in the art will appreciate that all or part of step in the various methods of above-described embodiment is can
To be completed by program come the device-dependent hardware of command terminal, the program can be stored in a computer-readable recording medium
In, storage medium can include:Flash disk, read-only storage (Read-Only Memory, ROM), random access device (Random
Access Memory, RAM), disk or CD etc..
Embodiment 4
Embodiments of the invention additionally provide a kind of storage medium.Alternatively, in the present embodiment, above-mentioned storage medium can
With at least one network equipment in multiple network equipments in the network shown in above-described embodiment.
Alternatively, in the present embodiment, storage medium is arranged to the program code that storage is used to perform following steps:
The initial target pair for allowing to collide with ray is determined in S1, the object shown in virtual reality scenario
As, wherein, ray is shown in virtual reality scenario;
S2, detects the final goal object collided with ray in initial target object;
S3, is operated to the control that ray in final goal object is chosen.
Alternatively, the specific example in the present embodiment may be referred to showing described in above-described embodiment 1 and embodiment 2
Example, the present embodiment will not be repeated here.
Alternatively, in the present embodiment, above-mentioned storage medium can include but is not limited to:USB flash disk, read-only storage (ROM,
Read-Only Memory), random access memory (RAM, Random Access Memory), mobile hard disk, magnetic disc or
CD etc. is various can be with the medium of store program codes.
The embodiments of the present invention are for illustration only, and the quality of embodiment is not represented.
If the integrated unit in above-described embodiment is realized using in the form of SFU software functional unit and is used as independent product
Sale or in use, the storage medium that above computer can be read can be stored in.Understood based on such, skill of the invention
The part or all or part of the technical scheme that art scheme substantially contributes to prior art in other words can be with soft
The form of part product is embodied, and the computer software product is stored in storage medium, including some instructions are to cause one
Platform or multiple stage computers equipment (can be personal computer, server or network equipment etc.) perform each embodiment institute of the invention
State all or part of step of method.
In the above embodiment of the present invention, the description to each embodiment all emphasizes particularly on different fields, and does not have in some embodiment
The part of detailed description, may refer to the associated description of other embodiment.
, can be by others side in several embodiments provided herein, it should be understood that disclosed client
Formula is realized.Wherein, device embodiment described above is only schematical, such as division of described unit, only one
Kind of division of logic function, can there is other dividing mode when actually realizing, such as multiple units or component can combine or
Another system is desirably integrated into, or some features can be ignored, or do not perform.It is another, it is shown or discussed it is mutual it
Between coupling or direct-coupling or communication connection can be the INDIRECT COUPLING or communication link of unit or module by some interfaces
Connect, can be electrical or other forms.
The unit illustrated as separating component can be or may not be it is physically separate, it is aobvious as unit
The part shown can be or may not be physical location, you can with positioned at a place, or can also be distributed to multiple
On NE.Some or all of unit therein can be selected to realize the mesh of this embodiment scheme according to the actual needs
's.
In addition, each functional unit in each embodiment of the invention can be integrated in a processing unit, can also
That unit is individually physically present, can also two or more units it is integrated in a unit.Above-mentioned integrated list
Member can both be realized in the form of hardware, it would however also be possible to employ the form of SFU software functional unit is realized.
Described above is only the preferred embodiment of the present invention, it is noted that for the ordinary skill people of the art
For member, under the premise without departing from the principles of the invention, some improvements and modifications can also be made, these improvements and modifications also should
It is considered as protection scope of the present invention.
Claims (14)
1. a kind of exchange method of virtual reality scenario, it is characterised in that including:
The initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, wherein,
The ray is shown in the virtual reality scenario;
The final goal object collided with the ray is detected in the initial target object;
The control that ray is chosen described in the final goal object is operated.
2. according to the method described in claim 1, it is characterised in that determined in the object shown in virtual reality scenario fair
Perhaps the initial target object collided with ray includes:
Obtain the object type of the object;
Determine that the ray type with the ray has the destination object type of corresponding relation in the object type;
Type is defined as the initial target object for the object of the destination object type in the object.
3. according to the method described in claim 1, it is characterised in that detected in the initial target object and the ray
The final goal object collided includes:
To in the initial target object each object perform following steps, wherein, the initial target object perform with
It is considered as current target object during lower step:
Detect that current target object space shared in the virtual reality scenario whether there is to overlap with the ray
Part;
If the space has the part overlapped with the ray, the current target object is defined as the final goal
Object.
4. method according to claim 3, it is characterised in that the detection current target object is in the virtual reality
Shared space whether there is the part overlapped with the ray and include in scape:
The space is detected whether comprising the point on the ray, wherein, the ray includes:Starting point and terminal, it is described to penetrate
The direction of line the terminal is pointed to by the starting point;
If detecting, the space includes the point on the ray, it is determined that goes out the space and there is the portion overlapped with the ray
Point.
5. method according to claim 4, it is characterised in that determine the terminal by following steps:
According to position of the starting point in the virtual reality scenario, the direction vector for indicating the direction, make a reservation for
The distance range value of determination, determines position of the terminal in the virtual reality scenario.
6. method according to any one of claim 1 to 5, it is characterised in that described in the final goal object
The control that ray is chosen, which carries out operation, to be included:
The control on the position that the ray overlaps with the final goal object is chosen in the final goal object;
The control is operated.
7. method according to claim 6, it is characterised in that carrying out operation to the control includes:
When detecting interactive button corresponding with the virtual reality scenario and being pressed, determine to detect and start to click on thing
Part;
When detecting interactive button corresponding with the virtual reality scenario and being lifted, determine that detecting end clicks on thing
Part;
Detect it is described start click event and the end and click on event when, clicking operation is carried out to the control.
8. a kind of interactive device of virtual reality scenario, it is characterised in that including:
Determining unit, for determining the initial mesh for allowing to collide with ray in the object that is shown in virtual reality scenario
Object is marked, wherein, the ray is shown in the virtual reality scenario;
Detection unit, for the final goal object for detecting to collide with the ray in the initial target object;
Operating unit, for being operated to the control that ray is chosen described in the final goal object.
9. device according to claim 8, it is characterised in that the determining unit includes:
Acquisition module, the object type for obtaining the object;
First determining module, for determining that the ray type with the ray has corresponding relation in the object type
Destination object type;
Second determining module, the initial target is defined as in the object by type for the object of the destination object type
Object.
10. device according to claim 8, it is characterised in that the detection unit includes:
Processing module, for performing following steps to each object in the initial target object, wherein, the initial target
Object is considered as current target object when performing following steps:
Detect that current target object space shared in the virtual reality scenario whether there is to overlap with the ray
Part;
If the space has the part overlapped with the ray, the current target object is defined as the final goal
Object.
11. device according to claim 10, it is characterised in that the processing module realizes detection institute by following steps
State current target object space shared in the virtual reality scenario and whether there is the part overlapped with the ray:
The space is detected whether comprising the point on the ray, wherein, the ray includes:Starting point and terminal, it is described to penetrate
The direction of line the terminal is pointed to by the starting point;
If detecting, the space includes the point on the ray, it is determined that goes out the space and there is the portion overlapped with the ray
Point.
12. device according to claim 11, it is characterised in that the processing module passes through following step by following steps
Suddenly the terminal is determined:
According to position of the starting point in the virtual reality scenario, the direction vector for indicating the direction, make a reservation for
The distance range value of determination, determines position of the terminal in the virtual reality scenario.
13. the device according to any one of claim 8 to 12, it is characterised in that the operating unit includes:
Module is chosen, for the position for choosing the ray to be overlapped with the final goal object in the final goal object
On control;
Operation module, for being operated to the control.
14. device according to claim 13, it is characterised in that the operation module includes:
First determination sub-module, for when detecting interactive button corresponding with the virtual reality scenario and being pressed, it is determined that
Go out to detect and start click event;
Second determination sub-module, for when detecting interactive button corresponding with the virtual reality scenario and being lifted, it is determined that
Go out to detect end click event;
3rd determination sub-module, for detect it is described start click event and the end and click on event when, to the control
Part carries out clicking operation.
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710263759.2A CN107145227B (en) | 2017-04-20 | 2017-04-20 | The exchange method and device of virtual reality scenario |
PCT/CN2018/082512 WO2018192394A1 (en) | 2017-04-20 | 2018-04-10 | Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710263759.2A CN107145227B (en) | 2017-04-20 | 2017-04-20 | The exchange method and device of virtual reality scenario |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107145227A true CN107145227A (en) | 2017-09-08 |
CN107145227B CN107145227B (en) | 2019-01-08 |
Family
ID=59775036
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710263759.2A Active CN107145227B (en) | 2017-04-20 | 2017-04-20 | The exchange method and device of virtual reality scenario |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN107145227B (en) |
WO (1) | WO2018192394A1 (en) |
Cited By (20)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107704397A (en) * | 2017-10-30 | 2018-02-16 | 歌尔科技有限公司 | Applied program testing method, device and electronic equipment |
CN108052253A (en) * | 2017-12-28 | 2018-05-18 | 灵图互动(武汉)科技有限公司 | A kind of virtual reality display method for manufacturing contents |
CN108536288A (en) * | 2018-03-27 | 2018-09-14 | 网易(杭州)网络有限公司 | The treating method and apparatus of interactive operation in virtual reality |
WO2018192394A1 (en) * | 2017-04-20 | 2018-10-25 | 腾讯科技(深圳)有限公司 | Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus |
CN108762617A (en) * | 2018-05-31 | 2018-11-06 | 苏州蜗牛数字科技股份有限公司 | UI interactive systems, method and storage medium in a kind of VR environment |
CN109407945A (en) * | 2018-11-08 | 2019-03-01 | 山东数字人科技股份有限公司 | A kind of 3 d mosaics implementation method and 3 d mosaics system based on VR |
CN109725782A (en) * | 2017-10-27 | 2019-05-07 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus that realizing virtual reality and smart machine, storage medium |
CN109782909A (en) * | 2018-12-29 | 2019-05-21 | 北京诺亦腾科技有限公司 | A kind of exchange method and device of VR interactive device and VR scene |
CN110032296A (en) * | 2019-04-10 | 2019-07-19 | 网易(杭州)网络有限公司 | Determination method, apparatus, terminal and the storage medium of virtual objects in terminal |
CN110262729A (en) * | 2019-05-20 | 2019-09-20 | 联想(上海)信息技术有限公司 | A kind of object processing method and equipment |
CN110969687A (en) * | 2019-11-29 | 2020-04-07 | 中国商用飞机有限责任公司北京民用飞机技术研究中心 | Collision detection method, device, equipment and medium |
CN110990098A (en) * | 2019-11-29 | 2020-04-10 | 北京新势界科技有限公司 | Method and device for overlapping AR information points triggered by focus |
CN111054066A (en) * | 2019-12-17 | 2020-04-24 | 腾讯科技(深圳)有限公司 | Operation method and device of virtual prop, storage medium and electronic device |
CN111192354A (en) * | 2020-01-02 | 2020-05-22 | 武汉瑞莱保能源技术有限公司 | Three-dimensional simulation method and system based on virtual reality |
CN111569424A (en) * | 2020-05-14 | 2020-08-25 | 北京代码乾坤科技有限公司 | Physical settlement method and device in game scene |
CN111580669A (en) * | 2020-05-12 | 2020-08-25 | 南京睿悦信息技术有限公司 | Interaction method and device for virtual reality and augmented reality mobile end plane application |
CN111803930A (en) * | 2020-07-20 | 2020-10-23 | 网易(杭州)网络有限公司 | Multi-platform interaction method and device and electronic equipment |
CN112107854A (en) * | 2020-09-01 | 2020-12-22 | 苏州紫焰网络科技有限公司 | Game task guide completion method, system and equipment |
CN114419287A (en) * | 2021-12-28 | 2022-04-29 | 北京光启元数字科技有限公司 | Three-dimensional scene processing method, device, equipment and medium |
CN114578972A (en) * | 2022-05-05 | 2022-06-03 | 江西科骏实业有限公司 | Trigger method and system for compatible plane and curved surface UI (user interface) event in VR (virtual reality) scene |
Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105892632A (en) * | 2015-11-16 | 2016-08-24 | 乐视致新电子科技(天津)有限公司 | Method and device for judging the selection of UI (User Interface) widgets of virtual reality application |
CN105913497A (en) * | 2016-05-27 | 2016-08-31 | 杭州映墨科技有限公司 | Virtual reality space mobile positioning system and virtual reality space mobile positioning method for virtual house inspecting |
CN106445118A (en) * | 2016-09-06 | 2017-02-22 | 网易(杭州)网络有限公司 | Virtual reality interaction method and apparatus |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101441542B (en) * | 2008-11-21 | 2011-05-11 | 广东威创视讯科技股份有限公司 | Method and apparatus for recognizing multiple target objects by interactive input apparatus |
CN101770324B (en) * | 2008-12-31 | 2011-12-28 | 商泰软件(上海)有限公司 | Method for realizing interactive operation of 3D graphical interface |
CN106527887B (en) * | 2016-10-18 | 2018-09-04 | 腾讯科技(深圳)有限公司 | Dummy object choosing method, device and VR systems |
CN107145227B (en) * | 2017-04-20 | 2019-01-08 | 腾讯科技(深圳)有限公司 | The exchange method and device of virtual reality scenario |
-
2017
- 2017-04-20 CN CN201710263759.2A patent/CN107145227B/en active Active
-
2018
- 2018-04-10 WO PCT/CN2018/082512 patent/WO2018192394A1/en active Application Filing
Patent Citations (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105892632A (en) * | 2015-11-16 | 2016-08-24 | 乐视致新电子科技(天津)有限公司 | Method and device for judging the selection of UI (User Interface) widgets of virtual reality application |
CN105913497A (en) * | 2016-05-27 | 2016-08-31 | 杭州映墨科技有限公司 | Virtual reality space mobile positioning system and virtual reality space mobile positioning method for virtual house inspecting |
CN106445118A (en) * | 2016-09-06 | 2017-02-22 | 网易(杭州)网络有限公司 | Virtual reality interaction method and apparatus |
Cited By (31)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2018192394A1 (en) * | 2017-04-20 | 2018-10-25 | 腾讯科技(深圳)有限公司 | Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus |
CN109725782B (en) * | 2017-10-27 | 2021-08-24 | 腾讯科技(深圳)有限公司 | Method and device for realizing virtual reality, intelligent equipment and storage medium |
CN109725782A (en) * | 2017-10-27 | 2019-05-07 | 腾讯科技(深圳)有限公司 | A kind of method, apparatus that realizing virtual reality and smart machine, storage medium |
CN107704397A (en) * | 2017-10-30 | 2018-02-16 | 歌尔科技有限公司 | Applied program testing method, device and electronic equipment |
CN107704397B (en) * | 2017-10-30 | 2020-10-09 | 歌尔科技有限公司 | Application program testing method and device and electronic equipment |
CN108052253A (en) * | 2017-12-28 | 2018-05-18 | 灵图互动(武汉)科技有限公司 | A kind of virtual reality display method for manufacturing contents |
CN108052253B (en) * | 2017-12-28 | 2020-09-25 | 灵图互动(武汉)科技有限公司 | Virtual reality display content manufacturing method |
CN108536288A (en) * | 2018-03-27 | 2018-09-14 | 网易(杭州)网络有限公司 | The treating method and apparatus of interactive operation in virtual reality |
CN108762617A (en) * | 2018-05-31 | 2018-11-06 | 苏州蜗牛数字科技股份有限公司 | UI interactive systems, method and storage medium in a kind of VR environment |
CN109407945B (en) * | 2018-11-08 | 2021-08-27 | 山东数字人科技股份有限公司 | Three-dimensional jigsaw realization method based on VR and three-dimensional jigsaw system |
CN109407945A (en) * | 2018-11-08 | 2019-03-01 | 山东数字人科技股份有限公司 | A kind of 3 d mosaics implementation method and 3 d mosaics system based on VR |
CN109782909A (en) * | 2018-12-29 | 2019-05-21 | 北京诺亦腾科技有限公司 | A kind of exchange method and device of VR interactive device and VR scene |
CN109782909B (en) * | 2018-12-29 | 2020-10-30 | 北京诺亦腾科技有限公司 | Interaction method and device for VR interaction equipment and VR scene |
CN110032296B (en) * | 2019-04-10 | 2022-06-03 | 网易(杭州)网络有限公司 | Method and device for determining virtual object in terminal, terminal and storage medium |
CN110032296A (en) * | 2019-04-10 | 2019-07-19 | 网易(杭州)网络有限公司 | Determination method, apparatus, terminal and the storage medium of virtual objects in terminal |
CN110262729B (en) * | 2019-05-20 | 2021-11-16 | 联想(上海)信息技术有限公司 | Object processing method and device |
CN110262729A (en) * | 2019-05-20 | 2019-09-20 | 联想(上海)信息技术有限公司 | A kind of object processing method and equipment |
CN110969687A (en) * | 2019-11-29 | 2020-04-07 | 中国商用飞机有限责任公司北京民用飞机技术研究中心 | Collision detection method, device, equipment and medium |
CN110990098A (en) * | 2019-11-29 | 2020-04-10 | 北京新势界科技有限公司 | Method and device for overlapping AR information points triggered by focus |
CN110969687B (en) * | 2019-11-29 | 2023-07-28 | 中国商用飞机有限责任公司北京民用飞机技术研究中心 | Collision detection method, device, equipment and medium |
CN111054066B (en) * | 2019-12-17 | 2021-01-15 | 腾讯科技(深圳)有限公司 | Operation method and device of virtual prop, storage medium and electronic device |
CN111054066A (en) * | 2019-12-17 | 2020-04-24 | 腾讯科技(深圳)有限公司 | Operation method and device of virtual prop, storage medium and electronic device |
CN111192354A (en) * | 2020-01-02 | 2020-05-22 | 武汉瑞莱保能源技术有限公司 | Three-dimensional simulation method and system based on virtual reality |
CN111580669A (en) * | 2020-05-12 | 2020-08-25 | 南京睿悦信息技术有限公司 | Interaction method and device for virtual reality and augmented reality mobile end plane application |
CN111569424A (en) * | 2020-05-14 | 2020-08-25 | 北京代码乾坤科技有限公司 | Physical settlement method and device in game scene |
CN111569424B (en) * | 2020-05-14 | 2023-04-28 | 北京代码乾坤科技有限公司 | Physical settlement method and device in game scene |
CN111803930A (en) * | 2020-07-20 | 2020-10-23 | 网易(杭州)网络有限公司 | Multi-platform interaction method and device and electronic equipment |
CN112107854A (en) * | 2020-09-01 | 2020-12-22 | 苏州紫焰网络科技有限公司 | Game task guide completion method, system and equipment |
CN112107854B (en) * | 2020-09-01 | 2021-03-16 | 苏州紫焰网络科技有限公司 | Game task guide completion method, system and equipment |
CN114419287A (en) * | 2021-12-28 | 2022-04-29 | 北京光启元数字科技有限公司 | Three-dimensional scene processing method, device, equipment and medium |
CN114578972A (en) * | 2022-05-05 | 2022-06-03 | 江西科骏实业有限公司 | Trigger method and system for compatible plane and curved surface UI (user interface) event in VR (virtual reality) scene |
Also Published As
Publication number | Publication date |
---|---|
WO2018192394A1 (en) | 2018-10-25 |
CN107145227B (en) | 2019-01-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107145227A (en) | The exchange method and device of virtual reality scenario | |
KR101863041B1 (en) | Creation of playable scene with an authoring system | |
CN107024981A (en) | Exchange method and device based on virtual reality | |
US20190236259A1 (en) | Method for 3d graphical authentication on electronic devices | |
CN108628533A (en) | Three-dimensional graphical user interface | |
CN110052021B (en) | Game object processing method, mobile terminal device, electronic device, and storage medium | |
CN102096548A (en) | Method and system for duplicating an object using a touch-sensitive display | |
WO2013106169A1 (en) | Menu selection using tangible interaction with mobile devices | |
CN106790828A (en) | The control method and mobile terminal of a kind of mobile terminal | |
RU2667720C1 (en) | Method of imitation modeling and controlling virtual sphere in mobile device | |
CN109701265A (en) | Remote desktop cursor mode switching method, system, client and remote server | |
CN112150602A (en) | Model image rendering method and device, storage medium and electronic equipment | |
CN110196654A (en) | The choosing method and device of virtual resource | |
CN114931746B (en) | Interaction method, device and medium for 3D game based on pen type and touch screen interaction | |
CN109806591A (en) | Signature generating method and device in scene of game | |
CN109701271A (en) | Show the methods, devices and systems of icon | |
CN108089713A (en) | A kind of interior decoration method based on virtual reality technology | |
CN108170338A (en) | Information processing method, device, electronic equipment and storage medium | |
CN113190142B (en) | Cubic model-based 3D environment interaction method and device | |
CN101833421A (en) | Electronic device and method for acquiring user operation | |
CN109375866B (en) | Screen touch click response method and system for realizing same | |
CN113244611A (en) | Virtual article processing method, device, equipment and storage medium | |
CN106547339A (en) | The control method and device of computer equipment | |
CN109089040A (en) | Image processing method, image processing apparatus and terminal device | |
CN108932098A (en) | Building demenstration method and device |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
TR01 | Transfer of patent right |
Effective date of registration: 20240102 Address after: 518000 Tencent Building, No. 1 High-tech Zone, Nanshan District, Shenzhen City, Guangdong Province, 35 Floors Patentee after: TENCENT TECHNOLOGY (SHENZHEN) Co.,Ltd. Patentee after: TENCENT CLOUD COMPUTING (BEIJING) Co.,Ltd. Address before: 518000 Tencent Building, No. 1 High-tech Zone, Nanshan District, Shenzhen City, Guangdong Province, 35 Floors Patentee before: TENCENT TECHNOLOGY (SHENZHEN) Co.,Ltd. |
|
TR01 | Transfer of patent right |