CN107145227A - The exchange method and device of virtual reality scenario - Google Patents

The exchange method and device of virtual reality scenario Download PDF

Info

Publication number
CN107145227A
CN107145227A CN201710263759.2A CN201710263759A CN107145227A CN 107145227 A CN107145227 A CN 107145227A CN 201710263759 A CN201710263759 A CN 201710263759A CN 107145227 A CN107145227 A CN 107145227A
Authority
CN
China
Prior art keywords
ray
virtual reality
reality scenario
target object
final goal
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201710263759.2A
Other languages
Chinese (zh)
Other versions
CN107145227B (en
Inventor
沈超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tencent Technology Shenzhen Co Ltd
Tencent Cloud Computing Beijing Co Ltd
Original Assignee
Tencent Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tencent Technology Shenzhen Co Ltd filed Critical Tencent Technology Shenzhen Co Ltd
Priority to CN201710263759.2A priority Critical patent/CN107145227B/en
Publication of CN107145227A publication Critical patent/CN107145227A/en
Priority to PCT/CN2018/082512 priority patent/WO2018192394A1/en
Application granted granted Critical
Publication of CN107145227B publication Critical patent/CN107145227B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/012Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment

Abstract

The invention discloses a kind of exchange method of virtual reality scenario and device.Wherein, this method includes:The initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, wherein, the ray is shown in the virtual reality scenario;The final goal object collided with the ray is detected in the initial target object;The control that ray is chosen described in the final goal object is operated.The present invention is solved can not also realize the technical problem interacted with the object in virtual reality scenario in the prior art.

Description

The exchange method and device of virtual reality scenario
Technical field
The present invention relates to computer realm, in particular to the exchange method and device of a kind of virtual reality scenario.
Background technology
Nowadays, in order that Consumer's Experience is on the spot in person, the things in three dimensions is really experienced.Increasing application is opened Hair business begins to focus on the application under virtual reality scenario.Wherein, virtual reality (Virtual Reality, abbreviation VR) technology, Be using computer simulation produce a three dimensions virtual world there is provided the various sense organ pseudo-experiences of user, held in user When row is mobile, computer is by complicated three-dimensional computing, to strengthen true telepresenc of the user in three-dimensional world.
At present, interactive mode commonly used in the prior art includes:Mouse is interacted.Wherein, mouse interaction is for any one The selection mode of program includes:
1) for two-dimensional world, shown object is respectively provided with corresponding position on two-dimentional (abbreviation 2D) screen, passes through Mouse clicks on above-mentioned object position on 2D screens, it is possible to realize the interaction with the object.
2) three-dimensional world shown for flat-panel screens, a first point A selected on two-dimentional (abbreviation 2D) screen, then In three-dimensional (abbreviation 3D) space corresponding with 2D screens, the position B that the point A selected is corresponded in 3D Virtual Space is obtained, Pass through point A and position B, it is possible to it is determined that the object to be interacted, is completed in flat-panel screens to realize by mouse interaction The interaction of the three-dimensional world of display.That is, participate in interaction in the three-dimensional world that flat-panel screens is shown is still The object in two dimensional surface space is embodied, in interaction, mouse is still to be handed over the object implementatio8 in two dimensional surface space Mutually.
However, the mouse interactive mode provided in currently available technology, be mainly used in shown by two dimensional surface space Object interact, and for the three dimensional object in Virtual Reality scene, do not provide also and a kind of operate simplified interaction Mode.
For it is above-mentioned the problem of, effective solution is not yet proposed at present.
The content of the invention
The embodiments of the invention provide a kind of exchange method of virtual reality scenario and device, at least to solve prior art In can not also realize the technical problem interacted with the object in virtual reality scenario.
One side according to embodiments of the present invention there is provided a kind of exchange method of virtual reality scenario, including:In void The initial target object for allowing to collide with ray is determined in the object shown in plan reality scene, wherein, it is above-mentioned virtual Above-mentioned ray is shown in reality scene;The final mesh collided with above-mentioned ray is detected in above-mentioned initial target object Mark object;The control that above-mentioned ray in above-mentioned final goal object is chosen is operated.
Another aspect according to embodiments of the present invention, additionally provides a kind of interactive device of virtual reality scenario, including:Really Order member, for determining the initial target pair for allowing to collide with ray in the object that is shown in virtual reality scenario As, wherein, above-mentioned ray is shown in above-mentioned virtual reality scenario;Detection unit, for being examined in above-mentioned initial target object Measure the final goal object collided with above-mentioned ray;Operating unit, for being penetrated to above-mentioned in above-mentioned final goal object The control that line is chosen is operated.
In embodiments of the present invention, determine that permission collides with ray in the object shown in virtual reality scenario Initial target object after, wherein, ray is shown in virtual reality scenario, then detect in above-mentioned initial target object with The final goal object that ray collides, is operated to the control that ray in final goal object is chosen.So as to realize By ray collision detection from initial target object, the final goal object to be interacted in virtual reality scenario is obtained, The purpose that control in the final goal object chosen with reaching to ray is operated, and then overcome can not in correlation technique The problem of to the object in virtual reality scenario precisely interact.Further, realized and virtual existing by ray collision detection Object in real field scape is interacted, and also simplify interactive operation, without carrying out complicated friendship using hardware devices such as mouses Mutual computing, and then reach in the case where ensureing interaction accuracy, improve the effect of interactive efficiency.
Brief description of the drawings
Accompanying drawing described herein is used for providing a further understanding of the present invention, constitutes the part of the application, this hair Bright schematic description and description is used to explain the present invention, does not constitute inappropriate limitation of the present invention.In the accompanying drawings:
Fig. 1 is the schematic diagram of the hardware environment of the exchange method of virtual reality scenario according to embodiments of the present invention;
Fig. 2 is a kind of flow chart of the exchange method of optional virtual reality scenario according to embodiments of the present invention;
Fig. 3 is a kind of schematic diagram of the exchange method of optional virtual reality scenario according to embodiments of the present invention;
Fig. 4 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 5 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 6 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 7 is the schematic diagram of the exchange method of another optional virtual reality scenario according to embodiments of the present invention;
Fig. 8 is a kind of schematic diagram of the interactive device of optional virtual reality scenario according to embodiments of the present invention;
Fig. 9 is a kind of structured flowchart of the interactive terminal of virtual reality scenario according to embodiments of the present invention.
Embodiment
In order that those skilled in the art more fully understand the present invention program, below in conjunction with the embodiment of the present invention Accompanying drawing, the technical scheme in the embodiment of the present invention is clearly and completely described, it is clear that described embodiment is only The embodiment of a part of the invention, rather than whole embodiments.Based on the embodiment in the present invention, ordinary skill people The every other embodiment that member is obtained under the premise of creative work is not made, should all belong to the model that the present invention is protected Enclose.
It should be noted that term " first " in description and claims of this specification and above-mentioned accompanying drawing, " Two " etc. be for distinguishing similar object, without for describing specific order or precedence.It should be appreciated that so using Data can exchange in the appropriate case, so as to embodiments of the invention described herein can with except illustrating herein or Order beyond those of description is implemented.In addition, term " comprising " and " having " and their any deformation, it is intended that cover Lid is non-exclusive to be included, for example, the process, method, system, product or the equipment that contain series of steps or unit are not necessarily limited to Those steps or unit clearly listed, but may include not list clearly or for these processes, method, product Or the intrinsic other steps of equipment or unit.
Embodiment 1
There is provided a kind of embodiment of the exchange method of above-mentioned virtual reality scenario in embodiments of the present invention.It is used as one Optional embodiment is planted, the exchange method of the virtual reality scenario can be, but not limited to be applied to application ring as shown in Figure 1 In border, display needs to be interacted in the equipment 102 for realizing virtual reality (Virtual Reality, abbreviation VR) scene interactivity Object, the initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, its In, ray is shown in virtual reality scenario;The final goal pair collided with ray is detected in initial target object As;Then the control that ray in final goal object is chosen is operated.
In the present embodiment, determine that permission and ray collide in the object shown in virtual reality scenario first After beginning destination object, wherein, ray is shown in virtual reality scenario, then detected and ray in above-mentioned initial target object The final goal object collided, is operated to the control that ray in final goal object is chosen.Pass through so as to realize Ray collision detection obtains the final goal object to be interacted in virtual reality scenario, and ray is chosen most with reaching The purpose that control in whole destination object is operated, and then overcome can not be to the object in virtual reality scenario in correlation technique The problem of precisely interact.
Alternatively, in the present embodiment, it is above-mentioned to be used to realize that the equipment of virtual reality scenario interaction include but not limit In at least one of:VR glasses, and other hardware devices for realizing virtual reality scenario interaction.Above-mentioned simply one kind is shown Example, the present embodiment does not do any restriction to this.
According to embodiments of the present invention there is provided a kind of exchange method of virtual reality scenario, as shown in Fig. 2 this method bag Include:
The initial target pair for allowing to collide with ray is determined in S202, the object shown in virtual reality scenario As, wherein, ray is shown in virtual reality scenario;
S204, detects the final goal object collided with ray in initial target object;
S206, is operated to the control that ray in final goal object is chosen.
Alternatively, in the present embodiment, the exchange method of above-mentioned virtual reality scenario can be, but not limited to be applied to use Virtual reality device carry out virtual reality interactive application during, for example, above-mentioned virtual reality interactive application can include but It is not limited to reality-virtualizing game, the broadcasting of virtual reality video display, virtual reality shopping etc..Above-mentioned is only a kind of example, this implementation Any restriction is not done to this in example.Wherein, the virtual reality device can include but is not limited to VR glasses.Further, above-mentioned friendship Mutual process can be, but not limited to interact with the object in the plane in virtual reality scenario, for example, as shown in figure 3, dotted line Virtual reality scenario content shown in VR glasses (equipment 102 of virtual reality scenario interaction) is shown in square frame:Two Display needs interactive object A to object E in dimensional plane, wherein, the final goal object collided with ray is object E; In addition, in the present embodiment, above-mentioned interaction can also but be not limited to enter with the three-dimensional character object in virtual reality scenario Row interaction, for example, showing as shown in figure 4, being shown in dashed rectangle in VR glasses (equipment 102 of virtual reality scenario interaction) The virtual reality scenario content shown:Display needs interactive object P, object X, object Y and object Z in three dimensions, wherein, The final goal object collided with ray is object Y.Above-mentioned is only not do any limit to this in a kind of example, the present embodiment It is fixed.
It should be noted that in the present embodiment, determining permission with penetrating in the object shown in virtual reality scenario After the initial target object that line collides, wherein, ray is shown in virtual reality scenario, then in above-mentioned initial target object In detect the final goal object that is collided with ray, the control that ray in final goal object is chosen is grasped Make.So as to realize by ray collision detection from initial target object, acquisition to be interacted most in virtual reality scenario Whole destination object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes phase The problem of to the object in virtual reality scenario can not precisely interact in the technology of pass.Further, ray collision detection is passed through Realization is interacted with the object in virtual reality scenario, also simplify interactive operation, without using the hardware devices such as mouse Complicated interacting operation is carried out, and then is reached in the case where ensureing interaction accuracy, the effect of interactive efficiency is improved.
In addition, in the present embodiment, above-mentioned collision detection (also referred to as collision detection (Collision detection)), Typically refer to a kind of judge the method whether two or more objects produce common factor.It can apply to electronic game and other calculated In physics, it can also be applied in artificial intelligence.Except being determined for whether two objects have collided, it can also use In the time (TOI) for calculating impact, and obtain the position of object intersection-type collision.
Alternatively, in the present embodiment, determine that permission occurs with ray in the object shown in virtual reality scenario The initial target object of collision includes:Determine that the ray type with ray has corresponding relation according to the object type of object Destination object type, the object of destination object type is defined as the initial target object for allowing to collide with ray.From And realize using the mapping relations between ray type and object type, determine the initial target for allowing to collide with ray Object, to avoid object unrelated with ray in virtual reality scenario from participating in interaction deterministic process, reduces the judgement time, Jin Erda To the effect for simplifying interaction judgment step.
Alternatively, in the present embodiment, the final goal pair collided with ray is detected in initial target object As including:Each initial target object is traveled through, detects that each initial target object space shared in virtual reality scenario is It is no to there is the part overlapped with ray;The initial target object that there will be the part overlapped with ray is defined as final goal pair As.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to Amount, above-mentioned detection current target object space shared in virtual reality scenario whether there is the part that is overlapped with ray can be with But it is not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.That is, by virtual reality Each point position in scape on shared space and ray carries out collision detection respectively, is accurately judged to collide with ray to reach Final goal object effect.If, in the present embodiment, can be with addition, the multiframe picture shown by virtual reality scenario But it is not limited to all do collision detection to each frame, to determine whether include above-mentioned ray in space shared in virtual reality scenario Point in the vector of composition, so as to ensure the accuracy of collision detection.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting It is dynamic to operate, choose operation etc..Execution event corresponding with aforesaid operations can include but is not limited at least one of:
1) entry event (Enter):Ray starts to point to some object, by the ray for comparing this frame and previous frame The change for hitting list object, see with the presence or absence of the object that newly adds, the event of the new addition object of triggering;
2) event (Leave) is left:Ray, which leaves, points to some object, by the ray for comparing this frame and previous frame The change for hitting list object, see with the presence or absence of the object newly deleted, the new event for deleting object of triggering;
3) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list The event of all objects;
4) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list There is the event of object;
5) moving event (Move):Ray is moved on each control in final goal object, in triggering interaction list The event of all objects.
6) other:For some special events, such as drag events of drag operation triggering etc..
It should be noted that above-mentioned control can be, but not limited to the function for a series of pre-productions for constituting interface Control, for example, button, check box, sliding block, progress bar etc..Above-mentioned is only not do any limit to this in a kind of example, the present embodiment It is fixed.
Determine that permission is sent out with ray in the embodiment provided by the application, the object shown in virtual reality scenario After the initial target object of raw collision, wherein, ray is shown in virtual reality scenario, then examined in above-mentioned initial target object The final goal object collided with ray is measured, the control that ray in final goal object is chosen is operated.From And realize by ray collision detection from initial target object, obtain the final goal to be interacted in virtual reality scenario Object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes correlation technique In can not in virtual reality scenario object carry out precisely interact the problem of.
As a kind of optional scheme, determine that permission is touched with ray in the object shown in virtual reality scenario The initial target object hit includes:
S1, obtains the object type of object;
S2, determines that the ray type with ray has the destination object type of corresponding relation in object type;
S3, initial target object is defined as in object by type for the object of destination object type.
Alternatively, in the present embodiment, the type of the object to allowing to collide with ray is set to destination object class Type, is represented with object collision objects, by for the ray type of collision detection laser collision Object tables Show.Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.
It should be noted that in the present embodiment, the above-mentioned collision mode for being used to carrying out ray collision detection can with but not It is limited to use the collision bounding box in correlation technique, by the way that the individual being detected is surrounded with collision bounding box, detects that each is touched Hit whether bounding box has intersection, when producing intersection, detect to produce collision.Above-mentioned is only a kind of example, this reality Apply in example and the mode of collision detection is not limited, other collision detection modes can also be used.
The embodiment provided by the application, by classifying by type to the object in virtual reality scenario, with reality Now pair destination object type for having corresponding relation with ray type carries out collision detection, so as to realize for not being destination object The object of type directly skips over detection, reaches that ray reduces the quantity of the object of collision detection, realization greatlys save collision detection Time loss effect, and then reach simplify interaction judgment step effect.
As a kind of optional scheme, the final goal object collided with ray is detected in initial target object Including:
S1, following steps are performed to each object in initial target object, wherein, initial target object is below performing It is considered as current target object during step:
S11, detection current target object space shared in virtual reality scenario whether there is the portion overlapped with ray Point;
S12, if space has the part overlapped with ray, final goal object is defined as by current target object.
Alternatively, in the present embodiment, it can be, but not limited to hold the initial target object for being defined as destination object type Row Ergodic judgement process, wherein, it regard each initial target object as current target object respectively:Detection current target object exists Shared space whether there is the part overlapped with ray in virtual reality scenario;There will be the initial of the part that is overlapped with ray Destination object is defined as final goal object.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to Amount.In the present embodiment, above-mentioned detection current target object space shared in virtual reality scenario whether there is and ray The part of coincidence can be, but not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.Namely Say, each point position on space and ray shared in virtual reality scenario is subjected to collision detection respectively, it is accurate to reach Judge the effect of final goal object collided with ray.Wherein, above-mentioned current target object institute in virtual reality scenario The space accounted for can be, but not limited to as shown in figure 3, by taking object E as an example, space shared in virtual reality scenario object E is one Individual rectangle plane region;As shown in figure 4, by taking object Y as an example, space shared in virtual reality scenario object Y is a people Object angle color Y shared solid spaces in three dimensions.
It is specific to be illustrated with reference to the example below, for example, by taking object E shown in Fig. 3 as an example, if detection emergent ray and object E Corresponding rectangle plane region, which exists, to intersect, then can detect that object E has the part overlapped with ray, can determine object E For final goal object.In another example, by taking object Y shown in Fig. 4 as an example, if detection emergent ray character Y institutes corresponding with object Y The solid space accounted for, which exists, to intersect, then can detect that object Y has the part overlapped with ray, can be defined as object Y finally Destination object.
The embodiment provided by the application, by traveling through each initial target object, judges that each is initial successively respectively Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, to realize according to judged result Accurately detect for interactive final goal object, it is ensured that interactive accuracy.
As a kind of optional scheme, detection current target object space shared in virtual reality scenario whether there is The part overlapped with ray includes:
Whether S1, detection space includes the point on ray, wherein, ray includes:Starting point and terminal, the direction of ray by Starting point points to terminal;
S2, if detecting, space includes the point on ray, it is determined that goes out space and there is the part overlapped with ray.
Alternatively, in the present embodiment, above-mentioned ray can be, but not limited to be to include the Directed Vector of starting point and terminal. Wherein, above-mentioned starting point can be, but not limited to send the position of ray in virtual reality scenario, and terminal is according to ray institute The position reached after finger direction extension certain distance, each position on this ray is required for carrying out collision detection.
For example, as shown in fig. 5, it is assumed that starting point (such as uses Hand with position of the finger in virtual reality scenario Position is represented) exemplified by, according to the extended distance (such as being represented with Max Distance) of the ray set in advance, acquisition pair The terminal (such as being represented with Max Range Position) answered, whether being taken up space according to finger orientation detection object Y, it is upper to include State the point on ray.For example, as shown in fig. 6, collision detection result is represented:Solid shared by the corresponding character Y of object Y is empty Between with above-mentioned ray (starting point be Hand Position, terminal be Max Range Position) there is intersection, crosspoint is Position shown in Fig. 6 rhombuses, then can detect that object Y has the part overlapped with ray, object Y can be defined as into final goal pair As.
Alternatively, in the present embodiment, terminal is determined by following steps:According to starting point in virtual reality scenario Position, the direction vector for direction indication, the predetermined distance range value determined, determine terminal in virtual reality scenario Position.
For example, as shown in fig. 7, the formula of the terminal for determining above-mentioned ray is as follows:
Max Range Position=Hand Position+Hand Forward Vector*Max Distance
Wherein, above-mentioned Hand Forward Vector are used for the unit vector for representing directions of rays.That is in starting On the three-dimensional coordinate of point, the three-dimensional coordinate along the extended distance in directions of rays is superimposed, so as to obtain the three-dimensional coordinate of terminal.
The embodiment provided by the application, by whether detecting space comprising the point on ray, to realize that detection is current Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, that is to say, that will virtually show Each point position in real field scape on shared space and ray carries out collision detection respectively, is accurately judged to and ray with reaching The effect of the final goal object of collision, so as to ensure the accuracy of collision detection, and then realizes the effect for improving interaction accuracy Really.
As a kind of optional scheme, the control chosen to ray in final goal object, which carries out operation, to be included:
S1, chooses the control on the position that ray overlaps with final goal object in final goal object;
S2, is operated to control.
Alternatively, in the present embodiment, carrying out operation to control includes:
S21, when detecting interactive button corresponding with virtual reality scenario and being pressed, determines to detect and starts to click on Event;
S22, when detecting interactive button corresponding with virtual reality scenario and being lifted, determines that detecting end clicks on Event;
S23, detect start click event and terminate click event when, to control carry out clicking operation.
Alternatively, in the present embodiment, object can include but is not limited in above-mentioned virtual reality scenario:Illusory Dynamic Graph Object in object (being referred to as UMG objects) designed by shape (Unreal Mobile) UI designers, destination object type.
It should be noted that in the present embodiment, if what is collided with ray is UMG objects, need to call UMG phases The control logic of pass, for example, in UE4 environment, for UMG objects, by calling Widget On series of functions to trigger Widget event response, so as to reach purpose of the compatible general game engine for UMG support.That is, virtual Real content developer can use general UI preparation method to make user interface (User for virtual reality scenario Interface, abbreviation UI) menu, the menu may be directly applied to virtual reality scenario and interacts.In addition, in this implementation In example, if what is collided with ray is not UMG objects, but the object in destination object type, then it can refer to this implementation Interactive mode described in example is interacted.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting It is dynamic to operate, choose operation etc..Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.For example, to click on behaviour As example, it can be, but not limited to determine by following event:
1) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list The event of all objects.The readjustment machine that the On series of functions triggering that widget has been present has been present is directly invoked for UMG System, for the call back function of other then self-defined events;
2) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list There is the event of object.The callback mechanism that the On series of functions triggering that widget has been present has been present is directly invoked for UMG, For the call back function of other then self-defined events.
That is, when detecting interactive button corresponding with virtual reality scenario and being pressed, determining and detecting out Initial point hits event;When detecting interactive button corresponding with virtual reality scenario and being lifted, determine that detecting end clicks on Event;Then detect start click event and terminate click event when, confirm to control carry out clicking operation.
For example, by taking Fig. 5 as an example, when selected object Y is final goal object, choosing ray to be overlapped with final goal object Position (such as object Y right crus of diaphragm), then on the position (such as object Y right crus of diaphragm) of coincidence control perform clicking operation, example Such as detect that interactive button is pressed start click event and event is clicked in end when interaction button is lifted, then can confirm Clicking operation is performed to above-mentioned control.So as to realize the interactive controlling of character Y corresponding to object Y right crus of diaphragm, for example, lead to Cross click right crus of diaphragm and realize that control lifts right crus of diaphragm and performs striking action.Above-mentioned is only this not to be appointed in a kind of example, the present embodiment What is limited.
The embodiment provided by the application, the position for choosing ray to be overlapped with final goal object in final goal object The control put, is operated to the control.Completed so as to realize in virtual reality scenario by ray between object Interaction, to ensure to interact accuracy.
It should be noted that for foregoing each method embodiment, in order to be briefly described, therefore it is all expressed as a series of Combination of actions, but those skilled in the art should know, the present invention is not limited by described sequence of movement because According to the present invention, some steps can be carried out sequentially or simultaneously using other.Secondly, those skilled in the art should also know Know, embodiment described in this description belongs to preferred embodiment, involved action and module is not necessarily of the invention It is necessary.
Through the above description of the embodiments, those skilled in the art can be understood that according to above-mentioned implementation The method of example can add the mode of required general hardware platform to realize by software, naturally it is also possible to by hardware, but a lot In the case of the former be more preferably embodiment.Understood based on such, technical scheme is substantially in other words to existing The part that technology contributes can be embodied in the form of software product, and the computer software product is stored in a storage In medium (such as ROM/RAM, magnetic disc, CD), including some instructions are to cause a station terminal equipment (can be mobile phone, calculate Machine, server, or network equipment etc.) perform method described in each of the invention embodiment.
Embodiment 2
According to embodiments of the present invention, a kind of the virtual of exchange method for being used to implement above-mentioned virtual reality scenario is additionally provided The interactive device of reality scene.As shown in figure 8, the device can include:
1) determining unit 802, for determining that permission is touched with ray in the object that is shown in virtual reality scenario The initial target object hit, wherein, ray is shown in virtual reality scenario;
2) detection unit 804, for detecting the final goal object collided with ray in initial target object;
3) operating unit 806, for being operated to the control that ray in final goal object is chosen.
Alternatively, in the present embodiment, the interactive device of above-mentioned virtual reality scenario can be, but not limited to be applied to use Virtual reality device carry out virtual reality interactive application during, for example, above-mentioned virtual reality interactive application can include but It is not limited to reality-virtualizing game, the broadcasting of virtual reality video display, virtual reality shopping etc..Above-mentioned is only a kind of example, this implementation Any restriction is not done to this in example.Wherein, the virtual reality device can include but is not limited to VR glasses.Further, above-mentioned friendship Mutual process can be, but not limited to interact with the object in the plane in virtual reality scenario, for example, as shown in figure 3, dotted line Virtual reality scenario content shown in VR glasses (equipment 102 of virtual reality scenario interaction) is shown in square frame:Two Display needs interactive object A to object E in dimensional plane, wherein, the final goal object collided with ray is object E; In addition, in the present embodiment, above-mentioned interaction can also but be not limited to enter with the three-dimensional character object in virtual reality scenario Row interaction, for example, showing as shown in figure 4, being shown in dashed rectangle in VR glasses (equipment 102 of virtual reality scenario interaction) The virtual reality scenario content shown:Display needs interactive object P, object X, object Y and object Z in three dimensions, wherein, The final goal object collided with ray is object Y.Above-mentioned is only not do any limit to this in a kind of example, the present embodiment It is fixed.
It should be noted that in the present embodiment, determining permission with penetrating in the object shown in virtual reality scenario After the initial target object that line collides, wherein, ray is shown in virtual reality scenario, then in above-mentioned initial target object In detect the final goal object that is collided with ray, the control that ray in final goal object is chosen is grasped Make.So as to realize by ray collision detection from initial target object, acquisition to be interacted most in virtual reality scenario Whole destination object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes phase The problem of to the object in virtual reality scenario can not precisely interact in the technology of pass.Further, ray collision detection is passed through Realization is interacted with the object in virtual reality scenario, also simplify interactive operation, without using the hardware devices such as mouse Complicated interacting operation is carried out, and then is reached in the case where ensureing interaction accuracy, the effect of interactive efficiency is improved.
In addition, in the present embodiment, above-mentioned collision detection (also referred to as collision detection (Collision detection)), Typically refer to a kind of judge the method whether two or more objects produce common factor.It can apply to electronic game and other calculated In physics, it can also be applied in artificial intelligence.Except being determined for whether two objects have collided, it can also use In the time (TOI) for calculating impact, and obtain the position of object intersection-type collision.
Alternatively, in the present embodiment, determine that permission occurs with ray in the object shown in virtual reality scenario The initial target object of collision includes:Determine that the ray type with ray has corresponding relation according to the object type of object Destination object type, the object of destination object type is defined as the initial target object for allowing to collide with ray.From And realize using the mapping relations between ray type and object type, determine the initial target for allowing to collide with ray Object, to avoid object unrelated with ray in virtual reality scenario from participating in interaction deterministic process, reduces the judgement time, Jin Erda To the effect for simplifying interaction judgment step.
Alternatively, in the present embodiment, the final goal pair collided with ray is detected in initial target object As including:Each initial target object is traveled through, detects that each initial target object space shared in virtual reality scenario is It is no to there is the part overlapped with ray;The initial target object that there will be the part overlapped with ray is defined as final goal pair As.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to Amount, above-mentioned detection current target object space shared in virtual reality scenario whether there is the part that is overlapped with ray can be with But it is not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.That is, by virtual reality Each point position in scape on shared space and ray carries out collision detection respectively, is accurately judged to collide with ray to reach Final goal object effect.If, in the present embodiment, can be with addition, the multiframe picture shown by virtual reality scenario But it is not limited to all do collision detection to each frame, to determine whether include above-mentioned ray in space shared in virtual reality scenario Point in the vector of composition, so as to ensure the accuracy of collision detection.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting It is dynamic to operate, choose operation etc..Execution event corresponding with aforesaid operations can include but is not limited at least one of:
1) entry event (Enter):Ray starts to point to some object, by the ray for comparing this frame and previous frame The change for hitting list object, see with the presence or absence of the object that newly adds, the event of the new addition object of triggering;
2) event (Leave) is left:Ray, which leaves, points to some object, by the ray for comparing this frame and previous frame The change for hitting list object, see with the presence or absence of the object newly deleted, the new event for deleting object of triggering;
3) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list The event of all objects;
4) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list There is the event of object;
5) moving event (Move):Ray is moved on each control in final goal object, in triggering interaction list The event of all objects.
6) other:For some special events, such as drag events of drag operation triggering etc..
It should be noted that above-mentioned control can be, but not limited to the function for a series of pre-productions for constituting interface Control, for example, button, check box, sliding block, progress bar etc..Above-mentioned is only not do any limit to this in a kind of example, the present embodiment It is fixed.
Determine that permission is sent out with ray in the embodiment provided by the application, the object shown in virtual reality scenario After the initial target object of raw collision, wherein, ray is shown in virtual reality scenario, then examined in above-mentioned initial target object The final goal object collided with ray is measured, the control that ray in final goal object is chosen is operated.From And realize by ray collision detection from initial target object, obtain the final goal to be interacted in virtual reality scenario Object, to reach purpose that the control in the final goal object chosen to ray is operated, and then overcomes correlation technique In can not in virtual reality scenario object carry out precisely interact the problem of.
As a kind of optional scheme, determining unit 802 includes:
1) acquisition module, the object type for obtaining object;
2) the first determining module, the ray type for being determined in object type with ray has the mesh of corresponding relation Mark object type;
3) the second determining module, initial target object is defined as in object by type for the object of destination object type.
Alternatively, in the present embodiment, the type of the object to allowing to collide with ray is set to destination object class Type, is represented with object collision objects, by for the ray type of collision detection laser collision Object tables Show.Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.
It should be noted that in the present embodiment, the above-mentioned collision mode for being used to carrying out ray collision detection can with but not It is limited to use the collision bounding box in correlation technique, by the way that the individual being detected is surrounded with collision bounding box, detects that each is touched Hit whether bounding box has intersection, when producing intersection, detect to produce collision.Above-mentioned is only a kind of example, this reality Apply in example and the mode of collision detection is not limited, other collision detection modes can also be used.
The embodiment provided by the application, by classifying by type to the object in virtual reality scenario, with reality Now pair destination object type for having corresponding relation with ray type carries out collision detection, so as to realize for not being destination object The object of type directly skips over detection, reaches that ray reduces the quantity of the object of collision detection, realization greatlys save collision detection Time loss effect, and then reach simplify interaction judgment step effect.
As a kind of optional scheme, detection unit 804 includes:
1) processing module, for performing following steps to each object in initial target object, wherein, initial target pair As being considered as current target object when performing following steps:
S1, detection current target object space shared in virtual reality scenario whether there is the portion overlapped with ray Point;
S2, if space has the part overlapped with ray, final goal object is defined as by current target object.
Alternatively, in the present embodiment, it can be, but not limited to hold the initial target object for being defined as destination object type Row Ergodic judgement process, wherein, it regard each initial target object as current target object respectively:Detection current target object exists Shared space whether there is the part overlapped with ray in virtual reality scenario;There will be the initial of the part that is overlapped with ray Destination object is defined as final goal object.
It should be noted that in the present embodiment, above-mentioned ray can be, but not limited to for starting point and terminal to Amount.In the present embodiment, above-mentioned detection current target object space shared in virtual reality scenario whether there is and ray The part of coincidence can be, but not limited to:Detect the point whether included in above-mentioned space in the vector that above-mentioned ray is constituted.Namely Say, each point position on space and ray shared in virtual reality scenario is subjected to collision detection respectively, it is accurate to reach Judge the effect of final goal object collided with ray.Wherein, above-mentioned current target object institute in virtual reality scenario The space accounted for can be, but not limited to as shown in figure 3, by taking object E as an example, space shared in virtual reality scenario object E is one Individual rectangle plane region;As shown in figure 4, by taking object Y as an example, space shared in virtual reality scenario object Y is a people Object angle color Y shared solid spaces in three dimensions.
It is specific to be illustrated with reference to the example below, for example, by taking object E shown in Fig. 3 as an example, if detection emergent ray and object E Corresponding rectangle plane region, which exists, to intersect, then can detect that object E has the part overlapped with ray, can determine object E For final goal object.In another example, by taking object Y shown in Fig. 4 as an example, if detection emergent ray character Y institutes corresponding with object Y The solid space accounted for, which exists, to intersect, then can detect that object Y has the part overlapped with ray, can be defined as object Y finally Destination object.
The embodiment provided by the application, by traveling through each initial target object, judges that each is initial successively respectively Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, to realize according to judged result Accurately detect for interactive final goal object, it is ensured that interactive accuracy.
As a kind of optional scheme, processing module realizes detection current target object in virtual reality by following steps Shared space whether there is the part overlapped with ray in scene:
Whether S1, detection space includes the point on ray, wherein, ray includes:Starting point and terminal, the direction of ray by Starting point points to terminal;
S2, if detecting, space includes the point on ray, it is determined that goes out space and there is the part overlapped with ray.
Alternatively, in the present embodiment, above-mentioned ray can be, but not limited to be to include the Directed Vector of starting point and terminal. Wherein, above-mentioned starting point can be, but not limited to send the position of ray in virtual reality scenario, and terminal is according to ray institute The position reached after finger direction extension certain distance, each position on this ray is required for carrying out collision detection.
For example, as shown in fig. 5, it is assumed that starting point (such as uses Hand with position of the finger in virtual reality scenario Position is represented) exemplified by, according to the extended distance (such as being represented with Max Distance) of the ray set in advance, acquisition pair The terminal (such as being represented with Max Range Position) answered, whether being taken up space according to finger orientation detection object Y, it is upper to include State the point on ray.For example, as shown in fig. 6, collision detection result is represented:Solid shared by the corresponding character Y of object Y is empty Between with above-mentioned ray (starting point be Hand Position, terminal be Max Range Position) there is intersection, crosspoint is Position shown in Fig. 6 rhombuses, then can detect that object Y has the part overlapped with ray, object Y can be defined as into final goal pair As.
Alternatively, in the present embodiment, processing module determines terminal by following steps by following steps:According to starting Position of the point in virtual reality scenario, the direction vector for direction indication, the predetermined distance range value determined, determine end Position of the point in virtual reality scenario.
For example, as shown in fig. 7, the formula of the terminal for determining above-mentioned ray is as follows:
Max Range Position=Hand Position+Hand Forward Vector*Max Distance
Wherein, above-mentioned Hand Forward Vector are used for the unit vector for representing directions of rays.That is in starting On the three-dimensional coordinate of point, the three-dimensional coordinate along the extended distance in directions of rays is superimposed, so as to obtain the three-dimensional coordinate of terminal.
The embodiment provided by the application, by whether detecting space comprising the point on ray, to realize that detection is current Destination object space shared in virtual reality scenario whether there is the part overlapped with ray, that is to say, that will virtually show Each point position in real field scape on shared space and ray carries out collision detection respectively, is accurately judged to and ray with reaching The effect of the final goal object of collision, so as to ensure the accuracy of collision detection, and then realizes the effect for improving interaction accuracy Really.
As a kind of optional scheme, operating unit 806 includes:
1) module is chosen, for choosing the control on the position that ray overlaps with final goal object in final goal object Part;
2) operation module, for being operated to control.
Alternatively, in the present embodiment, operation module includes:
(1) first determination sub-module, for when detecting interactive button corresponding with virtual reality scenario and being pressed, really Make to detect and start click event;
(2) second determination sub-modules, for when detecting interactive button corresponding with virtual reality scenario and being lifted, really Make and detect end click event;
(3) the 3rd determination sub-modules, for detect start click event and terminate click event when, to control carry out Clicking operation.
Alternatively, in the present embodiment, object can include but is not limited in above-mentioned virtual reality scenario:Illusory Dynamic Graph Object in object (being referred to as UMG objects) designed by shape (Unreal Mobile) UI designers, destination object type.
It should be noted that in the present embodiment, if what is collided with ray is UMG objects, need to call UMG phases The control logic of pass, for example, in UE4 environment, for UMG objects, by calling Widget On series of functions to trigger Widget event response, so as to reach purpose of the compatible general game engine for UMG support.That is, virtual Real content developer can use general UI preparation method to make user interface (User for virtual reality scenario Interface, abbreviation UI) menu, the menu may be directly applied to virtual reality scenario and interacts.In addition, in this implementation In example, if what is collided with ray is not UMG objects, but the object in destination object type, then it can refer to this implementation Interactive mode described in example is interacted.
Alternatively, in the present embodiment, the operation carried out using ray can include but is not limited to:Clicking operation, shifting It is dynamic to operate, choose operation etc..Above-mentioned is only not do any restriction to this in a kind of example, the present embodiment.For example, to click on behaviour As example, it can be, but not limited to determine by following event:
1) event (TouchDown) is pressed:Ray points to control, when pressing interactive button, in triggering interaction list The event of all objects.The readjustment machine that the On series of functions triggering that widget has been present has been present is directly invoked for UMG System, for the call back function of other then self-defined events;
2) event (TouchUp) is lifted:Ray points to control, when lifting interactive button, institute in triggering interaction list There is the event of object.The callback mechanism that the On series of functions triggering that widget has been present has been present is directly invoked for UMG, For the call back function of other then self-defined events.
That is, when detecting interactive button corresponding with virtual reality scenario and being pressed, determining and detecting out Initial point hits event;When detecting interactive button corresponding with virtual reality scenario and being lifted, determine that detecting end clicks on Event;Then detect start click event and terminate click event when, confirm to control carry out clicking operation.
For example, by taking Fig. 5 as an example, when selected object Y is final goal object, choosing ray to be overlapped with final goal object Position (such as object Y right crus of diaphragm), then on the position (such as object Y right crus of diaphragm) of coincidence control perform clicking operation, example Such as detect that interactive button is pressed start click event and event is clicked in end when interaction button is lifted, then can confirm Clicking operation is performed to above-mentioned control.So as to realize the interactive controlling of character Y corresponding to object Y right crus of diaphragm, for example, lead to Cross click right crus of diaphragm and realize that control lifts right crus of diaphragm and performs striking action.Above-mentioned is only this not to be appointed in a kind of example, the present embodiment What is limited.
The embodiment provided by the application, the position for choosing ray to be overlapped with final goal object in final goal object The control put, is operated to the control.Completed so as to realize in virtual reality scenario by ray between object Interaction, to ensure to interact accuracy.
Herein it should be noted that above-mentioned module is identical with example and application scenarios that the step of correspondence is realized, but not It is limited to the disclosure of that of above-described embodiment 1.It should be noted that above-mentioned module as a part for device may operate in as It in hardware environment shown in Fig. 1, can be realized, can also be realized by hardware by software, wherein, hardware environment includes network Environment.
Embodiment 3
According to embodiments of the present invention, a kind of the virtual of exchange method for being used to implement above-mentioned virtual reality scenario is additionally provided The interactive terminal of reality scene, as shown in figure 9, the terminal includes:
1) communication interface 902, are set to obtain ray shown in virtual reality scenario;
2) processor 904, are connected with communication interface 902, are set to determine in the object that shows in virtual reality scenario Go out to allow the initial target object collided with ray, wherein, ray is shown in virtual reality scenario;It is also configured to first The final goal object collided with ray is detected in beginning destination object;It is also configured to ray institute in final goal object The control chosen is operated;
3) display 906, are connected with processor 904, are set to real ray, initial target object and final goal pair As;
4) memory 908, are connected with communication interface 902, processor 904 and display 906, are set to storage above-mentioned initial Destination object, final goal object, and ray position information.
Alternatively, the specific example in the present embodiment may be referred to showing described in above-described embodiment 1 and embodiment 2 Example, the present embodiment will not be repeated here.
One of ordinary skill in the art will appreciate that all or part of step in the various methods of above-described embodiment is can To be completed by program come the device-dependent hardware of command terminal, the program can be stored in a computer-readable recording medium In, storage medium can include:Flash disk, read-only storage (Read-Only Memory, ROM), random access device (Random Access Memory, RAM), disk or CD etc..
Embodiment 4
Embodiments of the invention additionally provide a kind of storage medium.Alternatively, in the present embodiment, above-mentioned storage medium can With at least one network equipment in multiple network equipments in the network shown in above-described embodiment.
Alternatively, in the present embodiment, storage medium is arranged to the program code that storage is used to perform following steps:
The initial target pair for allowing to collide with ray is determined in S1, the object shown in virtual reality scenario As, wherein, ray is shown in virtual reality scenario;
S2, detects the final goal object collided with ray in initial target object;
S3, is operated to the control that ray in final goal object is chosen.
Alternatively, the specific example in the present embodiment may be referred to showing described in above-described embodiment 1 and embodiment 2 Example, the present embodiment will not be repeated here.
Alternatively, in the present embodiment, above-mentioned storage medium can include but is not limited to:USB flash disk, read-only storage (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), mobile hard disk, magnetic disc or CD etc. is various can be with the medium of store program codes.
The embodiments of the present invention are for illustration only, and the quality of embodiment is not represented.
If the integrated unit in above-described embodiment is realized using in the form of SFU software functional unit and is used as independent product Sale or in use, the storage medium that above computer can be read can be stored in.Understood based on such, skill of the invention The part or all or part of the technical scheme that art scheme substantially contributes to prior art in other words can be with soft The form of part product is embodied, and the computer software product is stored in storage medium, including some instructions are to cause one Platform or multiple stage computers equipment (can be personal computer, server or network equipment etc.) perform each embodiment institute of the invention State all or part of step of method.
In the above embodiment of the present invention, the description to each embodiment all emphasizes particularly on different fields, and does not have in some embodiment The part of detailed description, may refer to the associated description of other embodiment.
, can be by others side in several embodiments provided herein, it should be understood that disclosed client Formula is realized.Wherein, device embodiment described above is only schematical, such as division of described unit, only one Kind of division of logic function, can there is other dividing mode when actually realizing, such as multiple units or component can combine or Another system is desirably integrated into, or some features can be ignored, or do not perform.It is another, it is shown or discussed it is mutual it Between coupling or direct-coupling or communication connection can be the INDIRECT COUPLING or communication link of unit or module by some interfaces Connect, can be electrical or other forms.
The unit illustrated as separating component can be or may not be it is physically separate, it is aobvious as unit The part shown can be or may not be physical location, you can with positioned at a place, or can also be distributed to multiple On NE.Some or all of unit therein can be selected to realize the mesh of this embodiment scheme according to the actual needs 's.
In addition, each functional unit in each embodiment of the invention can be integrated in a processing unit, can also That unit is individually physically present, can also two or more units it is integrated in a unit.Above-mentioned integrated list Member can both be realized in the form of hardware, it would however also be possible to employ the form of SFU software functional unit is realized.
Described above is only the preferred embodiment of the present invention, it is noted that for the ordinary skill people of the art For member, under the premise without departing from the principles of the invention, some improvements and modifications can also be made, these improvements and modifications also should It is considered as protection scope of the present invention.

Claims (14)

1. a kind of exchange method of virtual reality scenario, it is characterised in that including:
The initial target object for allowing to collide with ray is determined in the object shown in virtual reality scenario, wherein, The ray is shown in the virtual reality scenario;
The final goal object collided with the ray is detected in the initial target object;
The control that ray is chosen described in the final goal object is operated.
2. according to the method described in claim 1, it is characterised in that determined in the object shown in virtual reality scenario fair Perhaps the initial target object collided with ray includes:
Obtain the object type of the object;
Determine that the ray type with the ray has the destination object type of corresponding relation in the object type;
Type is defined as the initial target object for the object of the destination object type in the object.
3. according to the method described in claim 1, it is characterised in that detected in the initial target object and the ray The final goal object collided includes:
To in the initial target object each object perform following steps, wherein, the initial target object perform with It is considered as current target object during lower step:
Detect that current target object space shared in the virtual reality scenario whether there is to overlap with the ray Part;
If the space has the part overlapped with the ray, the current target object is defined as the final goal Object.
4. method according to claim 3, it is characterised in that the detection current target object is in the virtual reality Shared space whether there is the part overlapped with the ray and include in scape:
The space is detected whether comprising the point on the ray, wherein, the ray includes:Starting point and terminal, it is described to penetrate The direction of line the terminal is pointed to by the starting point;
If detecting, the space includes the point on the ray, it is determined that goes out the space and there is the portion overlapped with the ray Point.
5. method according to claim 4, it is characterised in that determine the terminal by following steps:
According to position of the starting point in the virtual reality scenario, the direction vector for indicating the direction, make a reservation for The distance range value of determination, determines position of the terminal in the virtual reality scenario.
6. method according to any one of claim 1 to 5, it is characterised in that described in the final goal object The control that ray is chosen, which carries out operation, to be included:
The control on the position that the ray overlaps with the final goal object is chosen in the final goal object;
The control is operated.
7. method according to claim 6, it is characterised in that carrying out operation to the control includes:
When detecting interactive button corresponding with the virtual reality scenario and being pressed, determine to detect and start to click on thing Part;
When detecting interactive button corresponding with the virtual reality scenario and being lifted, determine that detecting end clicks on thing Part;
Detect it is described start click event and the end and click on event when, clicking operation is carried out to the control.
8. a kind of interactive device of virtual reality scenario, it is characterised in that including:
Determining unit, for determining the initial mesh for allowing to collide with ray in the object that is shown in virtual reality scenario Object is marked, wherein, the ray is shown in the virtual reality scenario;
Detection unit, for the final goal object for detecting to collide with the ray in the initial target object;
Operating unit, for being operated to the control that ray is chosen described in the final goal object.
9. device according to claim 8, it is characterised in that the determining unit includes:
Acquisition module, the object type for obtaining the object;
First determining module, for determining that the ray type with the ray has corresponding relation in the object type Destination object type;
Second determining module, the initial target is defined as in the object by type for the object of the destination object type Object.
10. device according to claim 8, it is characterised in that the detection unit includes:
Processing module, for performing following steps to each object in the initial target object, wherein, the initial target Object is considered as current target object when performing following steps:
Detect that current target object space shared in the virtual reality scenario whether there is to overlap with the ray Part;
If the space has the part overlapped with the ray, the current target object is defined as the final goal Object.
11. device according to claim 10, it is characterised in that the processing module realizes detection institute by following steps State current target object space shared in the virtual reality scenario and whether there is the part overlapped with the ray:
The space is detected whether comprising the point on the ray, wherein, the ray includes:Starting point and terminal, it is described to penetrate The direction of line the terminal is pointed to by the starting point;
If detecting, the space includes the point on the ray, it is determined that goes out the space and there is the portion overlapped with the ray Point.
12. device according to claim 11, it is characterised in that the processing module passes through following step by following steps Suddenly the terminal is determined:
According to position of the starting point in the virtual reality scenario, the direction vector for indicating the direction, make a reservation for The distance range value of determination, determines position of the terminal in the virtual reality scenario.
13. the device according to any one of claim 8 to 12, it is characterised in that the operating unit includes:
Module is chosen, for the position for choosing the ray to be overlapped with the final goal object in the final goal object On control;
Operation module, for being operated to the control.
14. device according to claim 13, it is characterised in that the operation module includes:
First determination sub-module, for when detecting interactive button corresponding with the virtual reality scenario and being pressed, it is determined that Go out to detect and start click event;
Second determination sub-module, for when detecting interactive button corresponding with the virtual reality scenario and being lifted, it is determined that Go out to detect end click event;
3rd determination sub-module, for detect it is described start click event and the end and click on event when, to the control Part carries out clicking operation.
CN201710263759.2A 2017-04-20 2017-04-20 The exchange method and device of virtual reality scenario Active CN107145227B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201710263759.2A CN107145227B (en) 2017-04-20 2017-04-20 The exchange method and device of virtual reality scenario
PCT/CN2018/082512 WO2018192394A1 (en) 2017-04-20 2018-04-10 Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710263759.2A CN107145227B (en) 2017-04-20 2017-04-20 The exchange method and device of virtual reality scenario

Publications (2)

Publication Number Publication Date
CN107145227A true CN107145227A (en) 2017-09-08
CN107145227B CN107145227B (en) 2019-01-08

Family

ID=59775036

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710263759.2A Active CN107145227B (en) 2017-04-20 2017-04-20 The exchange method and device of virtual reality scenario

Country Status (2)

Country Link
CN (1) CN107145227B (en)
WO (1) WO2018192394A1 (en)

Cited By (18)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107704397A (en) * 2017-10-30 2018-02-16 歌尔科技有限公司 Applied program testing method, device and electronic equipment
CN108052253A (en) * 2017-12-28 2018-05-18 灵图互动(武汉)科技有限公司 A kind of virtual reality display method for manufacturing contents
CN108536288A (en) * 2018-03-27 2018-09-14 网易(杭州)网络有限公司 The treating method and apparatus of interactive operation in virtual reality
WO2018192394A1 (en) * 2017-04-20 2018-10-25 腾讯科技(深圳)有限公司 Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus
CN108762617A (en) * 2018-05-31 2018-11-06 苏州蜗牛数字科技股份有限公司 UI interactive systems, method and storage medium in a kind of VR environment
CN109407945A (en) * 2018-11-08 2019-03-01 山东数字人科技股份有限公司 A kind of 3 d mosaics implementation method and 3 d mosaics system based on VR
CN109725782A (en) * 2017-10-27 2019-05-07 腾讯科技(深圳)有限公司 A kind of method, apparatus that realizing virtual reality and smart machine, storage medium
CN109782909A (en) * 2018-12-29 2019-05-21 北京诺亦腾科技有限公司 A kind of exchange method and device of VR interactive device and VR scene
CN110032296A (en) * 2019-04-10 2019-07-19 网易(杭州)网络有限公司 Determination method, apparatus, terminal and the storage medium of virtual objects in terminal
CN110262729A (en) * 2019-05-20 2019-09-20 联想(上海)信息技术有限公司 A kind of object processing method and equipment
CN110969687A (en) * 2019-11-29 2020-04-07 中国商用飞机有限责任公司北京民用飞机技术研究中心 Collision detection method, device, equipment and medium
CN110990098A (en) * 2019-11-29 2020-04-10 北京新势界科技有限公司 Method and device for overlapping AR information points triggered by focus
CN111054066A (en) * 2019-12-17 2020-04-24 腾讯科技(深圳)有限公司 Operation method and device of virtual prop, storage medium and electronic device
CN111192354A (en) * 2020-01-02 2020-05-22 武汉瑞莱保能源技术有限公司 Three-dimensional simulation method and system based on virtual reality
CN111580669A (en) * 2020-05-12 2020-08-25 南京睿悦信息技术有限公司 Interaction method and device for virtual reality and augmented reality mobile end plane application
CN111569424A (en) * 2020-05-14 2020-08-25 北京代码乾坤科技有限公司 Physical settlement method and device in game scene
CN112107854A (en) * 2020-09-01 2020-12-22 苏州紫焰网络科技有限公司 Game task guide completion method, system and equipment
CN114578972A (en) * 2022-05-05 2022-06-03 江西科骏实业有限公司 Trigger method and system for compatible plane and curved surface UI (user interface) event in VR (virtual reality) scene

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105892632A (en) * 2015-11-16 2016-08-24 乐视致新电子科技(天津)有限公司 Method and device for judging the selection of UI (User Interface) widgets of virtual reality application
CN105913497A (en) * 2016-05-27 2016-08-31 杭州映墨科技有限公司 Virtual reality space mobile positioning system and virtual reality space mobile positioning method for virtual house inspecting
CN106445118A (en) * 2016-09-06 2017-02-22 网易(杭州)网络有限公司 Virtual reality interaction method and apparatus

Family Cites Families (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101441542B (en) * 2008-11-21 2011-05-11 广东威创视讯科技股份有限公司 Method and apparatus for recognizing multiple target objects by interactive input apparatus
CN101770324B (en) * 2008-12-31 2011-12-28 商泰软件(上海)有限公司 Method for realizing interactive operation of 3D graphical interface
CN106527887B (en) * 2016-10-18 2018-09-04 腾讯科技(深圳)有限公司 Dummy object choosing method, device and VR systems
CN107145227B (en) * 2017-04-20 2019-01-08 腾讯科技(深圳)有限公司 The exchange method and device of virtual reality scenario

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105892632A (en) * 2015-11-16 2016-08-24 乐视致新电子科技(天津)有限公司 Method and device for judging the selection of UI (User Interface) widgets of virtual reality application
CN105913497A (en) * 2016-05-27 2016-08-31 杭州映墨科技有限公司 Virtual reality space mobile positioning system and virtual reality space mobile positioning method for virtual house inspecting
CN106445118A (en) * 2016-09-06 2017-02-22 网易(杭州)网络有限公司 Virtual reality interaction method and apparatus

Cited By (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2018192394A1 (en) * 2017-04-20 2018-10-25 腾讯科技(深圳)有限公司 Interaction method and apparatus for virtual reality scene, storage medium and electronic apparatus
CN109725782A (en) * 2017-10-27 2019-05-07 腾讯科技(深圳)有限公司 A kind of method, apparatus that realizing virtual reality and smart machine, storage medium
CN109725782B (en) * 2017-10-27 2021-08-24 腾讯科技(深圳)有限公司 Method and device for realizing virtual reality, intelligent equipment and storage medium
CN107704397A (en) * 2017-10-30 2018-02-16 歌尔科技有限公司 Applied program testing method, device and electronic equipment
CN107704397B (en) * 2017-10-30 2020-10-09 歌尔科技有限公司 Application program testing method and device and electronic equipment
CN108052253A (en) * 2017-12-28 2018-05-18 灵图互动(武汉)科技有限公司 A kind of virtual reality display method for manufacturing contents
CN108052253B (en) * 2017-12-28 2020-09-25 灵图互动(武汉)科技有限公司 Virtual reality display content manufacturing method
CN108536288A (en) * 2018-03-27 2018-09-14 网易(杭州)网络有限公司 The treating method and apparatus of interactive operation in virtual reality
CN108762617A (en) * 2018-05-31 2018-11-06 苏州蜗牛数字科技股份有限公司 UI interactive systems, method and storage medium in a kind of VR environment
CN109407945A (en) * 2018-11-08 2019-03-01 山东数字人科技股份有限公司 A kind of 3 d mosaics implementation method and 3 d mosaics system based on VR
CN109407945B (en) * 2018-11-08 2021-08-27 山东数字人科技股份有限公司 Three-dimensional jigsaw realization method based on VR and three-dimensional jigsaw system
CN109782909A (en) * 2018-12-29 2019-05-21 北京诺亦腾科技有限公司 A kind of exchange method and device of VR interactive device and VR scene
CN109782909B (en) * 2018-12-29 2020-10-30 北京诺亦腾科技有限公司 Interaction method and device for VR interaction equipment and VR scene
CN110032296A (en) * 2019-04-10 2019-07-19 网易(杭州)网络有限公司 Determination method, apparatus, terminal and the storage medium of virtual objects in terminal
CN110032296B (en) * 2019-04-10 2022-06-03 网易(杭州)网络有限公司 Method and device for determining virtual object in terminal, terminal and storage medium
CN110262729B (en) * 2019-05-20 2021-11-16 联想(上海)信息技术有限公司 Object processing method and device
CN110262729A (en) * 2019-05-20 2019-09-20 联想(上海)信息技术有限公司 A kind of object processing method and equipment
CN110969687A (en) * 2019-11-29 2020-04-07 中国商用飞机有限责任公司北京民用飞机技术研究中心 Collision detection method, device, equipment and medium
CN110969687B (en) * 2019-11-29 2023-07-28 中国商用飞机有限责任公司北京民用飞机技术研究中心 Collision detection method, device, equipment and medium
CN110990098A (en) * 2019-11-29 2020-04-10 北京新势界科技有限公司 Method and device for overlapping AR information points triggered by focus
CN111054066B (en) * 2019-12-17 2021-01-15 腾讯科技(深圳)有限公司 Operation method and device of virtual prop, storage medium and electronic device
CN111054066A (en) * 2019-12-17 2020-04-24 腾讯科技(深圳)有限公司 Operation method and device of virtual prop, storage medium and electronic device
CN111192354A (en) * 2020-01-02 2020-05-22 武汉瑞莱保能源技术有限公司 Three-dimensional simulation method and system based on virtual reality
CN111580669A (en) * 2020-05-12 2020-08-25 南京睿悦信息技术有限公司 Interaction method and device for virtual reality and augmented reality mobile end plane application
CN111569424B (en) * 2020-05-14 2023-04-28 北京代码乾坤科技有限公司 Physical settlement method and device in game scene
CN111569424A (en) * 2020-05-14 2020-08-25 北京代码乾坤科技有限公司 Physical settlement method and device in game scene
CN112107854B (en) * 2020-09-01 2021-03-16 苏州紫焰网络科技有限公司 Game task guide completion method, system and equipment
CN112107854A (en) * 2020-09-01 2020-12-22 苏州紫焰网络科技有限公司 Game task guide completion method, system and equipment
CN114578972A (en) * 2022-05-05 2022-06-03 江西科骏实业有限公司 Trigger method and system for compatible plane and curved surface UI (user interface) event in VR (virtual reality) scene

Also Published As

Publication number Publication date
WO2018192394A1 (en) 2018-10-25
CN107145227B (en) 2019-01-08

Similar Documents

Publication Publication Date Title
CN107145227A (en) The exchange method and device of virtual reality scenario
KR101863041B1 (en) Creation of playable scene with an authoring system
US20190236259A1 (en) Method for 3d graphical authentication on electronic devices
CN108628533A (en) Three-dimensional graphical user interface
US20130176202A1 (en) Menu selection using tangible interaction with mobile devices
CN102096548A (en) Method and system for duplicating an object using a touch-sensitive display
CN106790828A (en) The control method and mobile terminal of a kind of mobile terminal
CN105988933B (en) The operable node recognition methods in interface, application testing method, apparatus and system
RU2667720C1 (en) Method of imitation modeling and controlling virtual sphere in mobile device
CN109407940A (en) A kind of target object chooses method, system, terminal and storage medium
CN109806591A (en) Signature generating method and device in scene of game
CN109701265A (en) Remote desktop cursor mode switching method, system, client and remote server
CN109701271A (en) Show the methods, devices and systems of icon
CN108170338A (en) Information processing method, device, electronic equipment and storage medium
CN113190142B (en) Cubic model-based 3D environment interaction method and device
CN109375866B (en) Screen touch click response method and system for realizing same
CN113244611A (en) Virtual article processing method, device, equipment and storage medium
CN108089713A (en) A kind of interior decoration method based on virtual reality technology
CN112150602A (en) Model image rendering method and device, storage medium and electronic equipment
CN106547339A (en) The control method and device of computer equipment
US11301125B2 (en) Vector object interaction
CN103793053B (en) Gesture projection method and device for mobile terminals
CN108932098A (en) Building demenstration method and device
CN112451967A (en) Game interaction method and device based on 3D screen interaction and computer equipment
CN114931746B (en) Interaction method, device and medium for 3D game based on pen type and touch screen interaction

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right

Effective date of registration: 20240102

Address after: 518000 Tencent Building, No. 1 High-tech Zone, Nanshan District, Shenzhen City, Guangdong Province, 35 Floors

Patentee after: TENCENT TECHNOLOGY (SHENZHEN) Co.,Ltd.

Patentee after: TENCENT CLOUD COMPUTING (BEIJING) Co.,Ltd.

Address before: 518000 Tencent Building, No. 1 High-tech Zone, Nanshan District, Shenzhen City, Guangdong Province, 35 Floors

Patentee before: TENCENT TECHNOLOGY (SHENZHEN) Co.,Ltd.

TR01 Transfer of patent right