CN107506040A - A kind of space path method and system for planning - Google Patents
A kind of space path method and system for planning Download PDFInfo
- Publication number
- CN107506040A CN107506040A CN201710754154.3A CN201710754154A CN107506040A CN 107506040 A CN107506040 A CN 107506040A CN 201710754154 A CN201710754154 A CN 201710754154A CN 107506040 A CN107506040 A CN 107506040A
- Authority
- CN
- China
- Prior art keywords
- space
- rigid body
- thickness
- layer
- deep
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/011—Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/01—Input arrangements or combined input and output arrangements for interaction between user and computer
- G06F3/048—Interaction techniques based on graphical user interfaces [GUI]
- G06F3/0481—Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
- G06F3/04815—Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/01—Indexing scheme relating to G06F3/01
- G06F2203/012—Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F2203/00—Indexing scheme relating to G06F3/00 - G06F3/048
- G06F2203/048—Indexing scheme relating to G06F3/048
- G06F2203/04802—3D-info-object: information is displayed on the internal or external surface of a three dimensional manipulable object, e.g. on the faces of a cube that can be rotated by the user
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
Landscapes
- Engineering & Computer Science (AREA)
- General Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- User Interface Of Digital Computer (AREA)
- Processing Or Creating Images (AREA)
Abstract
The embodiment of the present application discloses a kind of space path method and system for planning, is related to intelligent terminal technical field.Methods described includes:Start the camera device of terminal;Obtain depth and width, the height in the first space, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan the space path of first spatial model.The space path method and system for planning of the application, by obtaining the object in space and space, it is determined that identification rigid body and record figure layer, to build spatial model, so as to planning space path, object movement is got around the space object of real world, improve entertainment effect, the experience in the augmented reality world.
Description
Technical field
The application is related to intelligent terminal technical field, more particularly to space path method and system for planning.
Background technology
At present, augmented reality AR (Augmented Reality) is used as a kind of by real world information and virtual generation
The technology that boundary's information is " seamless " to be integrated, by computer photograph model, real environment and virtual object are superimposed to together in real time
One picture or space.Wherein, Google glass is as one of typical AR equipment, by Trackpad, camera and light-emitting diode display
Combine, user can be networked, and the service such as map, Email is used in the visual field.
AR possesses three principal characters, and first, fusion is virtual and reality:Different from virtual reality technology VR (Virtual
Reality), AR technologies will not isolate user with real world, but the dummy object and/or information that computer is generated
It is superimposed in the scene of real world, intuitively in depth to understand reality scene, in limited time and limited scene
Understand real related knowledge domain.Enhancement information can include the non-geometry information related to real-world object, such as video, word
Deng;And/or geological information, such as virtual three-dimensional body/scene.Second, real-time, interactive:Pass through the interaction of augmented reality system
Interface equipment, user interact operation with natural way and augmented reality environment, and this interaction meets real-time.3rd, three
Dimension registration:" registration " (tracking and positioning) is to be corresponded dummy object caused by computer with true environment, and
When user moves in true environment, correct aligned relationship is also persistently maintained.
The AR applications of mobile terminal (for example, mobile phone), mainly carry out live preview, by mobile phone by the camera device of mobile phone
Local or network image are combined display with camera picture.In addition, in other applications, user by click on mobile phone screen or
External device carries out man-machine interaction, real-time update display content.For example, Vuforia softwares are a quick external members for realizing AR,
It can quickly realize that AR is applied by the software.For example, when the plane picture of identification " prefabricated rule ", mobile phone shoots the plan
After piece, it can be shown by matching default 3D models, wherein, " 3D toy books " is one of most classical product.But
The object that Vuforia is created can not pass through the mobile space object for getting around real world, such as desk.
Accordingly, it is desired to provide a kind of space path method and system for planning, by obtaining the object in space and space,
It is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, object movement is set to get around existing
The space object in the real world, improve entertainment effect, the experience in the augmented reality world.
The content of the invention
According to the first aspect of some embodiments of the present application, there is provided a kind of space path planing method, applied to end
Hold in (for example, electronic equipment etc.), methods described can include:Start the camera device of terminal;Obtain the first space depth,
Width, height, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;The i-th figure layer Li is recorded, including
kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan first spatial mode
The space path of type.
In certain embodiments, methods described may further include:If it is not, in i+1 deep space, it is determined that identify
Rigid body set Bi+1;Record i+1 space figure layer Li+1, including ki+1Individual rigid body.
In certain embodiments, the rigid body set for determining identification further comprises:Identified by the camera device
The collection of objects in first space, the object include rigid body, define object;Obtain the object that user clicks on;It is it is determined that described
The object that user clicks on is rigid body.
In certain embodiments, methods described may further include:Obtain terminal preset rules or end users operation;
According to the terminal preset rules or end users operation, it is determined that the definition collection of objects of identification.
In certain embodiments, the preset thickness includes deep space thickness T, rigid body thickness TB;The rigid body thickness
No more than the deep space thickness.
In certain embodiments, methods described may further include:The depth D in first space is obtained, depth is empty
Between thickness T;Determine that first space includes D/T deep space;Obtain the figure layer set in first space, the figure layer
Set includes D/T figure layer corresponding to D/T deep space.
In certain embodiments, the first spatial model of the structure further comprises:Scaled using three dimensions, by described in
Figure layer set is converted into coffin.
In certain embodiments, the first spatial model of the structure further comprises:According to deep space thickness T, superposition
The figure layer set.
In certain embodiments, methods described may further include:Generate virtual 3D objects, the virtual 3D object bags
Include and actual object image and/or imaginary subject image is presented;The virtual 3D objects move according to the space path, the sky
Between path include getting around the path of the rigid body.
According to the second aspect of some embodiments of the present application, there is provided a system, including:One memory, by with
It is set to data storage and instruction;One is established the processor to communicate with memory, wherein, when performing the instruction in memory,
The processor is configured as:Start the camera device of terminal;Obtain depth and width, the height in the first space, and default thickness
Degree;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including ki rigid body;Judge whether to travel through
The depth in the first space;If so, the first spatial model of structure;Plan the space path of first spatial model.
Therefore, according to the space path method and system for planning of some embodiments of the present application, by obtain space and
Object in space, it is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, make object
Movement can get around the space object of real world, improve entertainment effect, the experience in the augmented reality world.
Brief description of the drawings
To more fully understand and illustrating some embodiments of the present application, below with reference to the description of accompanying drawing reference implementation example,
In the drawings, same digital number indicates corresponding part in the accompanying drawings.
Fig. 1 is the illustrative diagram of the Environment System provided according to some embodiments of the present application.
Fig. 2 is the exemplary cell schematic diagram that the electronic functionalities provided according to some embodiments of the present application configure.
Fig. 3 is the exemplary process diagram of the space path planing method provided according to some embodiments of the present application.
Fig. 4 is the illustrative diagram that the space path provided according to some embodiments of the present application is planned.
Embodiment
Below with reference to accompanying drawing description for ease of Integrated Understanding the application as defined in claim and its equivalent
Various embodiments.These embodiments include various specific details in order to understand, but these be considered only as it is exemplary.Cause
This, it will be appreciated by those skilled in the art that carrying out variations and modifications without departing from this to various embodiments described here
The scope and spirit of application.In addition, briefly and to be explicitly described the application, the application will be omitted to known function and structure
Description.
The term and phrase used in description below and claims is not limited to literal meaning, and be merely can
Understand and as one man understand the application.Therefore, for those skilled in the art, it is possible to understand that, there is provided to the various implementations of the application
The description of example is only the purpose to illustrate, rather than limitation appended claims and its application of Equivalent definitions.
Below in conjunction with the accompanying drawing in the application some embodiments, the technical scheme in the embodiment of the present application is carried out clear
Chu, it is fully described by, it is clear that described embodiment is only some embodiments of the present application, rather than whole embodiments.
Based on the embodiment in the application, those of ordinary skill in the art are obtained all under the premise of creative work is not made
Other embodiment, belong to the scope of the application protection.
It should be noted that the term used in the embodiment of the present application is only merely for the mesh of description specific embodiment
, and it is not intended to be limiting the application." one " of singulative used in the embodiment of the present application and appended claims,
"one", " one kind ", " described " and "the" be also intended to including most forms, unless context clearly shows that other implications.Also
It should be appreciated that term "and/or" used herein refers to and list items purposes comprising one or more mutually bindings are any
Or it is possible to combine.Expression " first ", " second ", " described the first " and " described the second " be used for modify respective element without
Consideration order or importance, are used only for distinguishing a kind of element and another element, without limiting respective element.
Terminal according to the application some embodiments can be electronic equipment, the electronic equipment can include smart mobile phone,
PC (PC, such as tablet personal computer, desktop computer, notebook, net book, palm PC PDA), mobile phone, e-book
Reader, portable media player (PMP), audio/video player (MP3/MP4), video camera, virtual reality device
And one or more of combinations in wearable device etc. (VR).According to some embodiments of the present application, the wearable device
Type of attachment (such as wrist-watch, ring, bracelet, glasses or wear-type device (HMD)), integrated type (such as electronics can be included
Clothes), decorated type (such as pad skin, tatoo or built in electronic device) etc., or several combination.In some realities of the application
Apply in example, the electronic equipment can be flexible, be not limited to the said equipment, or can be one kind in above-mentioned various equipment
Or several combination.In this application, term " user " can be indicated using the people of electronic equipment or setting using electronic equipment
Standby (such as artificial intelligence electronic equipment).
The embodiment of the present application provides a kind of space path planing method.For the ease of understanding the embodiment of the present application, below
The embodiment of the present application is described in detail refer to the attached drawing.
Fig. 1 is the illustrative diagram of the Environment System 100 provided according to some embodiments of the present application.Such as Fig. 1
Shown, Environment System 100 can include electronic equipment 110, network 120 and server 130 etc..Electronic equipment 110 can be with
Including bus 111, processor 112, memory 113, input/output module 114, display 115, communication module 116 and physics
Key 117 etc..In some embodiments of the present application, electronic equipment 110 can omit one or more elements, or can enter one
Step includes one or more of the other element.
Bus 111 can include circuit.The circuit can be with one or more element (examples in interconnection electronics 110
Such as, bus 111, processor 112, memory 113, input/output module 114, display 115, communication module 116 and secondary or physical bond
117).The circuit can also be realized between one or more elements in electronic equipment 110 communication (for example, obtain and/or
Send information).
Processor 112 can include one or more coprocessors (Co-processor), application processor (AP,
Application Processor) and communication processor (Communication Processor).As an example, processor
112 can perform with the control of one or more elements of electronic equipment 110 and/or data processing (for example, starting shooting dress
Put, identify the operations such as object, record figure layer, structure spatial model, planning space path).
Memory 113 can be with data storage.The data can include other with one or more of electronic equipment 110
The related instruction of element or data.For example, the data can include the initial data of the before processing of processor 112, intermediate data
And/or the data after processing.Memory 113 can include impermanent memory memory and/or permanent memory memory.Make
For example, memory 113 can be with the collection of objects of memory space, the map data mining platform recorded etc..
According to some embodiments of the present application, memory 113 can store software and/or program.Described program can wrap
Include kernel, middleware, API (API, Application Programming Interface) and/or using journey
Sequence (or " application ").
At least a portion of the kernel, the middleware or the API can include operating system (OS,
Operating System).As an example, the kernel can be controlled or managed for performing other programs (in for example,
Between part, API and application program) in realize operation or function system resource (for example, bus 111, processor
112nd, memory 113 etc.).In addition, the kernel can provide interface.The interface can by the middleware, it is described should
One or more elements of electronic equipment 110 are accessed with DLL or the application program to control or management system resource.
The middleware can be as the intermediate layer of data transfer.The data transfer can allow API or
Application program is with the kernel communication exchanging data.As an example, the middleware can be handled from the application program
One or more task requests of acquisition.For example, the middleware can be to one or more application assigned electronic equipments
The priority of 110 system resource (for example, bus 111, processor 112, memory 113 etc.), and processing it is one or
Multiple tasks are asked.The API can be that the application program is used to control from the kernel or the middleware
The interface of function is provided.The API can also include one or more interfaces or function (for example, instruction).It is described
Function can be used for starting control, data channel control, security control, Control on Communication, document control, window control, text control
System, image procossing, information processing etc..
Input/output module 114 can send what is inputted from user or external equipment to the other elements of electronic equipment 110
Instruction or data.Input/output module 114 can also be defeated by the instruction of the other elements acquisition from electronic equipment 110 or data
Go out to user or external equipment.In certain embodiments, input/output module 114 can include input block, and user can lead to
Cross the input block input information or instruction.
Display 115 can be with display content.The content can to user show all kinds (for example, text, image,
Video, icon and/or symbol etc., or several combinations).Display 115 can include liquid crystal display (LCD, Liquid
Crystal Display), light emitting diode (LED, Light-Emitting Diode) display, Organic Light Emitting Diode
(OLED, Organic Light Emitting Diode) display, Micro Electro Mechanical System (MEMS, Micro Electro
Mechanical Systems) display or electric paper display etc., or several combinations.Display 115 can include display
Screen, touch-screen etc..The display screen can show the 3d space model of structure, display space path etc..In certain embodiments,
Display 115 can show virtual key.The touch-screen can obtain the input of the virtual key.Display 115 can pass through
The touch-screen obtains input.The input can include touch input, gesture input, action input, close input, electronic pen
Or the input (for example, hovering input) of user's body part.
Communication module 116 can configure the communication between equipment.In certain embodiments, Environment System 100 can be with
Further comprise electronic equipment 140.As an example, the communication between the equipment can include electronic equipment 110 and other set
Communication between standby (for example, server 130 or electronic equipment 140).For example, communication module 116 can by radio communication or
Wire communication is connected to network 120, realizes and communicates with other equipment (for example, server 130 or electronic equipment 140).
The radio communication can include microwave communication and/or satellite communication etc..The radio communication can include honeycomb
Communication is (for example, global mobile communication (GSM, Global System for Mobile Communications), CDMA
(CDMA, Code Division MultipleAccess), 3G (Third Generation) Moblie (3G, The 3rd Generation
Telecommunication), forth generation mobile communication (4G), the 5th third-generation mobile communication (5G), Long Term Evolution (LTE,
Long Term Evolution), Long Term Evolution upgrade version (LTE-A, LTE-Advanced), WCDMA
(WCDMA, Wideband Code Division MultipleAccess), UMTS (UMTS, Universal
Mobile Telecommunications System), WiMAX (WiBro, Wireless Broadband) etc., it is or several
Combination.According to some embodiments of the present application, the radio communication can include WLAN (WiFi, Wireless
Fidelity), bluetooth, low-power consumption bluetooth (BLE, Bluetooth Low Energy), ZigBee protocol (ZigBee), near-field communication
(NFC, Near Field Communication), magnetic safe transmission, radio frequency and body area network (BAN, BodyAreaNetwork)
Deng, or several combinations.According to some embodiments of the present application, the wire communication can include GLONASS
(Glonass/GNSS, Global Navigation Satellite System), global positioning system (GPS, Global
Position System), Beidou navigation satellite system or Galileo (European GPS) etc..The cable modem
Letter can include USB (USB, Universal Serial Bus), high-definition media interface (HDMI, High-
Definition Multimedia Interface), proposed standard 232 (RS-232, Recommend Standard 232),
And/or plain old telephone service (POTS, Plain Old Telephone Service) etc., or several combinations.
Secondary or physical bond 117 can be used for user mutual.Secondary or physical bond 117 can include one or more entity keys.In some realities
Apply in example, user can be with the function of self-defined secondary or physical bond 117.As an example, secondary or physical bond 117 can send instruction.The instruction
It can include starting iris read write command, start-up space path planning etc..The iris read write command can include iris verification
When read iris instruction, iris typing when write-in iris instruction etc..
In certain embodiments, electronic equipment 110 may further include sensor.The sensor can be included but not
It is limited to light sensor, acoustic sensor, gas sensor, chemical sensor, voltage sensitive sensor, temp-sensitive sensor, fluid to pass
Sensor, biology sensor, laser sensor, Hall sensor, intelligence sensor etc., or several combinations.
In certain embodiments, electronic equipment 110 may further include infrared equipment, image capture device etc..As
Example, the infrared equipment can identify by infrared ray mode of delivery, and blink, watch the technical limit spacing eyes such as identification attentively
Information.For example, the infrared equipment is acted come certification user profile by gathering the blink of user.As an example, described image
Collecting device can include camera, iris device etc..The camera can realize the functions such as eyeball tracking.The iris dress
Authentication (for example, certification user profile) can be carried out using iris recognition technology by putting.The iris device can include rainbow
Film camera, the iris camera can obtain iris information, and the iris information can be stored in memory 113.
Network 120 can include communication network.The communication network can include computer network (for example, LAN
(LAN, Local Area Network) or wide area network (WAN, Wide Area Network)), internet and/or telephone network
Deng, or several combinations.Network 120 can be to the other equipment in Environment System 100 (for example, electronic equipment 110, clothes
Business device 130, electronic equipment 140 etc.) send information.
Server 130 can be by the other equipment in the connection Environment System 100 of network 120 (for example, electronic equipment
110th, electronic equipment 140 etc.).In certain embodiments, when electronic equipment 110 is lost, server 130 can pass through network
120 send startup iris read write command to electronic equipment;When iris verification fails, server 130 can be entered by network 120
One step locks electronic equipment 110.In certain embodiments, server 130 can send the virtual object of structure to electronic equipment 110
Body information etc..
Electronic equipment 140 can be identical or different with electronic equipment 110 type.According to some embodiments of the present application,
The part or all of operation performed in electronic equipment 110 can be in another equipment or multiple equipment (for example, electronic equipment 140
And/or server 130) in perform.In certain embodiments, when electronic equipment 110 be automatically or in response to request perform it is a kind of or
When multiple functions and/or service, electronic equipment 110 can ask other equipment (for example, electronic equipment 140 and/or server
130) perform function and/or service are substituted.In certain embodiments, electronic equipment 110 is in addition to perform function or service, further
Perform relative one or more functions.In certain embodiments, other equipment is (for example, electronic equipment 140 and/or clothes
Business device 130) asked function or other related one or more functions can be performed, implementing result can be sent to electricity
Sub- equipment 110.Electronic equipment 110 can repeat result or further handle implementing result, to provide asked function
Or service.As an example, electronic equipment 110 can use cloud computing, distributed computing technology and/or client-server end to calculate meter
Calculate etc., or several combinations.In certain embodiments, can be included according to the difference of cloud computing service property, the cloud computing
Public cloud, private clound and mixed cloud etc..In certain embodiments, when electronic equipment 110 is lost, electronic equipment 140 can be to
Electronic equipment 110 sends positioning instruction, to obtain the positional information of electronic equipment 110.In certain embodiments, electronic equipment
110 can be with the communal space information of electronic equipment 140.
It should be noted that the description for Environment System 100 above, only for convenience of description, can not be this Shen
It please be limited within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system can
Each element can be combined on the premise of without departing substantially from the principle, or forms subsystem and be connected with other elements,
To implementing the various modifications and variations on the above method and systematic difference field progress form and details.For example, network environment
System 100 may further include database.In another example electronic equipment 110 can not include secondary or physical bond 117 etc..It is all such
The deformation of class, within the protection domain of the application.
Fig. 2 is the exemplary cell block diagram that the electronic functionalities provided according to some embodiments of the present application configure.Such as
Shown in Fig. 2, processor 112 can include processing module 200, and the processing module 200 can include acquiring unit 210, control
Unit 220, determining unit 230, generation unit 240, processing unit 250.
According to some embodiments of the present application, acquiring unit 210 can obtain information.In certain embodiments, the letter
Breath can include but is not limited to text, picture, audio, video, action, gesture, sound, eyes (for example, iris information etc.), gas
Breath, light etc., or several combinations.In certain embodiments, described information can include but is not limited to input information, system information
And/or communication information etc..As an example, acquiring unit 210 can pass through the touch of input/output module 114, display 115
Screen, secondary or physical bond 117 and/or sensor obtain the input information of electronic equipment 110.The input information can include other equipment
(for example, electronic equipment 140) and/or the input of user, for example, the input of key-press input, touch-control, gesture input, action input, remote
Journey input, transmission input, eyes input, sound input, breath input, light input etc., or several combinations.The input information
Obtaining widget can include but is not limited to infrared equipment, image capture device, sensor etc., or several combinations.As showing
Example, acquiring unit 210 can obtain spatial information, collection of objects etc. by image capture device (for example, camera device etc.).Institute
Rigid body (Rigid Body) can be included, define object etc. by stating object.
In certain embodiments, acquiring unit 210 can obtain the communication information by network 120.The communication information can
With including application software information, communication signal (for example, voice signal, vision signal etc.), short message etc..In some embodiments
In, acquiring unit 210 can obtain system information by network 120, memory 113 and/or sensor.The system information can
With include but is not limited to the system mode of electronic equipment 110, presupposed information, memory 113 store information (for example, iris is recognized
Demonstrate,prove information etc.) etc., or several combinations.As an example, acquiring unit 210 can obtain preset thickness information.The default thickness
Degree information can include the thickness of deep space, the thickness of rigid body, thickness of figure layer etc..
In certain embodiments, described information can include instruction.The instruction includes user instruction and/or system command
Deng, or several combinations.The instruction can include triggering command, certification instruction, fill in instruction etc., or several combinations.Institute
Certification user profile can be included, start structure spatial model instruction etc. by stating certification instruction.As an example, if iris information is verified
During failure, electronic equipment 110 can start user information authentication etc..In another example acquiring unit 210 can obtain startup structure sky
Between model instruction.
According to some embodiments of the present application, control unit 220 can be with control electronics.In certain embodiments, control
Unit 220 processed can start camera device (for example, it is first-class to start shooting), start structure spatial model etc..The camera
Can be first-class including black and white camera (Mono), colour imagery shot, iris shooting.In certain embodiments, control unit 220 can
With the camera device for the instruction startup terminal for building spatial model according to starting.In certain embodiments, control unit 220 can be with
Record map data mining platform.
According to some embodiments of the present application, determining unit 230 can determine information.In certain embodiments, it is it is determined that single
Member 230 can determine the object identified by camera device.The object can include rigid body (RigidBody), define object.
The rigid body can include the object that can not be passed through.The object for defining object and can including passing through, for example, non-firm
Body.In certain embodiments, determining unit 230 can determine whether flow travels through the depth in space.
According to some embodiments of the present application, generation unit 240 can generate information.In certain embodiments, generation is single
Member 240 can build spatial model etc..As an example, generation unit 240 can be according to spatial information, collection of objects and figure layer
Information etc. builds 3d space model.In certain embodiments, generation unit 240 can generate virtual 3D objects, the virtual 3D
Object includes actual object image and/or imaginary subject image is presented, for example, the virtual 3D rendering of terminal user or game are fabricated
Virtual 3D rendering etc..
According to some embodiments of the present application, processing unit 250 can be with processing information.In certain embodiments, processing is single
Member 250 can utilize three dimensions to scale, and figure layer set is converted into coffin.In certain embodiments, processing unit 250
Figure layer set can be superimposed according to figure layer thickness.The figure layer set can include multiple figure layers corresponding to multiple deep spaces.
In certain embodiments, processing unit 250 can delete the definition object of virtual 3D objects contact.In certain embodiments, locate
Managing unit 250 can be according to spatial model planning space path.As an example, the space path can bypass the space
Rigid body.
It should be noted that described above for the unit in processing module 200, only for convenience of description, can not be this
Application is limited within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system,
Unit may be combined on the premise of without departing substantially from the principle, or form submodule and connect with other units
Connect, the various modifications and variations in form and details are carried out to the function of implementing above-mentioned module and unit.For example, processing module
200 may further include memory cell, and the memory cell can be with storage space information, record map data mining platform etc..It is all such
The deformation of class, within the protection domain of the application.
Fig. 3 is the exemplary process diagram of the space path planing method provided according to some embodiments of the present application.Such as Fig. 3
Shown, flow 300 can be realized by processing module 200.In certain embodiments, the space path planing method can be certainly
It is dynamic to start or started by instructing.The automatic start can be when opening application program, the planning of automatic start space path.Institute
User instruction, system command, action command etc., or several combinations can be included by stating instruction.As an example, the system command
The information that can be obtained by sensor generates.The user instruction can include voice, gesture, action, secondary or physical bond 117 and/
Or virtual key etc., or several combinations.As an example, when user starts camera device, it can directly initiate or pass through pressing
Secondary or physical bond start-up space path planning.
301, start the camera device of terminal.Operation 301 can be realized by the control unit 220 of processing module 200.
In certain embodiments, control unit 220 can start camera device.
302, the depth D, width, height in the first space, and preset thickness are obtained.Operation 302 can pass through processing
The acquiring unit 210 of module 200 is realized.In certain embodiments, acquiring unit 210 can pass through the first of camera device identification
Spatial information.The spatial information includes depth and width, height etc..As an example, the depth in first space can be D.
In certain embodiments, acquiring unit 210 can obtain preset thickness.The preset thickness can include deep space thickness,
Rigid body thickness, define object thickness, figure layer thickness etc..
303, in the i-th deep space, it is determined that the rigid body set Bi identified.Operation 303 can pass through processing module 200
Determining unit 230, control unit 220 are realized.In certain embodiments, determining unit 230 can determine camera device identification
The rigid body set Bi of i-th deep space in the first space.In certain embodiments, the camera device can identify the first space
Collection of objects, the collection of objects includes rigid body set, defines collection of objects etc..As an example, determining unit 230 can lead to
Cross acquiring unit 210 and obtain the object that user clicks on, determine that the object that the user clicks on is rigid body.For example, the rigid body collection
Multiple rigid bodies that user clicks in deep space can be included by closing, or in multiple deep spaces user click on it is multiple just
Body.In certain embodiments, first space can divide N number of deep space according to different deep space thickness.Example
Such as, when deep space thickness is T, N=D/T.I-th deep space can be any one depth of N number of deep space
Spend space, i=1,2 ..., N.
In 304, the i-th figure layer Li of record, including kiIndividual rigid body:Bi1, Bi2... ..., Bik.Operation 304 can be by handling mould
The control unit 220 of block 200 is realized.In certain embodiments, control unit 220 can be by the i-th deep space and its rigid body collection
Close Bi and be recorded as the i-th figure layer Li.The i-th figure layer Li includes kiIndividual rigid body.
305, judge whether to travel through the depth in the first space.Operation 305 can pass through the determining unit of processing module 200
230 realize.In certain embodiments, determining unit 230 can determine flow 300 whether the depth in the first space.As an example,
When the first space includes N number of deep space, determining unit 230 may determine whether from i=1 the first deep space travel through to
I=N N deep spaces.
If not traveling through the depth in the first space, operation 303 is returned to, 303, in the i-th deep space, it is determined that the rigid body identified
Set Bi.Operation 302 can be realized by determining unit 230, the control unit 220 of processing module 200.In some embodiments
In, if traveled through i-th of deep space, i=i+1, that is, when returning to operation 303,303, in i+1 deep space, it is determined that
The rigid body set B of identificationi+1.In certain embodiments, the terminal can be moved to D depth from depth zero, travel through the first space
Depth.In certain embodiments, control unit 220 can adjust the camera device, from depth zero to D extreme saturations first
The depth in space.
If having traveled through the depth in the first space, into operation 306,306, the first spatial model is built.Operation 306 can be with
Realized by the generation unit 240 of processing module 200.In certain embodiments, generation unit 240 can be believed according to the first space
Breath, N number of deep space and its rigid body set B, the map data mining platform of record etc. in the first space build the first spatial model.At some
In embodiment, the rigid body thickness can be not more than the deep space thickness.In certain embodiments, processing unit 250 can
To be scaled using three dimensions, figure layer set is converted into coffin;Generation unit 240 can according to the coffin,
Build 3d space model.In certain embodiments, generation unit 240 can be superimposed figure layer set, structure the according to figure layer thickness
One spatial model.The figure layer thickness can be identical with the deep space thickness.The figure layer set can include and N number of depth
Spend the corresponding N number of figure layer in space.As an example, figure layer set L={ L1, L2..., LN}。
307, the space path of first spatial model is planned.Operation 307 can pass through the processing of processing module 200
Unit 250 is realized.In certain embodiments, processing unit 250 according to the first spatial model, can plan the space in the first space
Path.The space path can get around the rigid body set B in first space.
It should be noted that the description for flow 300 above, only for convenience of description, can not be limited in the application
Within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system, may not carry on the back
On the premise of from the principle, each operation is combined, or forms sub-process and other operative combinations, in implementation
State the various modifications and variations in flow and the function progress form and details of operation.For example, flow 300 may further include
It is determined that define the operations such as collection of objects, the virtual 3D objects of generation.Such deformation, the application protection domain it
It is interior.
Fig. 4 is the illustrative diagram that the space path provided according to some embodiments of the present application is planned.Such as Fig. 4 institutes
Show, be the top view in space 400, the spatial information in space 400 includes depth D.As an example, the deep space in space 400 is thick
Spend and correspond to figure layer L for T, the first deep space1, including rigid body B11, define object O11.Wherein, D=2T, space 400 include N number of
Deep space, N=D/T=2.The energy and disaster in space 400 corresponds to figure layer L2, including rigid body B21With rigid body B22.It is described firm
Body thickness is TB, TB≤T.In certain embodiments, the rigid body thickness can be different from actual rigid body thickness, space 400 it is firm
Body set B rigid body thickness can be with identical.In certain embodiments, the space path of planning can include the first space path
401 and second space path 402.As an example, the rigid body can be the object that can not be passed through, the object that defines can be
The object that can be passed through, first space path 401 can bypass rigid body B11With rigid body B21.The second space path 402
Rigid body B can be bypassed11, connection definition object O11, further around rigid body B excessively22。
In summary, according to the space path method and system for planning of the embodiment of the present application, by obtaining space and sky
In object, it is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, move object
Kinetic energy enough gets around the space object of real world, improves entertainment effect, the experience in the augmented reality world.
It should be noted that the above embodiments are intended merely as example, the application is not limited to such example, but can
To carry out various change.
It should be noted that in this manual, term " comprising ", "comprising" or its any other variant are intended to
Nonexcludability includes, so that process, method, article or equipment including a series of elements not only will including those
Element, but also the other element including being not expressly set out, or it is this process, method, article or equipment also to include
Intrinsic key element.In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that
Other identical element also be present in process, method, article or equipment including the key element.
Finally, it is to be noted that, a series of above-mentioned processing are not only included with order described here in temporal sequence
The processing of execution, and the processing including performing parallel or respectively rather than in chronological order.
One of ordinary skill in the art will appreciate that realize all or part of flow in above-described embodiment method, being can be with
To be completed by the related hardware of computer program instructions, described program can be stored in a computer-readable recording medium,
The program is upon execution, it may include such as the flow of the embodiment of above-mentioned each method.Wherein, described storage medium can be magnetic disc,
CD, read-only storage (Read-Only Memory, ROM) or random access memory (Random Access Memory, RAM)
Deng.
Above disclosed is only some preferred embodiments of the application, it is impossible to the right model of the application is limited with this
Enclose, one of ordinary skill in the art will appreciate that all or part of flow of above-described embodiment is realized, and will according to the application right
Made equivalent variations are sought, still falls within and invents covered scope.
Claims (10)
- A kind of 1. space path planing method, it is characterised in that including:Start the camera device of terminal;Obtain depth and width, the height in the first space, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan the space path of first spatial model.
- 2. space path planing method according to claim 1, it is characterised in that further comprise:If it is not, in i+1 deep space, it is determined that the rigid body set B identifiedi+1;Record i+1 space figure layer Li+1, including ki+1Individual rigid body.
- 3. space path planing method according to claim 1, it is characterised in that the rigid body set for determining identification is entered One step includes:The collection of objects in first space is identified by the camera device, the object includes rigid body, defines object;Obtain the object that user clicks on;Determine that the object that the user clicks on is rigid body.
- 4. space path planing method according to claim 3, it is characterised in that further comprise:Obtain terminal preset rules or end users operation;According to the terminal preset rules or end users operation, it is determined that the definition collection of objects of identification.
- 5. space path planing method according to claim 1, it is characterised in that the preset thickness includes deep space Thickness T, rigid body thickness TB;The rigid body thickness is not more than the deep space thickness.
- 6. space path planing method according to claim 5, it is characterised in that further comprise:Obtain the depth D in first space, deep space thickness T;Determine that first space includes D/T deep space;The figure layer set in first space is obtained, the figure layer set includes D/T figure layer corresponding to D/T deep space.
- 7. space path planing method according to claim 6, it is characterised in that the first spatial model of the structure enters one Step includes:Scaled using three dimensions, the figure layer set is converted into coffin.
- 8. space path planing method according to claim 6, it is characterised in that the first spatial model of the structure enters one Step includes:According to deep space thickness T, the figure layer set is superimposed.
- 9. space path planing method according to claim 1, it is characterised in that further comprise:Virtual 3D objects are generated, the virtual 3D objects include actual object image and/or imaginary subject image is presented;The virtual 3D objects move according to the space path, and the space path includes getting around the path of the rigid body.
- A 10. system, it is characterised in that including:One memory, is configured as data storage and instruction;One is established the processor to communicate with memory, wherein, when performing the instruction in memory, the processor is configured For:Start the camera device of terminal;Obtain depth and width, the height in the first space, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan the space path of first spatial model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710754154.3A CN107506040A (en) | 2017-08-29 | 2017-08-29 | A kind of space path method and system for planning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710754154.3A CN107506040A (en) | 2017-08-29 | 2017-08-29 | A kind of space path method and system for planning |
Publications (1)
Publication Number | Publication Date |
---|---|
CN107506040A true CN107506040A (en) | 2017-12-22 |
Family
ID=60693021
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710754154.3A Pending CN107506040A (en) | 2017-08-29 | 2017-08-29 | A kind of space path method and system for planning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107506040A (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109935318A (en) * | 2019-03-06 | 2019-06-25 | 珠海市万瑙特健康科技有限公司 | Display methods, device, computer equipment and the storage medium of three-dimensional pulse wave |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101726296A (en) * | 2009-12-22 | 2010-06-09 | 哈尔滨工业大学 | Vision measurement, path planning and GNC integrated simulation system for space robot |
CN101770235A (en) * | 2009-01-01 | 2010-07-07 | 索尼株式会社 | Path planning device, path planning method, and computer program |
CN101907459A (en) * | 2010-07-12 | 2010-12-08 | 清华大学 | Monocular video based real-time posture estimation and distance measurement method for three-dimensional rigid body object |
CN103247075A (en) * | 2013-05-13 | 2013-08-14 | 北京工业大学 | Variational mechanism-based indoor scene three-dimensional reconstruction method |
CN103649996A (en) * | 2011-05-10 | 2014-03-19 | 皇家飞利浦有限公司 | User-steered on-the-fly path planning |
CN103761737A (en) * | 2014-01-22 | 2014-04-30 | 北京工业大学 | Robot motion estimation method based on dense optical flow |
CN105122304A (en) * | 2012-11-14 | 2015-12-02 | 微软技术许可有限责任公司 | Real-time design of living spaces with augmented reality |
CN105143907A (en) * | 2013-04-22 | 2015-12-09 | 阿尔卡特朗讯 | Localization systems and methods |
CN105320140A (en) * | 2015-12-01 | 2016-02-10 | 浙江宇视科技有限公司 | Robot cleaner and cleaning path planning method thereof |
CN105493155A (en) * | 2013-08-30 | 2016-04-13 | 高通股份有限公司 | Method and apparatus for representing physical scene |
CN106052674A (en) * | 2016-05-20 | 2016-10-26 | 青岛克路德机器人有限公司 | Indoor robot SLAM method and system |
CN106127739A (en) * | 2016-06-16 | 2016-11-16 | 华东交通大学 | A kind of RGB D SLAM method of combination monocular vision |
CN106197263A (en) * | 2016-06-28 | 2016-12-07 | 大连理工大学 | Many bodies based on machine vision throw in pose measuring method |
CN106643692A (en) * | 2016-09-28 | 2017-05-10 | 深圳乐行天下科技有限公司 | Robot navigation and positioning method, system and robot |
CN107003728A (en) * | 2014-11-25 | 2017-08-01 | 微软技术许可有限责任公司 | Virtual measurement instrument for wearable visualization device |
-
2017
- 2017-08-29 CN CN201710754154.3A patent/CN107506040A/en active Pending
Patent Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101770235A (en) * | 2009-01-01 | 2010-07-07 | 索尼株式会社 | Path planning device, path planning method, and computer program |
CN101726296A (en) * | 2009-12-22 | 2010-06-09 | 哈尔滨工业大学 | Vision measurement, path planning and GNC integrated simulation system for space robot |
CN101907459A (en) * | 2010-07-12 | 2010-12-08 | 清华大学 | Monocular video based real-time posture estimation and distance measurement method for three-dimensional rigid body object |
CN103649996A (en) * | 2011-05-10 | 2014-03-19 | 皇家飞利浦有限公司 | User-steered on-the-fly path planning |
CN105122304A (en) * | 2012-11-14 | 2015-12-02 | 微软技术许可有限责任公司 | Real-time design of living spaces with augmented reality |
CN105143907A (en) * | 2013-04-22 | 2015-12-09 | 阿尔卡特朗讯 | Localization systems and methods |
CN103247075A (en) * | 2013-05-13 | 2013-08-14 | 北京工业大学 | Variational mechanism-based indoor scene three-dimensional reconstruction method |
CN105493155A (en) * | 2013-08-30 | 2016-04-13 | 高通股份有限公司 | Method and apparatus for representing physical scene |
CN103761737A (en) * | 2014-01-22 | 2014-04-30 | 北京工业大学 | Robot motion estimation method based on dense optical flow |
CN107003728A (en) * | 2014-11-25 | 2017-08-01 | 微软技术许可有限责任公司 | Virtual measurement instrument for wearable visualization device |
CN105320140A (en) * | 2015-12-01 | 2016-02-10 | 浙江宇视科技有限公司 | Robot cleaner and cleaning path planning method thereof |
CN106052674A (en) * | 2016-05-20 | 2016-10-26 | 青岛克路德机器人有限公司 | Indoor robot SLAM method and system |
CN106127739A (en) * | 2016-06-16 | 2016-11-16 | 华东交通大学 | A kind of RGB D SLAM method of combination monocular vision |
CN106197263A (en) * | 2016-06-28 | 2016-12-07 | 大连理工大学 | Many bodies based on machine vision throw in pose measuring method |
CN106643692A (en) * | 2016-09-28 | 2017-05-10 | 深圳乐行天下科技有限公司 | Robot navigation and positioning method, system and robot |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109935318A (en) * | 2019-03-06 | 2019-06-25 | 珠海市万瑙特健康科技有限公司 | Display methods, device, computer equipment and the storage medium of three-dimensional pulse wave |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107835365A (en) | Auxiliary shooting method and system | |
CN107657953A (en) | Sound control method and system | |
CN107609914A (en) | A kind of intelligent menu design method and system | |
CN108009140A (en) | A kind of end message edit methods and system | |
US11579744B2 (en) | Systems and methods for seat selection in virtual reality | |
CN107358179A (en) | A kind of living management system, medium and method based on iris verification | |
CN107423100A (en) | A kind of methods, devices and systems of terminal OTA firmware updatings | |
CN107423585A (en) | The concealed application method and system of a kind of application | |
CN107368793A (en) | A kind of colored method for collecting iris and system | |
CN105190469A (en) | Causing specific location of an object provided to a device | |
CN107786979A (en) | A kind of multiple terminals shared communication method and system | |
CN107862518A (en) | A kind of method of payment and system based on terminal location | |
CN107404484A (en) | A kind of method, apparatus and system of mechanical release authority | |
CN107220531A (en) | A kind of information processing method of convenient login | |
CN107018153A (en) | A kind of safe login method | |
CN107506040A (en) | A kind of space path method and system for planning | |
CN206908680U (en) | Intelligent terminal with secure log function | |
CN108154556A (en) | A kind of virtual trailing of terminal and system | |
CN107835117A (en) | A kind of instant communicating method and system | |
CN107402690A (en) | A kind of global collecting method and system | |
CN107277625A (en) | A kind of management method and system of video playback history | |
CN108897479A (en) | A kind of terminal touch control method and system | |
CN107592398A (en) | A kind of intelligent information storage method and system | |
CN109189536A (en) | A kind of terminal applies display methods and system | |
CN107623736A (en) | A kind of equipment connection method and system |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20171222 |