CN107506040A - A kind of space path method and system for planning - Google Patents

A kind of space path method and system for planning Download PDF

Info

Publication number
CN107506040A
CN107506040A CN201710754154.3A CN201710754154A CN107506040A CN 107506040 A CN107506040 A CN 107506040A CN 201710754154 A CN201710754154 A CN 201710754154A CN 107506040 A CN107506040 A CN 107506040A
Authority
CN
China
Prior art keywords
space
rigid body
thickness
layer
deep
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710754154.3A
Other languages
Chinese (zh)
Inventor
朱洪达
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Aiyouwei Software Development Co Ltd
Original Assignee
Shanghai Aiyouwei Software Development Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Aiyouwei Software Development Co Ltd filed Critical Shanghai Aiyouwei Software Development Co Ltd
Priority to CN201710754154.3A priority Critical patent/CN107506040A/en
Publication of CN107506040A publication Critical patent/CN107506040A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/011Arrangements for interaction with the human body, e.g. for user immersion in virtual reality
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/01Input arrangements or combined input and output arrangements for interaction between user and computer
    • G06F3/048Interaction techniques based on graphical user interfaces [GUI]
    • G06F3/0481Interaction techniques based on graphical user interfaces [GUI] based on specific properties of the displayed interaction object or a metaphor-based environment, e.g. interaction with desktop elements like windows or icons, or assisted by a cursor's changing behaviour or appearance
    • G06F3/04815Interaction with a metaphor-based environment or interaction object displayed as three-dimensional, e.g. changing the user viewpoint with respect to the environment or object
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/01Indexing scheme relating to G06F3/01
    • G06F2203/012Walk-in-place systems for allowing a user to walk in a virtual environment while constraining him to a given position in the physical environment
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2203/00Indexing scheme relating to G06F3/00 - G06F3/048
    • G06F2203/048Indexing scheme relating to G06F3/048
    • G06F2203/048023D-info-object: information is displayed on the internal or external surface of a three dimensional manipulable object, e.g. on the faces of a cube that can be rotated by the user
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/50Depth or shape recovery

Landscapes

  • Engineering & Computer Science (AREA)
  • General Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • User Interface Of Digital Computer (AREA)
  • Processing Or Creating Images (AREA)

Abstract

The embodiment of the present application discloses a kind of space path method and system for planning, is related to intelligent terminal technical field.Methods described includes:Start the camera device of terminal;Obtain depth and width, the height in the first space, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan the space path of first spatial model.The space path method and system for planning of the application, by obtaining the object in space and space, it is determined that identification rigid body and record figure layer, to build spatial model, so as to planning space path, object movement is got around the space object of real world, improve entertainment effect, the experience in the augmented reality world.

Description

A kind of space path method and system for planning
Technical field
The application is related to intelligent terminal technical field, more particularly to space path method and system for planning.
Background technology
At present, augmented reality AR (Augmented Reality) is used as a kind of by real world information and virtual generation The technology that boundary's information is " seamless " to be integrated, by computer photograph model, real environment and virtual object are superimposed to together in real time One picture or space.Wherein, Google glass is as one of typical AR equipment, by Trackpad, camera and light-emitting diode display Combine, user can be networked, and the service such as map, Email is used in the visual field.
AR possesses three principal characters, and first, fusion is virtual and reality:Different from virtual reality technology VR (Virtual Reality), AR technologies will not isolate user with real world, but the dummy object and/or information that computer is generated It is superimposed in the scene of real world, intuitively in depth to understand reality scene, in limited time and limited scene Understand real related knowledge domain.Enhancement information can include the non-geometry information related to real-world object, such as video, word Deng;And/or geological information, such as virtual three-dimensional body/scene.Second, real-time, interactive:Pass through the interaction of augmented reality system Interface equipment, user interact operation with natural way and augmented reality environment, and this interaction meets real-time.3rd, three Dimension registration:" registration " (tracking and positioning) is to be corresponded dummy object caused by computer with true environment, and When user moves in true environment, correct aligned relationship is also persistently maintained.
The AR applications of mobile terminal (for example, mobile phone), mainly carry out live preview, by mobile phone by the camera device of mobile phone Local or network image are combined display with camera picture.In addition, in other applications, user by click on mobile phone screen or External device carries out man-machine interaction, real-time update display content.For example, Vuforia softwares are a quick external members for realizing AR, It can quickly realize that AR is applied by the software.For example, when the plane picture of identification " prefabricated rule ", mobile phone shoots the plan After piece, it can be shown by matching default 3D models, wherein, " 3D toy books " is one of most classical product.But The object that Vuforia is created can not pass through the mobile space object for getting around real world, such as desk.
Accordingly, it is desired to provide a kind of space path method and system for planning, by obtaining the object in space and space, It is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, object movement is set to get around existing The space object in the real world, improve entertainment effect, the experience in the augmented reality world.
The content of the invention
According to the first aspect of some embodiments of the present application, there is provided a kind of space path planing method, applied to end Hold in (for example, electronic equipment etc.), methods described can include:Start the camera device of terminal;Obtain the first space depth, Width, height, and preset thickness;In the i-th deep space, it is determined that the rigid body set Bi identified;The i-th figure layer Li is recorded, including kiIndividual rigid body;Judge whether the depth in the first space of traversal;If so, the first spatial model of structure;Plan first spatial mode The space path of type.
In certain embodiments, methods described may further include:If it is not, in i+1 deep space, it is determined that identify Rigid body set Bi+1;Record i+1 space figure layer Li+1, including ki+1Individual rigid body.
In certain embodiments, the rigid body set for determining identification further comprises:Identified by the camera device The collection of objects in first space, the object include rigid body, define object;Obtain the object that user clicks on;It is it is determined that described The object that user clicks on is rigid body.
In certain embodiments, methods described may further include:Obtain terminal preset rules or end users operation; According to the terminal preset rules or end users operation, it is determined that the definition collection of objects of identification.
In certain embodiments, the preset thickness includes deep space thickness T, rigid body thickness TB;The rigid body thickness No more than the deep space thickness.
In certain embodiments, methods described may further include:The depth D in first space is obtained, depth is empty Between thickness T;Determine that first space includes D/T deep space;Obtain the figure layer set in first space, the figure layer Set includes D/T figure layer corresponding to D/T deep space.
In certain embodiments, the first spatial model of the structure further comprises:Scaled using three dimensions, by described in Figure layer set is converted into coffin.
In certain embodiments, the first spatial model of the structure further comprises:According to deep space thickness T, superposition The figure layer set.
In certain embodiments, methods described may further include:Generate virtual 3D objects, the virtual 3D object bags Include and actual object image and/or imaginary subject image is presented;The virtual 3D objects move according to the space path, the sky Between path include getting around the path of the rigid body.
According to the second aspect of some embodiments of the present application, there is provided a system, including:One memory, by with It is set to data storage and instruction;One is established the processor to communicate with memory, wherein, when performing the instruction in memory, The processor is configured as:Start the camera device of terminal;Obtain depth and width, the height in the first space, and default thickness Degree;In the i-th deep space, it is determined that the rigid body set Bi identified;Record the i-th figure layer Li, including ki rigid body;Judge whether to travel through The depth in the first space;If so, the first spatial model of structure;Plan the space path of first spatial model.
Therefore, according to the space path method and system for planning of some embodiments of the present application, by obtain space and Object in space, it is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, make object Movement can get around the space object of real world, improve entertainment effect, the experience in the augmented reality world.
Brief description of the drawings
To more fully understand and illustrating some embodiments of the present application, below with reference to the description of accompanying drawing reference implementation example, In the drawings, same digital number indicates corresponding part in the accompanying drawings.
Fig. 1 is the illustrative diagram of the Environment System provided according to some embodiments of the present application.
Fig. 2 is the exemplary cell schematic diagram that the electronic functionalities provided according to some embodiments of the present application configure.
Fig. 3 is the exemplary process diagram of the space path planing method provided according to some embodiments of the present application.
Fig. 4 is the illustrative diagram that the space path provided according to some embodiments of the present application is planned.
Embodiment
Below with reference to accompanying drawing description for ease of Integrated Understanding the application as defined in claim and its equivalent Various embodiments.These embodiments include various specific details in order to understand, but these be considered only as it is exemplary.Cause This, it will be appreciated by those skilled in the art that carrying out variations and modifications without departing from this to various embodiments described here The scope and spirit of application.In addition, briefly and to be explicitly described the application, the application will be omitted to known function and structure Description.
The term and phrase used in description below and claims is not limited to literal meaning, and be merely can Understand and as one man understand the application.Therefore, for those skilled in the art, it is possible to understand that, there is provided to the various implementations of the application The description of example is only the purpose to illustrate, rather than limitation appended claims and its application of Equivalent definitions.
Below in conjunction with the accompanying drawing in the application some embodiments, the technical scheme in the embodiment of the present application is carried out clear Chu, it is fully described by, it is clear that described embodiment is only some embodiments of the present application, rather than whole embodiments. Based on the embodiment in the application, those of ordinary skill in the art are obtained all under the premise of creative work is not made Other embodiment, belong to the scope of the application protection.
It should be noted that the term used in the embodiment of the present application is only merely for the mesh of description specific embodiment , and it is not intended to be limiting the application." one " of singulative used in the embodiment of the present application and appended claims, "one", " one kind ", " described " and "the" be also intended to including most forms, unless context clearly shows that other implications.Also It should be appreciated that term "and/or" used herein refers to and list items purposes comprising one or more mutually bindings are any Or it is possible to combine.Expression " first ", " second ", " described the first " and " described the second " be used for modify respective element without Consideration order or importance, are used only for distinguishing a kind of element and another element, without limiting respective element.
Terminal according to the application some embodiments can be electronic equipment, the electronic equipment can include smart mobile phone, PC (PC, such as tablet personal computer, desktop computer, notebook, net book, palm PC PDA), mobile phone, e-book Reader, portable media player (PMP), audio/video player (MP3/MP4), video camera, virtual reality device And one or more of combinations in wearable device etc. (VR).According to some embodiments of the present application, the wearable device Type of attachment (such as wrist-watch, ring, bracelet, glasses or wear-type device (HMD)), integrated type (such as electronics can be included Clothes), decorated type (such as pad skin, tatoo or built in electronic device) etc., or several combination.In some realities of the application Apply in example, the electronic equipment can be flexible, be not limited to the said equipment, or can be one kind in above-mentioned various equipment Or several combination.In this application, term " user " can be indicated using the people of electronic equipment or setting using electronic equipment Standby (such as artificial intelligence electronic equipment).
The embodiment of the present application provides a kind of space path planing method.For the ease of understanding the embodiment of the present application, below The embodiment of the present application is described in detail refer to the attached drawing.
Fig. 1 is the illustrative diagram of the Environment System 100 provided according to some embodiments of the present application.Such as Fig. 1 Shown, Environment System 100 can include electronic equipment 110, network 120 and server 130 etc..Electronic equipment 110 can be with Including bus 111, processor 112, memory 113, input/output module 114, display 115, communication module 116 and physics Key 117 etc..In some embodiments of the present application, electronic equipment 110 can omit one or more elements, or can enter one Step includes one or more of the other element.
Bus 111 can include circuit.The circuit can be with one or more element (examples in interconnection electronics 110 Such as, bus 111, processor 112, memory 113, input/output module 114, display 115, communication module 116 and secondary or physical bond 117).The circuit can also be realized between one or more elements in electronic equipment 110 communication (for example, obtain and/or Send information).
Processor 112 can include one or more coprocessors (Co-processor), application processor (AP, Application Processor) and communication processor (Communication Processor).As an example, processor 112 can perform with the control of one or more elements of electronic equipment 110 and/or data processing (for example, starting shooting dress Put, identify the operations such as object, record figure layer, structure spatial model, planning space path).
Memory 113 can be with data storage.The data can include other with one or more of electronic equipment 110 The related instruction of element or data.For example, the data can include the initial data of the before processing of processor 112, intermediate data And/or the data after processing.Memory 113 can include impermanent memory memory and/or permanent memory memory.Make For example, memory 113 can be with the collection of objects of memory space, the map data mining platform recorded etc..
According to some embodiments of the present application, memory 113 can store software and/or program.Described program can wrap Include kernel, middleware, API (API, Application Programming Interface) and/or using journey Sequence (or " application ").
At least a portion of the kernel, the middleware or the API can include operating system (OS,
Operating System).As an example, the kernel can be controlled or managed for performing other programs (in for example, Between part, API and application program) in realize operation or function system resource (for example, bus 111, processor 112nd, memory 113 etc.).In addition, the kernel can provide interface.The interface can by the middleware, it is described should One or more elements of electronic equipment 110 are accessed with DLL or the application program to control or management system resource.
The middleware can be as the intermediate layer of data transfer.The data transfer can allow API or Application program is with the kernel communication exchanging data.As an example, the middleware can be handled from the application program One or more task requests of acquisition.For example, the middleware can be to one or more application assigned electronic equipments The priority of 110 system resource (for example, bus 111, processor 112, memory 113 etc.), and processing it is one or Multiple tasks are asked.The API can be that the application program is used to control from the kernel or the middleware The interface of function is provided.The API can also include one or more interfaces or function (for example, instruction).It is described Function can be used for starting control, data channel control, security control, Control on Communication, document control, window control, text control System, image procossing, information processing etc..
Input/output module 114 can send what is inputted from user or external equipment to the other elements of electronic equipment 110 Instruction or data.Input/output module 114 can also be defeated by the instruction of the other elements acquisition from electronic equipment 110 or data Go out to user or external equipment.In certain embodiments, input/output module 114 can include input block, and user can lead to Cross the input block input information or instruction.
Display 115 can be with display content.The content can to user show all kinds (for example, text, image, Video, icon and/or symbol etc., or several combinations).Display 115 can include liquid crystal display (LCD, Liquid Crystal Display), light emitting diode (LED, Light-Emitting Diode) display, Organic Light Emitting Diode (OLED, Organic Light Emitting Diode) display, Micro Electro Mechanical System (MEMS, Micro Electro Mechanical Systems) display or electric paper display etc., or several combinations.Display 115 can include display Screen, touch-screen etc..The display screen can show the 3d space model of structure, display space path etc..In certain embodiments, Display 115 can show virtual key.The touch-screen can obtain the input of the virtual key.Display 115 can pass through The touch-screen obtains input.The input can include touch input, gesture input, action input, close input, electronic pen Or the input (for example, hovering input) of user's body part.
Communication module 116 can configure the communication between equipment.In certain embodiments, Environment System 100 can be with Further comprise electronic equipment 140.As an example, the communication between the equipment can include electronic equipment 110 and other set Communication between standby (for example, server 130 or electronic equipment 140).For example, communication module 116 can by radio communication or Wire communication is connected to network 120, realizes and communicates with other equipment (for example, server 130 or electronic equipment 140).
The radio communication can include microwave communication and/or satellite communication etc..The radio communication can include honeycomb Communication is (for example, global mobile communication (GSM, Global System for Mobile Communications), CDMA (CDMA, Code Division MultipleAccess), 3G (Third Generation) Moblie (3G, The 3rd Generation Telecommunication), forth generation mobile communication (4G), the 5th third-generation mobile communication (5G), Long Term Evolution (LTE, Long Term Evolution), Long Term Evolution upgrade version (LTE-A, LTE-Advanced), WCDMA (WCDMA, Wideband Code Division MultipleAccess), UMTS (UMTS, Universal Mobile Telecommunications System), WiMAX (WiBro, Wireless Broadband) etc., it is or several Combination.According to some embodiments of the present application, the radio communication can include WLAN (WiFi, Wireless Fidelity), bluetooth, low-power consumption bluetooth (BLE, Bluetooth Low Energy), ZigBee protocol (ZigBee), near-field communication (NFC, Near Field Communication), magnetic safe transmission, radio frequency and body area network (BAN, BodyAreaNetwork) Deng, or several combinations.According to some embodiments of the present application, the wire communication can include GLONASS (Glonass/GNSS, Global Navigation Satellite System), global positioning system (GPS, Global Position System), Beidou navigation satellite system or Galileo (European GPS) etc..The cable modem Letter can include USB (USB, Universal Serial Bus), high-definition media interface (HDMI, High- Definition Multimedia Interface), proposed standard 232 (RS-232, Recommend Standard 232), And/or plain old telephone service (POTS, Plain Old Telephone Service) etc., or several combinations.
Secondary or physical bond 117 can be used for user mutual.Secondary or physical bond 117 can include one or more entity keys.In some realities Apply in example, user can be with the function of self-defined secondary or physical bond 117.As an example, secondary or physical bond 117 can send instruction.The instruction It can include starting iris read write command, start-up space path planning etc..The iris read write command can include iris verification When read iris instruction, iris typing when write-in iris instruction etc..
In certain embodiments, electronic equipment 110 may further include sensor.The sensor can be included but not It is limited to light sensor, acoustic sensor, gas sensor, chemical sensor, voltage sensitive sensor, temp-sensitive sensor, fluid to pass Sensor, biology sensor, laser sensor, Hall sensor, intelligence sensor etc., or several combinations.
In certain embodiments, electronic equipment 110 may further include infrared equipment, image capture device etc..As Example, the infrared equipment can identify by infrared ray mode of delivery, and blink, watch the technical limit spacing eyes such as identification attentively Information.For example, the infrared equipment is acted come certification user profile by gathering the blink of user.As an example, described image Collecting device can include camera, iris device etc..The camera can realize the functions such as eyeball tracking.The iris dress Authentication (for example, certification user profile) can be carried out using iris recognition technology by putting.The iris device can include rainbow Film camera, the iris camera can obtain iris information, and the iris information can be stored in memory 113.
Network 120 can include communication network.The communication network can include computer network (for example, LAN (LAN, Local Area Network) or wide area network (WAN, Wide Area Network)), internet and/or telephone network Deng, or several combinations.Network 120 can be to the other equipment in Environment System 100 (for example, electronic equipment 110, clothes Business device 130, electronic equipment 140 etc.) send information.
Server 130 can be by the other equipment in the connection Environment System 100 of network 120 (for example, electronic equipment 110th, electronic equipment 140 etc.).In certain embodiments, when electronic equipment 110 is lost, server 130 can pass through network 120 send startup iris read write command to electronic equipment;When iris verification fails, server 130 can be entered by network 120 One step locks electronic equipment 110.In certain embodiments, server 130 can send the virtual object of structure to electronic equipment 110 Body information etc..
Electronic equipment 140 can be identical or different with electronic equipment 110 type.According to some embodiments of the present application, The part or all of operation performed in electronic equipment 110 can be in another equipment or multiple equipment (for example, electronic equipment 140 And/or server 130) in perform.In certain embodiments, when electronic equipment 110 be automatically or in response to request perform it is a kind of or When multiple functions and/or service, electronic equipment 110 can ask other equipment (for example, electronic equipment 140 and/or server 130) perform function and/or service are substituted.In certain embodiments, electronic equipment 110 is in addition to perform function or service, further Perform relative one or more functions.In certain embodiments, other equipment is (for example, electronic equipment 140 and/or clothes Business device 130) asked function or other related one or more functions can be performed, implementing result can be sent to electricity Sub- equipment 110.Electronic equipment 110 can repeat result or further handle implementing result, to provide asked function Or service.As an example, electronic equipment 110 can use cloud computing, distributed computing technology and/or client-server end to calculate meter Calculate etc., or several combinations.In certain embodiments, can be included according to the difference of cloud computing service property, the cloud computing Public cloud, private clound and mixed cloud etc..In certain embodiments, when electronic equipment 110 is lost, electronic equipment 140 can be to Electronic equipment 110 sends positioning instruction, to obtain the positional information of electronic equipment 110.In certain embodiments, electronic equipment 110 can be with the communal space information of electronic equipment 140.
It should be noted that the description for Environment System 100 above, only for convenience of description, can not be this Shen It please be limited within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system can Each element can be combined on the premise of without departing substantially from the principle, or forms subsystem and be connected with other elements, To implementing the various modifications and variations on the above method and systematic difference field progress form and details.For example, network environment System 100 may further include database.In another example electronic equipment 110 can not include secondary or physical bond 117 etc..It is all such The deformation of class, within the protection domain of the application.
Fig. 2 is the exemplary cell block diagram that the electronic functionalities provided according to some embodiments of the present application configure.Such as Shown in Fig. 2, processor 112 can include processing module 200, and the processing module 200 can include acquiring unit 210, control Unit 220, determining unit 230, generation unit 240, processing unit 250.
According to some embodiments of the present application, acquiring unit 210 can obtain information.In certain embodiments, the letter Breath can include but is not limited to text, picture, audio, video, action, gesture, sound, eyes (for example, iris information etc.), gas Breath, light etc., or several combinations.In certain embodiments, described information can include but is not limited to input information, system information And/or communication information etc..As an example, acquiring unit 210 can pass through the touch of input/output module 114, display 115 Screen, secondary or physical bond 117 and/or sensor obtain the input information of electronic equipment 110.The input information can include other equipment (for example, electronic equipment 140) and/or the input of user, for example, the input of key-press input, touch-control, gesture input, action input, remote Journey input, transmission input, eyes input, sound input, breath input, light input etc., or several combinations.The input information Obtaining widget can include but is not limited to infrared equipment, image capture device, sensor etc., or several combinations.As showing Example, acquiring unit 210 can obtain spatial information, collection of objects etc. by image capture device (for example, camera device etc.).Institute Rigid body (Rigid Body) can be included, define object etc. by stating object.
In certain embodiments, acquiring unit 210 can obtain the communication information by network 120.The communication information can With including application software information, communication signal (for example, voice signal, vision signal etc.), short message etc..In some embodiments In, acquiring unit 210 can obtain system information by network 120, memory 113 and/or sensor.The system information can With include but is not limited to the system mode of electronic equipment 110, presupposed information, memory 113 store information (for example, iris is recognized Demonstrate,prove information etc.) etc., or several combinations.As an example, acquiring unit 210 can obtain preset thickness information.The default thickness Degree information can include the thickness of deep space, the thickness of rigid body, thickness of figure layer etc..
In certain embodiments, described information can include instruction.The instruction includes user instruction and/or system command Deng, or several combinations.The instruction can include triggering command, certification instruction, fill in instruction etc., or several combinations.Institute Certification user profile can be included, start structure spatial model instruction etc. by stating certification instruction.As an example, if iris information is verified During failure, electronic equipment 110 can start user information authentication etc..In another example acquiring unit 210 can obtain startup structure sky Between model instruction.
According to some embodiments of the present application, control unit 220 can be with control electronics.In certain embodiments, control Unit 220 processed can start camera device (for example, it is first-class to start shooting), start structure spatial model etc..The camera Can be first-class including black and white camera (Mono), colour imagery shot, iris shooting.In certain embodiments, control unit 220 can With the camera device for the instruction startup terminal for building spatial model according to starting.In certain embodiments, control unit 220 can be with Record map data mining platform.
According to some embodiments of the present application, determining unit 230 can determine information.In certain embodiments, it is it is determined that single Member 230 can determine the object identified by camera device.The object can include rigid body (RigidBody), define object. The rigid body can include the object that can not be passed through.The object for defining object and can including passing through, for example, non-firm Body.In certain embodiments, determining unit 230 can determine whether flow travels through the depth in space.
According to some embodiments of the present application, generation unit 240 can generate information.In certain embodiments, generation is single Member 240 can build spatial model etc..As an example, generation unit 240 can be according to spatial information, collection of objects and figure layer Information etc. builds 3d space model.In certain embodiments, generation unit 240 can generate virtual 3D objects, the virtual 3D Object includes actual object image and/or imaginary subject image is presented, for example, the virtual 3D rendering of terminal user or game are fabricated Virtual 3D rendering etc..
According to some embodiments of the present application, processing unit 250 can be with processing information.In certain embodiments, processing is single Member 250 can utilize three dimensions to scale, and figure layer set is converted into coffin.In certain embodiments, processing unit 250 Figure layer set can be superimposed according to figure layer thickness.The figure layer set can include multiple figure layers corresponding to multiple deep spaces. In certain embodiments, processing unit 250 can delete the definition object of virtual 3D objects contact.In certain embodiments, locate Managing unit 250 can be according to spatial model planning space path.As an example, the space path can bypass the space Rigid body.
It should be noted that described above for the unit in processing module 200, only for convenience of description, can not be this Application is limited within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system, Unit may be combined on the premise of without departing substantially from the principle, or form submodule and connect with other units Connect, the various modifications and variations in form and details are carried out to the function of implementing above-mentioned module and unit.For example, processing module 200 may further include memory cell, and the memory cell can be with storage space information, record map data mining platform etc..It is all such The deformation of class, within the protection domain of the application.
Fig. 3 is the exemplary process diagram of the space path planing method provided according to some embodiments of the present application.Such as Fig. 3 Shown, flow 300 can be realized by processing module 200.In certain embodiments, the space path planing method can be certainly It is dynamic to start or started by instructing.The automatic start can be when opening application program, the planning of automatic start space path.Institute User instruction, system command, action command etc., or several combinations can be included by stating instruction.As an example, the system command The information that can be obtained by sensor generates.The user instruction can include voice, gesture, action, secondary or physical bond 117 and/ Or virtual key etc., or several combinations.As an example, when user starts camera device, it can directly initiate or pass through pressing Secondary or physical bond start-up space path planning.
301, start the camera device of terminal.Operation 301 can be realized by the control unit 220 of processing module 200. In certain embodiments, control unit 220 can start camera device.
302, the depth D, width, height in the first space, and preset thickness are obtained.Operation 302 can pass through processing The acquiring unit 210 of module 200 is realized.In certain embodiments, acquiring unit 210 can pass through the first of camera device identification Spatial information.The spatial information includes depth and width, height etc..As an example, the depth in first space can be D. In certain embodiments, acquiring unit 210 can obtain preset thickness.The preset thickness can include deep space thickness, Rigid body thickness, define object thickness, figure layer thickness etc..
303, in the i-th deep space, it is determined that the rigid body set Bi identified.Operation 303 can pass through processing module 200 Determining unit 230, control unit 220 are realized.In certain embodiments, determining unit 230 can determine camera device identification The rigid body set Bi of i-th deep space in the first space.In certain embodiments, the camera device can identify the first space Collection of objects, the collection of objects includes rigid body set, defines collection of objects etc..As an example, determining unit 230 can lead to Cross acquiring unit 210 and obtain the object that user clicks on, determine that the object that the user clicks on is rigid body.For example, the rigid body collection Multiple rigid bodies that user clicks in deep space can be included by closing, or in multiple deep spaces user click on it is multiple just Body.In certain embodiments, first space can divide N number of deep space according to different deep space thickness.Example Such as, when deep space thickness is T, N=D/T.I-th deep space can be any one depth of N number of deep space Spend space, i=1,2 ..., N.
In 304, the i-th figure layer Li of record, including kiIndividual rigid body:Bi1, Bi2... ..., Bik.Operation 304 can be by handling mould The control unit 220 of block 200 is realized.In certain embodiments, control unit 220 can be by the i-th deep space and its rigid body collection Close Bi and be recorded as the i-th figure layer Li.The i-th figure layer Li includes kiIndividual rigid body.
305, judge whether to travel through the depth in the first space.Operation 305 can pass through the determining unit of processing module 200 230 realize.In certain embodiments, determining unit 230 can determine flow 300 whether the depth in the first space.As an example, When the first space includes N number of deep space, determining unit 230 may determine whether from i=1 the first deep space travel through to I=N N deep spaces.
If not traveling through the depth in the first space, operation 303 is returned to, 303, in the i-th deep space, it is determined that the rigid body identified Set Bi.Operation 302 can be realized by determining unit 230, the control unit 220 of processing module 200.In some embodiments In, if traveled through i-th of deep space, i=i+1, that is, when returning to operation 303,303, in i+1 deep space, it is determined that The rigid body set B of identificationi+1.In certain embodiments, the terminal can be moved to D depth from depth zero, travel through the first space Depth.In certain embodiments, control unit 220 can adjust the camera device, from depth zero to D extreme saturations first The depth in space.
If having traveled through the depth in the first space, into operation 306,306, the first spatial model is built.Operation 306 can be with Realized by the generation unit 240 of processing module 200.In certain embodiments, generation unit 240 can be believed according to the first space Breath, N number of deep space and its rigid body set B, the map data mining platform of record etc. in the first space build the first spatial model.At some In embodiment, the rigid body thickness can be not more than the deep space thickness.In certain embodiments, processing unit 250 can To be scaled using three dimensions, figure layer set is converted into coffin;Generation unit 240 can according to the coffin, Build 3d space model.In certain embodiments, generation unit 240 can be superimposed figure layer set, structure the according to figure layer thickness One spatial model.The figure layer thickness can be identical with the deep space thickness.The figure layer set can include and N number of depth Spend the corresponding N number of figure layer in space.As an example, figure layer set L={ L1, L2..., LN}。
307, the space path of first spatial model is planned.Operation 307 can pass through the processing of processing module 200 Unit 250 is realized.In certain embodiments, processing unit 250 according to the first spatial model, can plan the space in the first space Path.The space path can get around the rigid body set B in first space.
It should be noted that the description for flow 300 above, only for convenience of description, can not be limited in the application Within the scope of illustrated embodiment.It is appreciated that for those skilled in the art, the principle based on the system, may not carry on the back On the premise of from the principle, each operation is combined, or forms sub-process and other operative combinations, in implementation State the various modifications and variations in flow and the function progress form and details of operation.For example, flow 300 may further include It is determined that define the operations such as collection of objects, the virtual 3D objects of generation.Such deformation, the application protection domain it It is interior.
Fig. 4 is the illustrative diagram that the space path provided according to some embodiments of the present application is planned.Such as Fig. 4 institutes Show, be the top view in space 400, the spatial information in space 400 includes depth D.As an example, the deep space in space 400 is thick Spend and correspond to figure layer L for T, the first deep space1, including rigid body B11, define object O11.Wherein, D=2T, space 400 include N number of Deep space, N=D/T=2.The energy and disaster in space 400 corresponds to figure layer L2, including rigid body B21With rigid body B22.It is described firm Body thickness is TB, TB≤T.In certain embodiments, the rigid body thickness can be different from actual rigid body thickness, space 400 it is firm Body set B rigid body thickness can be with identical.In certain embodiments, the space path of planning can include the first space path 401 and second space path 402.As an example, the rigid body can be the object that can not be passed through, the object that defines can be The object that can be passed through, first space path 401 can bypass rigid body B11With rigid body B21.The second space path 402 Rigid body B can be bypassed11, connection definition object O11, further around rigid body B excessively22
In summary, according to the space path method and system for planning of the embodiment of the present application, by obtaining space and sky In object, it is determined that the rigid body of identification and recording figure layer, to build spatial model, so as to planning space path, move object Kinetic energy enough gets around the space object of real world, improves entertainment effect, the experience in the augmented reality world.
It should be noted that the above embodiments are intended merely as example, the application is not limited to such example, but can To carry out various change.
It should be noted that in this manual, term " comprising ", "comprising" or its any other variant are intended to Nonexcludability includes, so that process, method, article or equipment including a series of elements not only will including those Element, but also the other element including being not expressly set out, or it is this process, method, article or equipment also to include Intrinsic key element.In the absence of more restrictions, the key element limited by sentence "including a ...", it is not excluded that Other identical element also be present in process, method, article or equipment including the key element.
Finally, it is to be noted that, a series of above-mentioned processing are not only included with order described here in temporal sequence The processing of execution, and the processing including performing parallel or respectively rather than in chronological order.
One of ordinary skill in the art will appreciate that realize all or part of flow in above-described embodiment method, being can be with To be completed by the related hardware of computer program instructions, described program can be stored in a computer-readable recording medium, The program is upon execution, it may include such as the flow of the embodiment of above-mentioned each method.Wherein, described storage medium can be magnetic disc, CD, read-only storage (Read-Only Memory, ROM) or random access memory (Random Access Memory, RAM) Deng.
Above disclosed is only some preferred embodiments of the application, it is impossible to the right model of the application is limited with this Enclose, one of ordinary skill in the art will appreciate that all or part of flow of above-described embodiment is realized, and will according to the application right Made equivalent variations are sought, still falls within and invents covered scope.

Claims (10)

  1. A kind of 1. space path planing method, it is characterised in that including:
    Start the camera device of terminal;
    Obtain depth and width, the height in the first space, and preset thickness;
    In the i-th deep space, it is determined that the rigid body set Bi identified;
    Record the i-th figure layer Li, including kiIndividual rigid body;
    Judge whether the depth in the first space of traversal;
    If so, the first spatial model of structure;
    Plan the space path of first spatial model.
  2. 2. space path planing method according to claim 1, it is characterised in that further comprise:
    If it is not, in i+1 deep space, it is determined that the rigid body set B identifiedi+1
    Record i+1 space figure layer Li+1, including ki+1Individual rigid body.
  3. 3. space path planing method according to claim 1, it is characterised in that the rigid body set for determining identification is entered One step includes:
    The collection of objects in first space is identified by the camera device, the object includes rigid body, defines object;
    Obtain the object that user clicks on;
    Determine that the object that the user clicks on is rigid body.
  4. 4. space path planing method according to claim 3, it is characterised in that further comprise:
    Obtain terminal preset rules or end users operation;
    According to the terminal preset rules or end users operation, it is determined that the definition collection of objects of identification.
  5. 5. space path planing method according to claim 1, it is characterised in that the preset thickness includes deep space Thickness T, rigid body thickness TB;The rigid body thickness is not more than the deep space thickness.
  6. 6. space path planing method according to claim 5, it is characterised in that further comprise:
    Obtain the depth D in first space, deep space thickness T;
    Determine that first space includes D/T deep space;
    The figure layer set in first space is obtained, the figure layer set includes D/T figure layer corresponding to D/T deep space.
  7. 7. space path planing method according to claim 6, it is characterised in that the first spatial model of the structure enters one Step includes:
    Scaled using three dimensions, the figure layer set is converted into coffin.
  8. 8. space path planing method according to claim 6, it is characterised in that the first spatial model of the structure enters one Step includes:
    According to deep space thickness T, the figure layer set is superimposed.
  9. 9. space path planing method according to claim 1, it is characterised in that further comprise:
    Virtual 3D objects are generated, the virtual 3D objects include actual object image and/or imaginary subject image is presented;
    The virtual 3D objects move according to the space path, and the space path includes getting around the path of the rigid body.
  10. A 10. system, it is characterised in that including:
    One memory, is configured as data storage and instruction;
    One is established the processor to communicate with memory, wherein, when performing the instruction in memory, the processor is configured For:
    Start the camera device of terminal;
    Obtain depth and width, the height in the first space, and preset thickness;
    In the i-th deep space, it is determined that the rigid body set Bi identified;
    Record the i-th figure layer Li, including kiIndividual rigid body;
    Judge whether the depth in the first space of traversal;
    If so, the first spatial model of structure;
    Plan the space path of first spatial model.
CN201710754154.3A 2017-08-29 2017-08-29 A kind of space path method and system for planning Pending CN107506040A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710754154.3A CN107506040A (en) 2017-08-29 2017-08-29 A kind of space path method and system for planning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710754154.3A CN107506040A (en) 2017-08-29 2017-08-29 A kind of space path method and system for planning

Publications (1)

Publication Number Publication Date
CN107506040A true CN107506040A (en) 2017-12-22

Family

ID=60693021

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710754154.3A Pending CN107506040A (en) 2017-08-29 2017-08-29 A kind of space path method and system for planning

Country Status (1)

Country Link
CN (1) CN107506040A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109935318A (en) * 2019-03-06 2019-06-25 珠海市万瑙特健康科技有限公司 Display methods, device, computer equipment and the storage medium of three-dimensional pulse wave

Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101726296A (en) * 2009-12-22 2010-06-09 哈尔滨工业大学 Vision measurement, path planning and GNC integrated simulation system for space robot
CN101770235A (en) * 2009-01-01 2010-07-07 索尼株式会社 Path planning device, path planning method, and computer program
CN101907459A (en) * 2010-07-12 2010-12-08 清华大学 Monocular video based real-time posture estimation and distance measurement method for three-dimensional rigid body object
CN103247075A (en) * 2013-05-13 2013-08-14 北京工业大学 Variational mechanism-based indoor scene three-dimensional reconstruction method
CN103649996A (en) * 2011-05-10 2014-03-19 皇家飞利浦有限公司 User-steered on-the-fly path planning
CN103761737A (en) * 2014-01-22 2014-04-30 北京工业大学 Robot motion estimation method based on dense optical flow
CN105122304A (en) * 2012-11-14 2015-12-02 微软技术许可有限责任公司 Real-time design of living spaces with augmented reality
CN105143907A (en) * 2013-04-22 2015-12-09 阿尔卡特朗讯 Localization systems and methods
CN105320140A (en) * 2015-12-01 2016-02-10 浙江宇视科技有限公司 Robot cleaner and cleaning path planning method thereof
CN105493155A (en) * 2013-08-30 2016-04-13 高通股份有限公司 Method and apparatus for representing physical scene
CN106052674A (en) * 2016-05-20 2016-10-26 青岛克路德机器人有限公司 Indoor robot SLAM method and system
CN106127739A (en) * 2016-06-16 2016-11-16 华东交通大学 A kind of RGB D SLAM method of combination monocular vision
CN106197263A (en) * 2016-06-28 2016-12-07 大连理工大学 Many bodies based on machine vision throw in pose measuring method
CN106643692A (en) * 2016-09-28 2017-05-10 深圳乐行天下科技有限公司 Robot navigation and positioning method, system and robot
CN107003728A (en) * 2014-11-25 2017-08-01 微软技术许可有限责任公司 Virtual measurement instrument for wearable visualization device

Patent Citations (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101770235A (en) * 2009-01-01 2010-07-07 索尼株式会社 Path planning device, path planning method, and computer program
CN101726296A (en) * 2009-12-22 2010-06-09 哈尔滨工业大学 Vision measurement, path planning and GNC integrated simulation system for space robot
CN101907459A (en) * 2010-07-12 2010-12-08 清华大学 Monocular video based real-time posture estimation and distance measurement method for three-dimensional rigid body object
CN103649996A (en) * 2011-05-10 2014-03-19 皇家飞利浦有限公司 User-steered on-the-fly path planning
CN105122304A (en) * 2012-11-14 2015-12-02 微软技术许可有限责任公司 Real-time design of living spaces with augmented reality
CN105143907A (en) * 2013-04-22 2015-12-09 阿尔卡特朗讯 Localization systems and methods
CN103247075A (en) * 2013-05-13 2013-08-14 北京工业大学 Variational mechanism-based indoor scene three-dimensional reconstruction method
CN105493155A (en) * 2013-08-30 2016-04-13 高通股份有限公司 Method and apparatus for representing physical scene
CN103761737A (en) * 2014-01-22 2014-04-30 北京工业大学 Robot motion estimation method based on dense optical flow
CN107003728A (en) * 2014-11-25 2017-08-01 微软技术许可有限责任公司 Virtual measurement instrument for wearable visualization device
CN105320140A (en) * 2015-12-01 2016-02-10 浙江宇视科技有限公司 Robot cleaner and cleaning path planning method thereof
CN106052674A (en) * 2016-05-20 2016-10-26 青岛克路德机器人有限公司 Indoor robot SLAM method and system
CN106127739A (en) * 2016-06-16 2016-11-16 华东交通大学 A kind of RGB D SLAM method of combination monocular vision
CN106197263A (en) * 2016-06-28 2016-12-07 大连理工大学 Many bodies based on machine vision throw in pose measuring method
CN106643692A (en) * 2016-09-28 2017-05-10 深圳乐行天下科技有限公司 Robot navigation and positioning method, system and robot

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109935318A (en) * 2019-03-06 2019-06-25 珠海市万瑙特健康科技有限公司 Display methods, device, computer equipment and the storage medium of three-dimensional pulse wave

Similar Documents

Publication Publication Date Title
CN107835365A (en) Auxiliary shooting method and system
CN107657953A (en) Sound control method and system
CN107609914A (en) A kind of intelligent menu design method and system
CN108009140A (en) A kind of end message edit methods and system
US11579744B2 (en) Systems and methods for seat selection in virtual reality
CN107358179A (en) A kind of living management system, medium and method based on iris verification
CN107423100A (en) A kind of methods, devices and systems of terminal OTA firmware updatings
CN107423585A (en) The concealed application method and system of a kind of application
CN107368793A (en) A kind of colored method for collecting iris and system
CN105190469A (en) Causing specific location of an object provided to a device
CN107786979A (en) A kind of multiple terminals shared communication method and system
CN107862518A (en) A kind of method of payment and system based on terminal location
CN107404484A (en) A kind of method, apparatus and system of mechanical release authority
CN107220531A (en) A kind of information processing method of convenient login
CN107018153A (en) A kind of safe login method
CN107506040A (en) A kind of space path method and system for planning
CN206908680U (en) Intelligent terminal with secure log function
CN108154556A (en) A kind of virtual trailing of terminal and system
CN107835117A (en) A kind of instant communicating method and system
CN107402690A (en) A kind of global collecting method and system
CN107277625A (en) A kind of management method and system of video playback history
CN108897479A (en) A kind of terminal touch control method and system
CN107592398A (en) A kind of intelligent information storage method and system
CN109189536A (en) A kind of terminal applies display methods and system
CN107623736A (en) A kind of equipment connection method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20171222