WO2021142787A1 - 行进路线及空间模型生成方法、装置、系统 - Google Patents
行进路线及空间模型生成方法、装置、系统 Download PDFInfo
- Publication number
- WO2021142787A1 WO2021142787A1 PCT/CN2020/072814 CN2020072814W WO2021142787A1 WO 2021142787 A1 WO2021142787 A1 WO 2021142787A1 CN 2020072814 W CN2020072814 W CN 2020072814W WO 2021142787 A1 WO2021142787 A1 WO 2021142787A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- travel route
- images
- module
- image
- model
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 78
- 238000004590 computer program Methods 0.000 claims description 21
- 230000008569 process Effects 0.000 claims description 17
- 238000012545 processing Methods 0.000 claims description 17
- 230000006870 function Effects 0.000 description 33
- 238000010586 diagram Methods 0.000 description 14
- 230000001133 acceleration Effects 0.000 description 6
- 238000004891 communication Methods 0.000 description 6
- 230000003287 optical effect Effects 0.000 description 6
- 238000013135 deep learning Methods 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 4
- 238000006073 displacement reaction Methods 0.000 description 3
- 238000010276 construction Methods 0.000 description 2
- 239000011159 matrix material Substances 0.000 description 2
- 230000000644 propagated effect Effects 0.000 description 2
- 239000007787 solid Substances 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 238000005034 decoration Methods 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000000835 fiber Substances 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 239000004973 liquid crystal related substance Substances 0.000 description 1
- 239000013307 optical fiber Substances 0.000 description 1
- 239000004065 semiconductor Substances 0.000 description 1
- 230000009897 systematic effect Effects 0.000 description 1
- 238000012546 transfer Methods 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4038—Image mosaicing, e.g. composing plane images from plane sub-images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T11/00—2D [Two Dimensional] image generation
- G06T11/20—Drawing from basic elements, e.g. lines or circles
- G06T11/203—Drawing of straight lines or curves
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/32—Indexing scheme for image data processing or generation, in general involving image mosaicing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2210/00—Indexing scheme for image generation or computer graphics
- G06T2210/04—Architectural design, interior design
Definitions
- the present disclosure relates to the field of image processing, and in particular to a method, device, system, equipment, and storage medium for generating a travel route and a space model.
- the present disclosure is completed to solve the above-mentioned problems, and its purpose is to provide a travel route and a spatial model generation method, device, system, equipment, and storage medium that can automatically splice and generate spatial models and overcome the loss of travel routes due to unexpected interruptions.
- the embodiments of the present disclosure provide a method for generating a travel route, which adopts the following technical solutions, including:
- the first image shooting step moving from the first shooting point and shooting multiple images
- a first travel route generating step matching the feature points of the multiple images taken in the first image capturing step to generate a first travel route
- the second image shooting step moving from the second shooting point and shooting multiple images
- a second travel route generating step matching the feature points of the multiple images taken in the second image capturing step to generate a second travel route
- the feature points of the image on the second travel route are matched with the feature points of the image on the first travel route, so that the second travel route matches the first travel route. Stitching of travel routes.
- embodiments of the present disclosure also provide a travel route generation device, which adopts the following technical solutions, including:
- the receiving module receives multiple sets of images that are moved from different shooting points and shot separately;
- the travel route generation module respectively matches the feature points of the multiple sets of images to generate multiple travel routes respectively;
- the splicing module matches the feature points of different groups of images, so that the multiple travel routes are spliced.
- the embodiments of the present disclosure also provide a travel route generation system, which adopts the following technical solutions, including:
- Image shooting module which moves from different shooting points and shoots multiple sets of images respectively
- the travel route generation module respectively matches the feature points of the multiple sets of images to generate multiple travel routes respectively;
- the splicing module matches the feature points of different groups of images, so that the multiple travel routes are spliced.
- the embodiments of the present disclosure also provide a space model generation method, which adopts the following technical solutions, including:
- the travel route generation step uses the travel route generation method described in the preceding paragraph to generate the travel route
- a model image shooting step in the process of moving according to the travel route, shooting a model image for generating the space model in the space in which it is located;
- the model generation step is to generate a model of each space based on the model images taken in each space;
- the models of each of the spaces are spliced in the same coordinate system to form a splicing result of the respective models of each of the spaces.
- the embodiments of the present disclosure also provide a space model generation device, which adopts the following technical solutions, including:
- the travel route generating device as described in the preceding paragraph to generate a travel route
- a receiving device which receives multiple groups of model images taken separately from multiple spaces
- a model generating module based on the multiple sets of model images received by the receiving device, respectively generating models of each of the spaces;
- the model splicing module based on the position and orientation information of each space in the travel route, performs splicing processing on the models of each space in the same coordinate system to form a splicing result of the respective models of each space. Into the overall model.
- the embodiments of the present disclosure also provide a space model generation system, which adopts the following technical solutions, including:
- a model image capturing device which captures a model image for generating the space model in the space in which it is located;
- a model generation module based on the model images respectively captured by the model image capturing device for a plurality of the spaces, respectively generating models of each of the spaces;
- the travel route generation system as described in the preceding paragraph to generate travel routes
- the model splicing module based on the position and orientation information of each space in the travel route, performs splicing processing on the models of each space in the same coordinate system to form a splicing result of the respective models of each space. Into the overall model.
- the embodiments of the present disclosure also provide a computer device, which adopts the following technical solutions, including:
- a memory and a processor wherein a computer program is stored in the memory, and the processor implements the aforementioned method when the computer program is executed.
- the embodiments of the present disclosure also provide a computer-readable storage medium, which adopts the following technical solutions, including:
- the computer-readable storage medium stores a computer program, and when the computer program is executed by a processor, the method as described above is implemented.
- the present disclosure can overcome the loss of the travel route due to accidental interruption and can automatically splice and generate a spatial model, thereby improving user experience.
- Fig. 1 is an exemplary system architecture diagram to which the present disclosure can be applied;
- Fig. 2 is a flowchart of an embodiment of a method for generating a travel route according to the present disclosure
- FIG. 3 is a schematic diagram of an embodiment of a method for generating a travel route according to the present disclosure
- Fig. 4 is a schematic diagram of an embodiment of a travel route generating device according to the present disclosure.
- Fig. 5 is a schematic diagram of an embodiment of a travel route generation system according to the present disclosure.
- Fig. 6 is a flowchart of an embodiment of a spatial model generation method according to the present disclosure.
- Fig. 7 is a schematic diagram of an embodiment of a space model generating device according to the present disclosure.
- Fig. 8 is a schematic diagram of an embodiment of a space model generation system according to the present disclosure.
- Fig. 9 is a schematic structural diagram of an embodiment of a computer device according to the present disclosure.
- the system structure 100 may include terminal devices 101, 102, 103, 104, a network 105, and a server 106.
- the network 105 is used to provide a medium for communication links between the terminal devices 101, 102, 103, 104 and the server 106.
- the electronic device (such as the terminal device 101, 102, 103, or 104 shown in FIG. 1) on which the travel route generation method runs can transmit various information through the network 105.
- the network 105 may include various connection types, such as wired, wireless communication links, or fiber optic cables, and so on. It should be pointed out that the above wireless connection methods can include but are not limited to 3G/4G/5G connection, Wi-Fi connection, Bluetooth connection, WiMAX connection, Zigbee connection, UWB connection, local area network (“LAN”), wide area network (“WAN” ), the Internet (for example, the Internet), end-to-end networks (for example, ad hoc end-to-end networks), and other network connection methods currently known or developed in the future.
- the network 105 can communicate with any currently known or future developed network protocol such as HTTP (Hyper Text Transfer Protocol), and can communicate with any form or medium of digital data (for example, a communication network) interconnection.
- HTTP Hyper Text Transfer Protocol
- the user can use the terminal devices 101, 102, 103, 104 to interact with the server 106 through the network 105 to receive or send messages, and so on.
- Various client applications can be installed on the terminal device 101, 102, 103, or 104, such as live video and playback applications, web browser applications, shopping applications, search applications, instant messaging tools, email clients, and social platforms Software etc.
- the terminal device 101, 102, 103, or 104 can be various electronic devices with touch screens and/or support for web browsing, including but not limited to smart phones, tablet computers, e-book readers, MP3 players (moving image expert compression Standard audio level 3), MP4 (moving image expert compressed standard audio level 4) player, head-mounted display device, notebook computer, digital broadcast receiver, PDA (personal digital assistant), PMP (portable multimedia player), car Terminals (for example, car navigation terminals) and other mobile terminals, as well as digital TVs, desktop computers, and the like.
- MP3 players moving image expert compression Standard audio level 3
- MP4 moving image expert compressed standard audio level 4
- head-mounted display device notebook computer
- digital broadcast receiver personal digital assistant
- PMP portable multimedia player
- car Terminals for example, car navigation terminals
- other mobile terminals as well as digital TVs, desktop computers, and the like.
- the server 106 may be a server that provides various services, for example, a background server that provides support for pages displayed on the terminal device 101, 102, 103, or 104 or transmitted data.
- terminal devices, networks, and servers in FIG. 1 are merely illustrative. There can be any number of terminal devices, networks, and servers according to implementation needs.
- the terminal device can independently or cooperate with other electronic terminal devices to run applications in various operating systems, such as the Android system, to implement the embodiments of the present disclosure, and can also run applications in other operating systems, such as iOS, Windows, and Hongmeng.
- the application of the system and the like implements the embodiment method of the present disclosure.
- the method for generating a travel route includes the following steps:
- the first image shooting step S21 moving from the first shooting point and shooting multiple images
- the first photographing point is, for example, the starting point of the travel route, and the travel route is generated from the first photographing point;
- the image photographed from the first photographing point is, for example, an image for positioning, which may be photographs, previews, video frames, etc. , Can be stored or not stored but only used for feature point identification and matching.
- the first travel route generation step S22 is to match the feature points of the multiple images captured in the first image capturing step S21 to generate the first travel route;
- the relative displacement of each shooting point is obtained by matching the feature points of the positioning image of the close shooting point, thereby providing the relative position and direction of each shooting point, and connecting the shooting points to generate a travel route.
- the travel route may be in a visible form that shows the connection of each shooting point, or it may be invisible without displaying the connection of each shooting point.
- the travel route color, thickness, line shape, virtual and solid forms are not limited, and may be in any display form.
- the second image shooting step S23 moving from the second shooting point and shooting multiple images
- the second image capturing step S23 is started.
- the reason for the feature point mismatch is, for example, that the movement is too fast and there are not enough feature points in the adjacent two frames of images to match; or during the movement, there is interference in the environment or the environment changes, such as entering a rough room or Environments with poor lighting conditions (too dark or too strong); or the shooting was interrupted by external factors during the shooting process, such as receiving a phone call causing the shooting to be interrupted.
- the second shooting point can be on the first travel route, or certainly not on the first travel route, for example, it can be located near the point on the first travel route.
- it cannot be matched to find the route for a long time (for example, When the environment has undergone major changes and the feature points cannot be matched), it is also possible to restart an independent travel route without being located near the shooting point on the first travel route.
- the second travel route generation step S24 is to match the feature points of the multiple images captured in the second image capturing step S23 to generate a second travel route;
- the method of generating the travel route and the display form of the route are the same as the first travel route generation step S22, and will not be repeated.
- the feature points of the image on the second travel route are matched with the feature points of the image on the first travel route, so that the second travel route is spliced with the first travel route.
- the feature point of the image captured in the second image capturing step S23 is the same as
- the feature points of the images captured in the existing first image capturing step S21 are compared, and an attempt is made to find enough feature points for matching, so as to calculate the mutual positions of the two travel routes, and perform route stitching.
- the shooting point of the image on the second travel route may be one or multiple, as long as there are enough feature points for matching, which is not limited.
- the position information of the image photographed in step S21 is photographed by the first image and the image photographed in step S23 is photographed by the second image Calculate the mutual position of the two travel routes based on the location information, and perform route splicing.
- the travel route generation method may further include: a saving step of saving at least part of the information of the multiple images captured in the first image capturing step S21;
- the image of the information saved in the saving step can be one or more images, one or more images at the initial shooting position, or one or more images at the position before the feature point mismatch. It can be multiple images separated by a certain distance, or multiple images taken continuously, which is not limited here.
- part of the information of the image saved in the saving step can be saved locally or uploaded to the server for saving.
- part of the information of the image saved in the saving step includes at least the feature point information in at least one image.
- at least one image or picture information of the image can also be saved for extracting the feature point information, and of course, it can also be saved.
- the attribute information of at least one image such as the shooting time, shooting location, shooting direction, and shooting angle of the image, is not limited.
- the feature points of the image taken at the second shooting point are matched with the feature points of the image saved in the saving step, so that the second travel route is spliced with the first travel route.
- the travel route generation method may further include: a prompting step of prompting a point on the first travel route before the shooting point of the image in which at least part of the information was last saved in the saving step as the first travel route. 2. Shooting point.
- the point before the shooting point of the image in which at least a part of the information is last saved may include the last point.
- the point on the first travel route may be a certain point on the travel route, or a point within a certain range of a certain point on the travel route.
- the prompt method can be, for example:
- the screen displays a reminder: "The route is lost! Please go back to point XX and repeat the route!, for example, you can also set up multiple display modes on the screen to help users understand the route, such as a red mark moving on the route or the route flashes Active way, for example, you can also show the photos of the shooting point you need to return to at the same time;
- the above prompt content is broadcast by voice: "The route is lost! Please go back to point XX and repeat the route! This disclosure is not limited.
- the travel route generation method may further include:
- the first positioning step is to locate and record at least the position information of the shooting point of an image saved in the saving step, for example, the position information of the shooting point of the last image;
- the second positioning step is to locate and record the position information of at least one shooting point on the second travel route
- the feature points of the image on the second travel route cannot be matched with the feature points of the image saved in the saving step
- use the position information located and recorded in the first positioning step and locate and record in the second positioning step The location information of the first travel route and the second travel route are spliced.
- the feature points of the image on the second travel route may be that a set of feature points corresponding to the image of any shooting point on the second travel route cannot match one or more sets of feature points of the image saved in the saving step. It may also be the case that all the feature points corresponding to the image on the second travel route cannot be matched with all the feature points of the image saved in the saving step, and the number of unmatched feature points is not limited.
- the position information of one or more shooting points located and recorded in the first positioning step and the position information of one or more shooting points located and recorded in the second positioning step are used to determine the relative position during splicing.
- FIG. 3 it is a schematic diagram of an embodiment of the method for generating a travel route according to the present disclosure.
- the porch is the first shooting point 1
- the circular arrow is the location where the feature points are mismatched
- the living room shooting point 2 is a point saved on the first travel route, which can be saved on the first travel route.
- the last point can also be a point before the last point
- the triangle arrow is the direction of the travel route.
- Step 1 Move and shoot multiple images from the first shooting point
- a mobile device with a camera function (which can be a mobile phone, a tablet computer, etc.) is fixed on a shooting stand (which can be, for example, a tripod, etc.), as shown in Figure 3, the entrance is determined as the starting position Take the first shooting point 1, and start to move the stand, and take multiple positioning images during the process of moving the stand.
- Step 2 Match the feature points of the multiple positioning images taken in Step 1, as shown in Figure 3, to generate the first travel route 1-2;
- the travel route shows that the connection of each shooting point is in a visible form, and of course, it is also possible to not display the connection of each shooting point to be an invisible form.
- Step 3 Save the feature points of one or more images taken in step 1, in this embodiment, for example, the feature points of the image of the living room at the living room shooting point 2;
- Step 4 As shown in Figure 3, the circular arrow is the current feature point mismatch, that is, the point where the position is lost.
- the reason for the feature point mismatch in this embodiment is, for example, that the two adjacent frames of images are not enough due to too fast movement
- Many feature points are matched; or during the movement, there is interference in the environment or the environment changes, such as entering a rough room or an environment with poor light conditions (too dark or too strong); or being affected by external factors during the shooting process Interruption of shooting, such as shooting interruption caused by answering the phone.
- the shooting point of one or more images saved before the living room shooting point 2 or living room shooting point 2 on the first travel route is prompted as the second shooting point, In this embodiment, for example, shooting point 2 in the living room is prompted as the second shooting point.
- the way of reminding includes displaying a reminder on the screen: "The route is lost! Please move the stand to the living room and repeat the route after the living room", and at the same time, multiple display modes are set on the screen to help users understand
- the route of travel for example, the way of moving on the route with a triangular arrow and showing the photo of shooting point 2 that needs to be returned at the same time, and the above prompt content is broadcast by voice: "The route is lost! Please move the bracket to the living room and repeat the route after the living room ".
- Step 5 Move and take multiple images from the second shooting point 2;
- the second shooting point 2 is on the first travel route 1-2, of course, it may not be on the first travel route 1-2, for example, it may be near a point on the first travel route.
- Step 6 matching the feature points of the multiple images shot in step 5 to generate a second travel route after the second shooting point 2;
- Step 7 the feature points of the image taken in step 5 are matched with the feature points of the image taken in step 1, so that the second travel route is spliced with the first travel route 1-2.
- the feature points of the image taken at the second shooting point 2 are matched with the feature points of the image of the living room shooting point 2 saved in the saving step, so that the second travel route is consistent with the first
- the one-way route is 1-2 spliced.
- the computer program can be stored in a computer readable storage medium, and the program can be stored in a computer readable storage medium. When executed, it may include the procedures of the above-mentioned method embodiments.
- the aforementioned storage medium may be a non-volatile storage medium such as a magnetic disk, an optical disc, a read-only memory (ROM), or a random storage memory (RAM).
- an embodiment of the present disclosure provides a travel route generation device, which can be specifically applied to various electronic terminal equipment, including: a receiving module, a travel route generation module, and a splicing Module.
- the receiving module receives multiple sets of images that are moved from different shooting points and shot separately;
- the different shooting points include at least a first shooting point and a second shooting point
- the images taken from each shooting point are a group of images.
- the image captured at each shooting point is, for example, an image for positioning, which may be a captured photo, preview image, video frame, etc., which may be stored or not stored but only used for feature point identification and matching.
- the travel route generation module matches the feature points of multiple sets of images to generate multiple travel routes respectively;
- each group of images generates a travel route through feature point matching.
- the relative displacement of each shooting point is obtained by matching the feature points of the image for positioning of similar shooting points, so as to provide the relative position and direction of each shooting point.
- the travel route may be in a visible form that shows the connection of each shooting point, or it may be invisible without displaying the connection of each shooting point.
- the displayed route color, thickness, line shape, virtual and solid forms are not limited, and may be in any display form.
- the stitching module matches the feature points of different sets of images so that multiple travel routes can be stitched; here, when the shooting points of different travel routes overlap or are in nearby locations, the pictures taken on different travel routes.
- the feature points of the image are compared, and an attempt is made to find enough feature points to match, so as to calculate the mutual position of different travel routes and perform route stitching.
- the position information of the images taken on the different travel routes is used to calculate the mutual positions of the different travel routes, and route stitching is performed.
- the multiple sets of images include at least a first set of images and a second set of images.
- the traveling route generating device may include: a receiving module 401, a traveling route generating module 402, The splicing module 403, the saving module 404, the prompting module 405, and the positioning module 406.
- the functions of the receiving module 401, the travel route generating module 402, and the splicing module 403 may be the same as the functions of the corresponding modules in the foregoing embodiment, and are not repeated here.
- the saving module 404 saves at least part of the information of the first group of images received by the receiving module 401;
- the image saved by the saving module 404 may be one or more images, one or more images at the initial shooting position, or one or more images at the position before the feature point mismatch. It can be multiple images separated by a certain distance, or multiple images taken continuously, which is not limited here.
- part of the information of the image saved in the saving module 404 may be saved locally or uploaded to the server for saving.
- part of the information of the image saved in the saving module 404 includes at least the feature point information in at least one image.
- at least one image or picture information of the image can also be saved for extracting the feature point information.
- the attribute information of at least one image is stored, such as the shooting time, shooting location, shooting direction, shooting angle, etc. of the image, which are not limited.
- the stitching module 403 matches the feature points of the second group of images with the image saved in the saving module 404, so as to stitch multiple travel routes.
- the prompting module 405 generates information for indicating returning to a specific shooting position in the event that a feature point mismatch occurs during the process of matching the feature points of the image by the travel route generation module 402, for example, prompting the travel corresponding to the first group of images In the route, the point before the shooting point of the image in which at least part of the information is last saved in the saving module 404 is used as the starting shooting point of the second group of images.
- the point before the shooting point of the image in which at least a part of the information is last saved may include the last point.
- the point on the first travel route may be a certain point on the travel route, or a point within a certain range of a certain point on the travel route.
- the positioning module 406 locates and records at least one image saved in the storage module 404, such as the position information of the shooting point of the last image; and at least locates and records the position information of one shooting point of the second group of images;
- the location information of the shooting points of all the images stored in the storage module 404 can also be located and recorded, which is not limited.
- it can also be used to locate and record the position information of the shooting points of all the images of the second group of images, which is not limited.
- the location information located and recorded in the positioning module 406 is used to stitch the travel route.
- the position information of one or more shooting points of the first group of images and the position information of one or more shooting points of the second group of images located and recorded in the positioning module 406 are used to determine the relative position during splicing.
- each block in the block diagram of the accompanying drawings may represent a module, a part of the module contains one or more executable instructions for implementing the specified logical function, but these modules are not necessarily in order Execute in order.
- the modules and functional units in the device embodiments of the present disclosure may be integrated into one processing module, or each unit may exist alone physically, or two or more modules or functional units may be integrated into one module.
- the above-mentioned integrated modules can be implemented in the form of hardware or software function modules. If the integrated module is implemented in the form of a software functional module and sold or used as an independent product, it can also be stored in a computer readable storage medium.
- the aforementioned storage medium may be a read-only memory, a magnetic disk or an optical disk, etc.
- an embodiment of the present disclosure provides a travel route generation system, including:
- Image shooting module which moves from different shooting points and shoots multiple sets of images respectively
- the image capturing module can be implemented by terminal devices such as a camera and/or a mobile phone with a camera function.
- terminal devices such as a camera and/or a mobile phone with a camera function.
- a camera that implements the image capturing module and/or a mobile phone with a camera function can be fixed on On the same shooting stand; in the process of moving the stand, obtain positioning images taken by multiple cameras or mobile phones with camera functions, so as to obtain and record the position and location of the camera or mobile phone with camera function when taking images in the space.
- the positioning images taken by the camera or the mobile phone with the camera function can be used to obtain each shooting point by matching the feature points of the positioning images at similar shooting points.
- the relative displacement of each camera provides the relative position and direction of each shooting point.
- the location, direction and route map of the shooting point can also be calculated from the camera image.
- the travel route generation module respectively matches the feature points of multiple sets of images to generate multiple travel routes; here, the function of the travel route generation module may be the same as the travel route generation module in the travel route generation device, for example, here It will not be repeated, but it is not limited. Of course, it can have other functions according to the cooperation with other modules in the travel route generation system in this embodiment.
- the splicing module matches the feature points of different groups of images to splice multiple travel routes.
- the function of the splicing module may be the same as the splicing module in the travel route generating device described above, which will not be repeated here, but not It is not limited, and of course, it can have other functions according to the cooperation with other modules in the travel route generation system in this embodiment.
- the travel route generation system of this embodiment includes: an image capturing module 501, a travel route generation module 502, a splicing module 503, a saving module 504, a prompting module 505, Positioning module 506.
- the various modules of the travel route generation system in this embodiment can be installed in whole or in part in the terminal device that implements this embodiment.
- it can also be installed in whole or in part on the server.
- the route generation module 502 and the splicing The module 503, the saving module 504, and the positioning module 506 are provided in the server implementing this embodiment
- the image capturing module 501 and the prompting module 505 are provided in the terminal device implementing this embodiment.
- the functions of the image capturing module 501, the travel route generation module 502, and the splicing module 503 may be the same as those of the corresponding modules in the above-mentioned embodiment. Of course, they may also be used in cooperation with other modules in the travel route generation system in this embodiment. Other functions.
- the function of each module in this embodiment may be the same as the corresponding module in the above-mentioned travel route generation device, which will not be repeated here, but is not limited. Of course, it can be based on other modules in the travel route generation system in this embodiment.
- the coordination has other functions.
- the multiple groups of images include at least a first group of images and a second group of images
- the image capturing module 501 moves from the first shooting point and shoots the first group of images from the second shooting point. Start moving and take the second set of images;
- the travel route generation module 502 matches the feature points of the first group of images to generate a first travel route; and matches the feature points of the second group of images to generate a second travel route;
- the stitching module 503 matches the feature points of the second group of images with the feature points of the first group of images, so as to stitch the second travel route with the first travel route.
- the saving module 504 also saves at least a part of the information of the first group of images; and the stitching module 503 compares the feature points of the image captured at the second shooting point with the feature points of the image saved by the saving module 504 Matching is performed so that the second travel route is spliced with the first travel route.
- the prompting module 505 prompts the point before the shooting point of the image in which at least part of the information was last saved by the saving module 504 on the first travel route as the second shooting point.
- the positioning module 506 locates and records at least one image saved by the saving module 504, such as the location information of the shooting point of the last image; and at least locates and records the location information of a shooting point on the second travel route;
- the location information located and recorded by the positioning module 506 is used to perform the first travel route and the second travel route. Splicing.
- the travel route generation system may further include a receiving module, a sending module, etc., for example, and specific functions include:
- a receiving module which receives multiple positioning images that are moved and photographed by the image capturing module 501;
- the travel route generation module 502 matches the feature points of the positioning image to generate a travel route
- the saving module 504 saves at least part of the information of the multiple positioning images
- the prompt module 505 generates information for instructing the image capturing module 501 to return to a specific shooting position in the event that a feature point mismatch occurs during the process of matching the feature points of the image by the travel route generating module 502.
- the specific shooting position is The position before the shooting position of the image for positioning of at least part of the information last saved by the pre-mismatch saving module 504 on the generated travel route;
- the sending module sends the information generated by the prompt module 505 for instructing the image capturing module 501 to return to a specific shooting position to the image capturing module 501.
- the travel route generation module 502 compares the feature points of the positioning image with the positioning image saved by the storage module 504. The feature points of the image are matched for route stitching.
- a stitching module 503 is further included, and the image capturing module 501 is used to return to the specific shooting position and re-shoot the position information and the specific shooting position information to compare the new travel route with the travel route generated before the mismatch. Make splicing.
- the present disclosure provides a space model generation method, including:
- the travel route generation step S61 for example, a method such as one or more embodiments of the travel route generation method in the present disclosure may be used to generate the travel route;
- the travel route generation step S61 can also include other methods, which are not limited.
- the travel route is generated by manual splicing, or can be generated according to pre-input route position and direction information, or can be generated by, for example, the following spatial model generation system
- the acceleration information provided by the acceleration sensor, speed sensor, and movement speed information provided by the relevant module in the middle generates a travel route.
- Model image shooting step S62 in the process of moving according to the travel route, shooting a model image for generating a space model in the space in which it is located;
- different shooting points in the process of moving according to the route of travel use the binocular lens of the relevant module in the following space model generation system to shoot model images or use a panoramic camera to shoot and generate model images.
- the panorama image is used as the model image.
- Model generation step S63 generating models of each space based on the model images taken in each space respectively;
- Figure for example, by performing image comparison on the model images taken by each of the binocular lenses, the corresponding pixels are determined, and the depth information of each corresponding pixel is obtained for generating the spatial model;
- Figure here can also use deep learning technology to predict the depth of each pixel in the model image, calculate or directly use deep learning technology to predict the normal direction of each pixel, or predict the position of the wall and the outline of the room to generate each space Model.
- the model generation step S63 can be implemented locally or by a remote server.
- a remote server receives and sends model images of each space through the network, and generates each model image based on the model images taken in each space. Model of space.
- Model splicing step S64 based on the position and orientation information of each space in the travel route, splicing the models of each space in the same coordinate system to form an overall model formed by splicing the respective models of each space.
- a transformation matrix is used to convert the local coordinates of a single space model into global world coordinates, so as to obtain an overall model of all shooting points.
- the model splicing step S64 can also be implemented locally or by a remote server.
- a remote server receives the position and orientation information of each spatial model sent through the network, and completes it based on the position and orientation information. Stitching process to generate the overall model.
- the space model can be a three-dimensional space model, of course, it can also be a two-dimensional plane model; here, the method of generating a two-dimensional plane model can be, for example, generating a two-dimensional plane model of each space in the model generation step S63, and then splicing the models.
- step S64 based on the position and orientation information of each space in the travel route, the two-dimensional plane models of each space are spliced in the same coordinate system to form an overall two-dimensional plane model; of course, the method of generating a two-dimensional plane model
- the model generation step S63 and the model splicing step S64 a separate three-dimensional space model and an overall three-dimensional space model are respectively generated, and then the overall three-dimensional space model is converted into an overall two-dimensional plane model, which is not limited .
- the present disclosure provides a space model generation device, including:
- the travel route generation device 701 includes one or more embodiments of the travel route generation device in the present disclosure, to generate a travel route;
- the travel route generation device 701 can also generate travel routes in other ways, which are not limited, for example, travel routes are generated by manual splicing, or can be generated based on pre-input route position and direction information, or, for example, are equipped with acceleration sensors and speeds. Sensors and the like provide acceleration information and movement speed information to generate travel routes.
- a receiving device 702 which receives multiple sets of model images respectively taken from multiple spaces;
- each of the multiple spaces may correspond to a set of model images, and of course, each space may also correspond to multiple sets of model images, which is not limited.
- the model generation module 703 determines the corresponding pixels by performing image comparison on the model images, and obtains the depth information of each corresponding pixel for generating the spatial model; for the panoramic image, here It is also possible to predict the depth of each pixel in the model image through deep learning technology, calculate or directly use deep learning technology to predict the normal direction of each pixel, or predict the position of the wall and the outline of the room to generate various spatial models.
- the model splicing module 704 based on the position and orientation information of each space in the travel route, performs splicing processing of the models of each space in the same coordinate system to form an overall model formed by splicing the respective models of each space.
- the model stitching module 704 converts the local coordinates of a single space model into global world coordinates, for example, according to the position and orientation information of each space, for example, using a conversion matrix, so as to obtain the coordinates of all the shooting points.
- the overall model converts the local coordinates of a single space model into global world coordinates, for example, according to the position and orientation information of each space, for example, using a conversion matrix, so as to obtain the coordinates of all the shooting points.
- the space model can be a three-dimensional space model, of course, it can also be a two-dimensional plane model; here, for example, a two-dimensional plane model of each space can be generated by the model generation module 703, and then the model splicing module 704 is based on the travel route of each space.
- the position and orientation information in each space are stitched together in the same coordinate system to form an overall two-dimensional plane model; of course, it can also be generated by the model generation module 703 and the model stitching module 704 respectively.
- a separate three-dimensional space model and an overall three-dimensional space model are formed, and the model generation module 703 or the model splicing module 704 converts the overall three-dimensional space model into an overall two-dimensional plane model, which is not limited.
- the present disclosure provides a travel route generation system, including:
- a model image capturing device 801 which captures a model image for generating a space model in the space in which it is located;
- the model image shooting device 801 has, for example, a positioning sensor and a direction sensor, and can obtain positioning information and shooting direction information when shooting a model image in the space in which it is located.
- it may have a binocular lens to shoot model images at the same shooting point.
- a panoramic camera to shoot and generate a panoramic image as a model image.
- the model generation module 802 generates models of each space based on the model images respectively captured by the model image capturing device for multiple spaces; here, the function of the model generation module 802 may be the same as the model generation module 703 in the above-mentioned space model generation device, for example It will not be repeated here, but it is not limited. Of course, it can have other functions according to the cooperation with other modules in the space model generation system in this embodiment.
- the travel route generation system 803 includes a system of one or more embodiments of the travel route generation system in the present disclosure to generate a travel route;
- the model splicing module 804 based on the position and orientation information of each space in the travel route, splices the models of each space in the same coordinate system to form an overall model formed by splicing the respective models of each space.
- the function of the model splicing module 804 can be, for example, the same as the model splicing module 704 in the above-mentioned space model generation device, which will not be repeated here, but is not limited. Of course, it can be based on other modules in the space model generation system in this embodiment.
- the coordination has other functions.
- FIG. 9 shows a schematic structural diagram of an electronic device (for example, the terminal device or the server in FIG. 1) 900 suitable for implementing the embodiments of the present disclosure.
- the terminal device in the embodiment of the present disclosure may be various terminal devices in the above-mentioned system.
- the electronic device shown in FIG. 9 is only an example, and should not bring any limitation to the function and scope of use of the embodiments of the present disclosure.
- the electronic device 900 may include a processing device (such as a central processing unit, a graphics processor, etc.) 901 for controlling the overall operation of the electronic device.
- the processing device may include one or more processors to execute instructions to complete all or part of the steps of the above-mentioned method.
- the processing device 901 may also include one or more modules for processing interaction with other devices.
- the storage device 902 is used to store various types of data.
- the storage device 902 may include various types of computer-readable storage media or a combination thereof.
- it may be an electrical, magnetic, optical, electromagnetic, infrared, or semiconductor system, Device or device, or any combination of the above.
- More specific examples of computer-readable storage media may include, but are not limited to: electrical connections with one or more wires, portable computer disks, hard disks, random access memory (RAM), read-only memory (ROM), erasable removable Programmable read-only memory (EPROM or flash memory), optical fiber, portable compact disk read-only memory (CD-ROM), optical storage device, magnetic storage device, or any suitable combination of the above.
- a computer-readable storage medium may be any tangible medium that contains or stores a program, and the program may be used by or in combination with an instruction execution system, apparatus, or device.
- the sensor device 903 is used to sense prescribed measured information and convert it into a usable output signal according to a certain rule, and may include one or more sensors.
- it may include an acceleration sensor, a gyroscope sensor, a magnetic sensor, a pressure sensor or a temperature sensor, etc., which are used to detect changes in the on/off state, relative positioning, acceleration/deceleration, temperature, humidity, and light of the electronic device.
- the processing device 901, the storage device 902, and the sensor device 903 are connected to each other through a bus 904.
- An input/output (I/O) interface 905 is also connected to the bus 904.
- the multimedia device 906 may include input devices such as a touch screen, a touch pad, a keyboard, a mouse, a camera, a microphone, etc., to receive input signals from a user. Various input devices may cooperate with various sensors of the aforementioned sensor device 903 to complete, for example, gesture operations. Input, image recognition input, distance detection input, etc.; the multimedia device 906 may also include an output device such as a liquid crystal display (LCD), a speaker, a vibrator, and the like.
- LCD liquid crystal display
- the power supply device 907 is used to provide power to various devices in the electronic equipment, and may include a power management system, one or more power supplies, and components that distribute power to other devices.
- the communication device 908 may allow the electronic device 900 to perform wireless or wired communication with other devices to exchange data.
- All the above-mentioned devices can also be connected to the I/O interface 905 to implement the application of the electronic device 900.
- FIG. 9 shows an electronic device having various devices, it should be understood that it is not required to implement or have all of the illustrated devices. It may be implemented alternatively or provided with more or fewer devices.
- an embodiment of the present disclosure includes a computer program product, which includes a computer program carried on a non-transitory computer readable medium, and the computer program contains program code for executing the method shown in the flowchart.
- the computer program can be downloaded and installed from a network through a communication device, or can be installed from a storage device.
- the processing device When the computer program is executed by the processing device, the above-mentioned functions defined in the method of the embodiment of the present disclosure are executed.
- a machine-readable medium may be a tangible medium, which may contain or store a program for use by the instruction execution system, apparatus, or device or in combination with the instruction execution system, apparatus, or device.
- a computer-readable medium in the present disclosure may be a computer-readable signal medium or a computer-readable storage medium, or any combination of the two.
- a computer-readable signal medium may include a data signal propagated in a baseband or as a part of a carrier wave, and a computer-readable program code is carried therein. This propagated data signal can take many forms, including but not limited to electromagnetic signals, optical signals, or any suitable combination of the foregoing.
- the computer-readable signal medium may also be any computer-readable medium other than the computer-readable storage medium.
- the computer-readable signal medium may send, propagate, or transmit the program for use by or in combination with the instruction execution system, apparatus, or device .
- the program code contained on the computer-readable medium can be transmitted by any suitable medium, including but not limited to: wire, optical cable, RF (Radio Frequency), etc., or any suitable combination of the above.
- the above-mentioned computer-readable medium may be included in the above-mentioned electronic device; or it may exist alone without being assembled into the electronic device.
- the computer program code used to perform the operations of the present disclosure can be written in one or more programming languages or a combination thereof.
- the above-mentioned programming languages include but are not limited to object-oriented programming languages such as Java, Smalltalk, C++, and Including conventional procedural programming languages-such as "C" language or similar programming languages.
- the program code can be executed entirely on the user's computer, partly on the user's computer, executed as an independent software package, partly on the user's computer and partly executed on a remote computer, or entirely executed on the remote computer or server.
- the remote computer may be connected to the user's computer through any kind of network, or may be connected to an external computer (for example, using an Internet service provider to connect through the Internet).
- each block in the flowchart or block diagram can represent a module, program segment, or part of code, and the module, program segment, or part of code contains one or more for realizing the specified logic function.
- Executable instructions can also occur in a different order from the order marked in the drawings. For example, two blocks shown one after another can actually be executed substantially in parallel, and they can sometimes be executed in the reverse order, depending on the functions involved.
- each block in the block diagram and/or flowchart, and the combination of the blocks in the block diagram and/or flowchart can be implemented by a dedicated hardware-based system that performs the specified functions or operations Or it can be realized by a combination of dedicated hardware and computer instructions.
- the units involved in the embodiments described in the present disclosure can be implemented in software or hardware. Among them, the name of the unit does not constitute a limitation on the unit itself under certain circumstances.
- exemplary types of hardware logic components include: Field Programmable Gate Array (FPGA), Application Specific Integrated Circuit (ASIC), Application Specific Standard Product (ASSP), System on Chip (SOC), Complex Programmable Logical device (CPLD) and so on.
- FPGA Field Programmable Gate Array
- ASIC Application Specific Integrated Circuit
- ASSP Application Specific Standard Product
- SOC System on Chip
- CPLD Complex Programmable Logical device
- a method for generating a travel route characterized in that the method includes:
- the first image shooting step moving from the first shooting point and shooting multiple images
- a first travel route generating step matching the feature points of the multiple images taken in the first image capturing step to generate a first travel route
- the second image shooting step moving from the second shooting point and shooting multiple images
- a second travel route generating step matching the feature points of the multiple images taken in the second image capturing step to generate a second travel route
- the feature points of the image on the second travel route are matched with the feature points of the image on the first travel route, so that the second travel route matches the first travel route. Stitching of travel routes.
- a method for generating a travel route which is characterized in that it further includes:
- a saving step saving at least a part of the information of the plurality of images taken in the first image shooting step
- At least part of the information of the plurality of images includes at least one feature point of the image
- the feature points of the image on the second travel route are matched with the feature points of the image saved in the saving step, so that the second travel route matches the The first travel route is spliced.
- a method for generating a travel route which is characterized in that it further includes:
- the prompting step is to prompt a point on the first travel route before the shooting point of the image in which the at least part of the information was last saved in the saving step as the second shooting point.
- a method for generating a travel route which is characterized in that it further includes:
- the first positioning step at least positioning and recording the position information of the shooting point of an image saved in the saving step
- the second positioning step is to locate and record the position information of at least one shooting point on the second travel route
- the location information and the location information that are located and recorded in the first positioning step are used.
- the location information located and recorded in the second positioning step is spliced between the first travel route and the second travel route.
- a travel route generation device including:
- the receiving module receives multiple sets of images that are moved from different shooting points and shot separately;
- the travel route generation module respectively matches the feature points of the multiple sets of images to generate multiple travel routes respectively;
- the splicing module matches the feature points of different groups of images, so that the multiple travel routes are spliced.
- an apparatus for generating a travel route characterized in that the multiple sets of images include at least a first set of images and a second set of images, and the apparatus further includes:
- a saving module for saving at least part of the information of the first group of images received by the receiving module
- At least a part of the information of the first group of images includes at least a feature point of an image of the first group of images
- the stitching module matches the feature points of the second group of images with the feature points of the images saved in the saving module, so as to stitch the multiple travel routes.
- a travel route generation device which is characterized in that it further includes:
- a prompting module that prompts the point before the shooting point of the image that last saved the at least a part of the information in the travel route corresponding to the first group of images in the storage module as the starting point of the second group of images Shooting point.
- a travel route generation device which is characterized in that it further includes:
- a positioning module that at least locates and records the position information of a shooting point of an image saved in the saving module; and at least locates and records the position information of a shooting point of the second group of images;
- a travel route generation system including:
- Image shooting module which moves from different shooting points and shoots multiple sets of images respectively
- the travel route generation module respectively matches the feature points of the multiple sets of images to generate multiple travel routes respectively;
- the splicing module matches the feature points of different groups of images, so that the multiple travel routes are spliced.
- a travel route generation system which is characterized in that:
- the multiple sets of images include at least a first set of images and a second set of images
- the travel route generation module matches the characteristic points of the first group of images to generate a first travel route; and matches the characteristic points of the second group of images to generate a second travel route;
- the stitching module matches the feature points of the second group of images with the feature points of the first group of images, so as to stitch the second travel route with the first travel route.
- a travel route generation system which is characterized in that it further includes:
- a saving module for saving at least part of the information of the first group of images
- At least part of the information of the first group of images includes at least feature points of an image of the first group of images
- the splicing module matches the feature points of the second group of images with the feature points of the images saved by the saving module, so that the second travel route and the first travel route are spliced.
- a travel route generation system which is characterized in that it further includes:
- the prompting module prompts the point before the shooting point of the image in which the storage module last saved the at least part of the information on the first travel route as the second shooting point.
- a positioning module that at least locates and records the location information of a shooting point of an image saved by the saving module; and at least locates and records the location information of a shooting point on the second travel route;
- the position information located and recorded by the positioning module is used to perform the first travel route and the The splicing of the second travel route.
- a space model generation method including:
- a travel route generation step using the travel route generation method as described in any one of the preceding items to generate a travel route
- a model image shooting step in the process of moving according to the travel route, shooting a model image for generating the space model in the space in which it is located;
- the model generation step is to generate a model of each space based on the model images taken in each space;
- the models of each of the spaces are spliced in the same coordinate system to form a splicing result of the respective models of each of the spaces.
- a space model generation device including:
- the travel route generating device as described in any one of the preceding items to generate a travel route
- a receiving device which receives multiple groups of model images taken separately from multiple spaces
- a model generating module based on the multiple sets of model images received by the receiving device, respectively generating models of each of the spaces;
- the model splicing module based on the position and orientation information of each space in the travel route, performs splicing processing on the models of each space in the same coordinate system to form a splicing result of the respective models of each space. Into the overall model.
- a spatial model generation system including:
- a model image capturing device which captures a model image for generating the space model in the space in which it is located;
- a model generation module based on the model images respectively captured by the model image capturing device for a plurality of the spaces, respectively generating models of each of the spaces;
- the travel route generation system as described in any one of the preceding items to generate travel routes
- the model splicing module based on the position and orientation information of each space in the travel route, performs splicing processing on the models of each space in the same coordinate system to form a splicing result of the respective models of each space. Into the overall model.
- a computer device which is characterized in that it includes a memory and a processor, the memory stores a computer program, and the processor executes the computer program as described above. Any of the methods.
- a computer-readable storage medium characterized in that a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a processor, Any of the methods.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Navigation (AREA)
Abstract
Description
Claims (17)
- 一种行进路线生成方法,包括:第一图像拍摄步骤,从第一拍摄点起移动并拍摄多张图像;第一行进路线生成步骤,对在所述第一图像拍摄步骤中拍摄的多张所述图像的特征点进行匹配,以生成第一行进路线;第二图像拍摄步骤,从第二拍摄点起移动并拍摄多张图像;第二行进路线生成步骤,对在所述第二图像拍摄步骤中拍摄的多张所述图像的特征点进行匹配,以生成第二行进路线;拼接步骤,将在所述第二行进路线上的所述图像的特征点与在所述第一行进路线上的所述图像的特征点进行匹配,以使所述第二行进路线与所述第一行进路线拼接。
- 如权利要求1所述的行进路线生成方法,其特征在于,还包括:保存步骤,保存在所述第一图像拍摄步骤中拍摄的多张所述图像的至少一部分信息;所述多张所述图像的至少一部分信息至少包括一张所述图像的特征点;在所述拼接步骤中,对在所述第二行进路线上的所述图像的特征点与在所述保存步骤中保存的所述图像的特征点进行匹配,以使所述第二行进路线与所述第一行进路线拼接。
- 如权利要求2所述的行进路线生成方法,其特征在于,还包括:提示步骤,提示所述第一行进路线上的、在所述保存步骤中最后保存所述至少一部分信息的图像的拍摄点以前的点,作为所述第二拍摄点。
- 如权利要求1至3的任一项所述的行进路线生成方法,其特征在于,还包括:第一定位步骤,至少定位并记录在所述保存步骤中保存的一张图像的拍摄点的位置信息;第二定位步骤,至少定位并记录所述第二行进路线上的一个拍摄点的位置信息;在所述第二行进路线上的图像的特征点与在所述保存步骤中保存的 图像的特征点无法进行匹配的情况下,利用所述第一定位步骤中定位并记录的位置信息和所述第二定位步骤中定位并记录的位置信息,进行所述第一行进路线和所述第二行进路线的拼接。
- 一种行进路线生成装置,包括:接收模块,接收从不同拍摄点起移动并分别拍摄的多组图像;行进路线生成模块,分别对所述多组图像的特征点进行匹配,以分别生成多条行进路线;拼接模块,对不同组图像的特征点进行匹配,以使所述多条行进路线进行拼接。
- 如权利要求5所述的行进路线生成装置,其特征在于,所述多组图像至少包括第一组图像和第二组图像,所述装置还包括:保存模块,保存所述接收模块接收的所述第一组图像的至少一部分信息;所述第一组图像的至少一部分信息至少包括所述第一组图像中的一张图像的特征点;所述拼接模块对所述第二组图像的特征点与所述保存模块中保存的图像的特征点进行匹配,以使所述多条行进路线拼接。
- 如权利要求6所述的行进路线生成装置,其特征在于,还包括:提示模块,提示所述第一组图像对应的行进路线中的、所述保存模块中最后保存所述至少一部分信息的图像的拍摄点以前的点,作为所述第二组图像的起始拍摄点。
- 如权利要求5至7的任一项所述的行进路线生成装置,其特征在于,还包括:定位模块,至少定位并记录所述保存模块中保存的一张图像的拍摄点的位置信息;以及至少定位并记录所述第二组图像的一个拍摄点的位置信息;在所述第二组图像的特征点与在所述保存模块中保存的图像的特征点无法进行匹配的情况下,利用所述定位模块中定位并记录的位置信息进行所述行进路线的拼接。
- 一种行进路线生成系统,包括:图像拍摄模块,从不同拍摄点起移动并分别拍摄多组图像;行进路线生成模块,分别对所述多组图像的特征点进行匹配,以分别生成多条行进路线;拼接模块,对不同组图像的特征点进行匹配,以使所述多条行进路线进行拼接。
- 如权利要求9所述的行进路线生成系统,其特征在于,所述多组图像至少包括第一组图像和第二组图像,所述图像拍摄模块从第一拍摄点起移动并拍摄所述第一组图像;以及从第二拍摄点起移动并拍摄所述第二组图像;所述行进路线生成模块对所述第一组图像的特征点进行匹配,以生成第一行进路线;以及对所述第二组图像的特征点进行匹配,以生成第二行进路线;所述拼接模块将所述第二组图像的特征点与所述第一组图像的特征点进行匹配,以使所述第二行进路线与所述第一行进路线拼接。
- 如权利要求10所述的行进路线生成系统,其特征在于,还包括,保存模块,保存所述第一组图像的至少一部分信息;所述第一组图像的至少一部分信息至少包括所述第一组图像的一张图像的特征点;所述拼接模块对所述第二组图像的特征点与所述保存模块保存的图像的特征点进行匹配,以使所述第二行进路线与所述第一行进路线拼接。
- 如权利要求11所述的行进路线生成系统,其特征在于,还包括,提示模块,提示所述第一行进路线上的、所述保存模块最后保存所述至少一部分信息的图像的拍摄点以前的点,作为所述第二拍摄点。定位模块,至少定位并记录所述保存模块保存的一张图像的拍摄点的位置信息;以及至少定位并记录所述第二行进路线上的一个拍摄点的位置信息;在所述第二组图像的特征点与所述保存模块保存的图像的特征点无法进行匹配的情况下,利用所述定位模块定位并记录的位置信息进行所述第一行进路线和所述第二行进路线的拼接。
- 一种空间模型生成方法,包括:行进路线生成步骤,使用如权利要求1-4任一项所述的行进路线生成方法以生成行进路线;模型图像拍摄步骤,在根据所述行进路线进行移动的过程中对所处空间拍摄用于生成所述空间模型的模型图像;模型生成步骤,分别基于各个所述空间拍摄的所述模型图像,生成各个所述空间的模型;模型拼接步骤,基于各个所述空间的在所述行进路线中的位置和朝向信息,将各个所述空间的模型在同一个坐标系内进行拼接处理,形成由各个所述空间各自的模型拼接而成的整体模型。
- 一种空间模型生成装置,包括:如权利要求5-8中任一项所述的行进路线生成装置以生成行进路线;接收装置,接收对多个空间分别拍摄的多组模型图像;模型生成模块,基于所述接收装置接收的所述多组模型图像,分别生成各个所述空间的模型;模型拼接模块,基于各个所述空间的在所述行进路线中的位置和朝向信息,将各个所述空间的模型在同一个坐标系内进行拼接处理,形成由各个所述空间各自的模型拼接而成的整体模型。
- 一种空间模型生成系统,包括:模型图像拍摄装置,对所处空间拍摄用于生成所述空间模型的模型图像;模型生成模块,基于所述模型图像拍摄装置针对多个所述空间分别拍摄的所述模型图像,分别生成各个所述空间的模型;如权利要求9-12的任一项所述的行进路线生成系统以生成行进路线;模型拼接模块,基于各个所述空间的在所述行进路线中的位置和朝向信息,将各个所述空间的模型在同一个坐标系内进行拼接处理,形成由各个所述空间各自的模型拼接而成的整体模型。
- 一种计算机设备,包括存储器和处理器,所述存储器中存储有计算机程序,所述处理器执行所述计算机程序时实现如权利要求1-4或 13中任一项所述的方法。
- 一种计算机可读存储介质,其特征在于,所述计算机可读存储介质上存储有计算机程序,所述计算机程序被处理器执行时实现如权利要求1-4或13中任一项所述的方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2020/072814 WO2021142787A1 (zh) | 2020-01-17 | 2020-01-17 | 行进路线及空间模型生成方法、装置、系统 |
CN202080000591.6A CN111433809B (zh) | 2020-01-17 | 2020-01-17 | 行进路线及空间模型生成方法、装置、系统 |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
PCT/CN2020/072814 WO2021142787A1 (zh) | 2020-01-17 | 2020-01-17 | 行进路线及空间模型生成方法、装置、系统 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021142787A1 true WO2021142787A1 (zh) | 2021-07-22 |
Family
ID=71559084
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/072814 WO2021142787A1 (zh) | 2020-01-17 | 2020-01-17 | 行进路线及空间模型生成方法、装置、系统 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN111433809B (zh) |
WO (1) | WO2021142787A1 (zh) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2022051899A1 (zh) * | 2020-09-08 | 2022-03-17 | 上海亦我信息技术有限公司 | 移动路线的处理方法、装置、终端和存储介质 |
CN114500831A (zh) * | 2021-12-30 | 2022-05-13 | 北京城市网邻信息技术有限公司 | 图像采集过程中的提示方法、装置、电子设备及存储介质 |
CN114511622A (zh) * | 2021-12-30 | 2022-05-17 | 北京城市网邻信息技术有限公司 | 全景图像采集方法、装置、电子终端及介质 |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102298779A (zh) * | 2011-08-16 | 2011-12-28 | 淮安盈科伟力科技有限公司 | 全景辅助泊车系统的图像配准方法 |
KR101214471B1 (ko) * | 2011-04-11 | 2012-12-24 | 주식회사 이미지넥스트 | 감시 영상의 3차원 정합 방법 및 시스템 |
CN107248169A (zh) * | 2016-03-29 | 2017-10-13 | 中兴通讯股份有限公司 | 图像定位方法及装置 |
US20170309010A1 (en) * | 2016-04-25 | 2017-10-26 | Alstom Transport Technologies | Assembly completeness inspection method using active ranging |
KR101908952B1 (ko) * | 2017-02-27 | 2018-12-19 | (주)진명아이앤씨 | Uhd동영상 스티칭 방법 및 그 장치 |
US20190054937A1 (en) * | 2017-08-15 | 2019-02-21 | Bnsf Railway Company | Unmanned aerial vehicle system for inspecting railroad assets |
CN110505463A (zh) * | 2019-08-23 | 2019-11-26 | 上海亦我信息技术有限公司 | 基于拍照的实时自动3d建模方法 |
CN110532962A (zh) * | 2019-08-30 | 2019-12-03 | 上海秒针网络科技有限公司 | 轨迹的检测方法及装置、存储介质和电子装置 |
Family Cites Families (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3712857A1 (en) * | 2011-02-22 | 2020-09-23 | Midmark Corporation | Hybrid stitching for 3d reconstruction |
CN104135617B (zh) * | 2014-07-29 | 2017-07-14 | 努比亚技术有限公司 | 物体运动轨迹拍摄方法、终端及系统 |
CN108961395B (zh) * | 2018-07-03 | 2019-07-30 | 上海亦我信息技术有限公司 | 一种基于拍照重建三维空间场景的方法 |
CN110334568B (zh) * | 2019-03-30 | 2022-09-16 | 深圳市晓舟科技有限公司 | 轨迹生成与监控方法、装置、设备及存储介质 |
-
2020
- 2020-01-17 WO PCT/CN2020/072814 patent/WO2021142787A1/zh active Application Filing
- 2020-01-17 CN CN202080000591.6A patent/CN111433809B/zh active Active
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR101214471B1 (ko) * | 2011-04-11 | 2012-12-24 | 주식회사 이미지넥스트 | 감시 영상의 3차원 정합 방법 및 시스템 |
CN102298779A (zh) * | 2011-08-16 | 2011-12-28 | 淮安盈科伟力科技有限公司 | 全景辅助泊车系统的图像配准方法 |
CN107248169A (zh) * | 2016-03-29 | 2017-10-13 | 中兴通讯股份有限公司 | 图像定位方法及装置 |
US20170309010A1 (en) * | 2016-04-25 | 2017-10-26 | Alstom Transport Technologies | Assembly completeness inspection method using active ranging |
KR101908952B1 (ko) * | 2017-02-27 | 2018-12-19 | (주)진명아이앤씨 | Uhd동영상 스티칭 방법 및 그 장치 |
US20190054937A1 (en) * | 2017-08-15 | 2019-02-21 | Bnsf Railway Company | Unmanned aerial vehicle system for inspecting railroad assets |
CN110505463A (zh) * | 2019-08-23 | 2019-11-26 | 上海亦我信息技术有限公司 | 基于拍照的实时自动3d建模方法 |
CN110532962A (zh) * | 2019-08-30 | 2019-12-03 | 上海秒针网络科技有限公司 | 轨迹的检测方法及装置、存储介质和电子装置 |
Also Published As
Publication number | Publication date |
---|---|
CN111433809A (zh) | 2020-07-17 |
CN111433809B (zh) | 2021-08-27 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021036353A1 (zh) | 基于拍照的3d建模系统及方法、自动3d建模装置及方法 | |
US8661053B2 (en) | Method and apparatus for enabling virtual tags | |
US10389938B2 (en) | Device and method for panoramic image processing | |
WO2021142787A1 (zh) | 行进路线及空间模型生成方法、装置、系统 | |
US20230360337A1 (en) | Virtual image displaying method and apparatus, electronic device and storage medium | |
EP2974509B1 (en) | Personal information communicator | |
WO2023051185A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
WO2022082704A1 (zh) | 模型修正方法、装置、设备 | |
WO2022007565A1 (zh) | 增强现实的图像处理方法、装置、电子设备及存储介质 | |
WO2021027596A1 (zh) | 图像特效处理方法、装置、电子设备和计算机可读存储介质 | |
US20210407052A1 (en) | Method for processing image, related device and storage medium | |
TW202102026A (zh) | 坐標系對齊的方法及裝置、電子設備和計算機可讀存儲介質 | |
CN114554092A (zh) | 设备控制方法、装置和电子设备 | |
WO2021027547A1 (zh) | 图像特效处理方法、装置、电子设备和计算机可读存储介质 | |
WO2023231918A1 (zh) | 图像处理方法、装置、电子设备及存储介质 | |
WO2020155908A1 (zh) | 用于生成信息的方法和装置 | |
US11810336B2 (en) | Object display method and apparatus, electronic device, and computer readable storage medium | |
CN115941841A (zh) | 关联信息展示方法、装置、设备、存储介质和程序产品 | |
CN109472873B (zh) | 三维模型的生成方法、装置、硬件装置 | |
US11880919B2 (en) | Sticker processing method and apparatus | |
CN111696214A (zh) | 房屋展示方法、装置和电子设备 | |
WO2024016880A1 (zh) | 信息交互方法、装置、电子设备和存储介质 | |
US12019669B2 (en) | Method, apparatus, device, readable storage medium and product for media content processing | |
RU2802724C1 (ru) | Способ и устройство обработки изображений, электронное устройство и машиночитаемый носитель информации | |
US20240119082A1 (en) | Method, apparatus, device, readable storage medium and product for media content processing |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20913201 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20913201 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20913201 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 02.02.2023) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20913201 Country of ref document: EP Kind code of ref document: A1 |