WO2022147976A1 - 三维重建及相关交互、测量方法和相关装置、设备 - Google Patents
三维重建及相关交互、测量方法和相关装置、设备 Download PDFInfo
- Publication number
- WO2022147976A1 WO2022147976A1 PCT/CN2021/102882 CN2021102882W WO2022147976A1 WO 2022147976 A1 WO2022147976 A1 WO 2022147976A1 CN 2021102882 W CN2021102882 W CN 2021102882W WO 2022147976 A1 WO2022147976 A1 WO 2022147976A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- data set
- image
- dimensional
- preset
- data
- Prior art date
Links
- 238000000034 method Methods 0.000 title claims abstract description 118
- 230000003993 interaction Effects 0.000 title claims abstract description 20
- 238000000691 measurement method Methods 0.000 title claims abstract description 20
- 238000005457 optimization Methods 0.000 claims abstract description 149
- 238000012545 processing Methods 0.000 claims abstract description 16
- 230000009466 transformation Effects 0.000 claims description 113
- 238000005259 measurement Methods 0.000 claims description 39
- 230000006870 function Effects 0.000 claims description 34
- 238000003384 imaging method Methods 0.000 claims description 19
- 239000013598 vector Substances 0.000 claims description 19
- 230000002452 interceptive effect Effects 0.000 claims description 15
- 238000012216 screening Methods 0.000 claims description 13
- 238000013507 mapping Methods 0.000 claims description 11
- 238000007621 cluster analysis Methods 0.000 claims description 7
- 230000000007 visual effect Effects 0.000 claims description 7
- 238000010276 construction Methods 0.000 claims description 6
- 230000005484 gravity Effects 0.000 claims description 6
- 238000004458 analytical method Methods 0.000 claims description 5
- 238000004590 computer program Methods 0.000 claims description 4
- 230000001131 transforming effect Effects 0.000 claims 1
- 230000000694 effects Effects 0.000 abstract description 26
- 238000004364 calculation method Methods 0.000 abstract description 13
- 239000011159 matrix material Substances 0.000 description 17
- 238000010586 diagram Methods 0.000 description 12
- 230000008569 process Effects 0.000 description 12
- 230000003190 augmentative effect Effects 0.000 description 6
- 238000004422 calculation algorithm Methods 0.000 description 6
- 230000009286 beneficial effect Effects 0.000 description 5
- 239000011505 plaster Substances 0.000 description 5
- 210000000988 bone and bone Anatomy 0.000 description 4
- 230000008030 elimination Effects 0.000 description 4
- 238000003379 elimination reaction Methods 0.000 description 4
- 230000004927 fusion Effects 0.000 description 4
- 238000013519 translation Methods 0.000 description 4
- 241001465754 Metazoa Species 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 241001494479 Pecora Species 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 230000001186 cumulative effect Effects 0.000 description 2
- 238000000354 decomposition reaction Methods 0.000 description 2
- 238000003709 image segmentation Methods 0.000 description 2
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 description 1
- 238000010146 3D printing Methods 0.000 description 1
- 230000001154 acute effect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000008846 dynamic interplay Effects 0.000 description 1
- 238000000605 extraction Methods 0.000 description 1
- 238000007667 floating Methods 0.000 description 1
- 229910052602 gypsum Inorganic materials 0.000 description 1
- 239000010440 gypsum Substances 0.000 description 1
- 210000003128 head Anatomy 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000003058 natural language processing Methods 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/50—Depth or shape recovery
- G06T7/55—Depth or shape recovery from multiple images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2200/00—Indexing scheme for image data processing or generation, in general
- G06T2200/08—Indexing scheme for image data processing or generation, in general involving all processing steps from image acquisition to 3D model generation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10016—Video; Image sequence
Definitions
- the present disclosure relates to the technical field of computer vision, and in particular, to a three-dimensional reconstruction and related interaction and measurement method, and related devices and equipment.
- mobile terminals such as mobile phones, tablet computers and other integrated camera devices to perform 3D reconstruction of objects in real scenes, so as to use the 3D reconstruction obtained by 3D reconstruction.
- the model implements applications such as Augmented Reality (AR) and games on mobile terminals.
- AR Augmented Reality
- the present disclosure provides a three-dimensional reconstruction method and related devices and equipment.
- a first aspect of the present disclosure provides a three-dimensional reconstruction method, including: acquiring multiple frames of images to be processed obtained by scanning a target to be reconstructed by a camera device; The target pixel points of the target to be reconstructed and their camera pose parameters; according to the preset division strategy, the image data of each frame of the to-be-processed image is sequentially divided into corresponding data sets, wherein the image data at least includes the target pixel points; The image data of the data set, and the image data and pose optimization parameters of the data set whose time sequence is located before it, determine the pose optimization parameters of each data set; using the pose optimization parameters of each data set, Adjust the camera pose parameters of the to-be-processed image to which the image data in the image data belongs; use the preset three-dimensional reconstruction method and the adjusted camera pose parameters of the to-be-processed image to reconstruct the image data of the to-be-processed image to obtain the to-be-reconstructed target. 3D model.
- the image to be processed obtained by scanning the target to be reconstructed by the camera device and the calibration parameters of the camera device are used to determine the target pixels and camera pose parameters of each frame of the image to be processed belonging to the target to be reconstructed, and according to the preset division strategy, Divide the image data of each frame of the image to be processed into the corresponding data sets in turn, so as to sequentially use the image data of each data set, and the image data and pose optimization parameters of the data set before it in time sequence to determine the value of each data set.
- the pose optimization parameters, and then the pose optimization parameters of each data set can be determined based on the pose optimization parameters of the previous data set, so the pose optimization parameters of each data set are used to be included in the data set.
- the preset three-dimensional reconstruction method and the adjusted camera pose of the to-be-processed image are used. parameters, the image data of the image to be processed is reconstructed, and the effect of the obtained 3D model of the target to be reconstructed can be effectively improved, and the error elimination of the camera pose parameters in the unit of the data set can reduce the amount of calculation, which is conducive to reducing Calculate load.
- a second aspect of the present disclosure provides an interaction method based on three-dimensional reconstruction, including: acquiring a three-dimensional model of a target to be reconstructed, wherein the three-dimensional model is obtained by using the three-dimensional reconstruction method in the first aspect; using a preset visual inertial navigation method , build a three-dimensional map of the scene where the camera device is located, and obtain the current pose information of the camera device in the three-dimensional map; based on the pose information, display the three-dimensional model in the scene image currently captured by the camera device.
- the 3D model of the target to be reconstructed is displayed in the currently captured scene image, which can realize the geometric consistency fusion of the virtual object and the real scene, and because of the 3D model
- the model is obtained by the 3D reconstruction method in the above first aspect, so the effect of 3D reconstruction can be improved, thereby improving the effect of geometrically consistent fusion of virtual and reality, which is beneficial to improve user experience.
- a third aspect of the present disclosure provides a measurement method based on three-dimensional reconstruction, including: acquiring a three-dimensional model of a target to be reconstructed, wherein the three-dimensional model is obtained by using the three-dimensional reconstruction method in the first aspect; Multiple set measurement points; obtain distances between multiple measurement points, and obtain distances between positions corresponding to multiple measurement points on the target to be reconstructed.
- the distance between the multiple measurement points is obtained, and the distance between the positions corresponding to the multiple measurement points on the target to be reconstructed is obtained, so as to satisfy the requirements for Measurement requirements of objects in real scenes, and since the 3D model is obtained by using the 3D reconstruction method in the first aspect, the effect of 3D reconstruction can be improved, thereby improving the measurement accuracy.
- a fourth aspect of the present disclosure provides a three-dimensional reconstruction device, including an image acquisition module, a first determination module, a data division module, a second determination module, a parameter adjustment module, and a model reconstruction module, and an image acquisition module for acquiring a scan of a camera device
- the multi-frame to-be-processed images obtained by the target to be reconstructed;
- the first determination module is used to use each frame of the to-be-processed image and the calibration parameters of the imaging device to determine the target pixels of each frame of the to-be-processed image belonging to the target to be reconstructed and its camera pose parameters
- the data division module is used to divide the image data of each frame of images to be processed into corresponding data sets in turn according to the preset division strategy, wherein the image data at least includes target pixels;
- the second determination module sequentially utilizes the images of each data set data, and the image data and pose optimization parameters of the data set before it in time sequence, determine the pose optimization parameters of each data set;
- the parameter adjustment module
- a fifth aspect of the present disclosure provides an interaction device based on three-dimensional reconstruction, including a model acquisition module, a mapping positioning module, and a display interaction module.
- the model acquisition module is used to acquire a three-dimensional model of a target to be reconstructed, wherein the three-dimensional model is obtained by using the above obtained by the three-dimensional reconstruction device in the fourth aspect;
- the mapping and positioning module is used to construct a three-dimensional map of the scene where the camera device is located by using a preset visual inertial navigation method, and obtain the current pose information of the camera device in the three-dimensional map; display interaction
- the module is used to display the 3D model in the scene image currently captured by the camera device based on the pose information.
- a sixth aspect of the present disclosure provides a measurement device based on three-dimensional reconstruction, including a model acquisition module, a display interaction module, and a distance acquisition module, where the model acquisition module is used to acquire a three-dimensional model of a target to be reconstructed, wherein the three-dimensional model is obtained by using the above-mentioned No. obtained by the three-dimensional reconstruction device in the four aspects; the display interaction module is used to receive multiple measurement points set by the user on the three-dimensional model; the distance acquisition module is used to acquire the distances between the multiple measurement points, and obtain the corresponding values on the target to be reconstructed. The distance between the positions of multiple measurement points.
- a seventh aspect of the present disclosure provides an electronic device, including a memory and a processor coupled to each other, where the processor is configured to execute program instructions stored in the memory, so as to implement the three-dimensional reconstruction method in the first aspect, or implement the second The three-dimensional reconstruction-based interaction method in the aspect, or the three-dimensional reconstruction-based measurement method in the above-mentioned third aspect.
- An eighth aspect of the present disclosure provides a computer-readable storage medium on which program instructions are stored, and when the program instructions are executed by a processor, implement the three-dimensional reconstruction method in the first aspect above, or implement the three-dimensional reconstruction method in the second aspect above. Reconstruction interactive method, or implement the three-dimensional reconstruction-based measurement method in the third aspect.
- a ninth aspect of the present disclosure provides a computer program, including computer-readable codes, which, when the computer-readable codes are executed in an electronic device and executed by a processor in the electronic device, implement the above-mentioned first aspect
- a tenth aspect of the present disclosure provides a computer program product that, when run on a computer, causes the computer to execute the three-dimensional reconstruction method in the first aspect above, or the interactive method based on three-dimensional reconstruction in the second aspect above, or execute the The three-dimensional reconstruction-based measurement method in the third aspect.
- the pose optimization parameters of each data set can be determined based on the pose optimization parameters of the previous data set, so the pose optimization parameters of each data set are used for the image data contained in the data set.
- the preset three-dimensional reconstruction method and the adjusted camera pose parameters of the image to be processed are used to The image data of the processed image is reconstructed, and the effect of the obtained 3D model of the target to be reconstructed can be effectively improved, and the error elimination of the camera pose parameters in the unit of the data set can reduce the amount of calculation, thereby helping to reduce the calculation load.
- FIG. 1 is a schematic flowchart of an embodiment of a three-dimensional reconstruction method of the present disclosure
- FIG. 2 is a schematic state diagram of an embodiment of the three-dimensional reconstruction method of the present disclosure
- step S12 in FIG. 1 is a schematic flowchart of an embodiment of step S12 in FIG. 1;
- FIG. 4 is a schematic flowchart of an embodiment of step S13 in FIG. 1;
- FIG. 5 is a schematic flowchart of an embodiment of step S14 in FIG. 1;
- FIG. 6 is a schematic flowchart of an embodiment of step S141 in FIG. 5;
- FIG. 7 is a schematic flowchart of an embodiment of step S142 in FIG. 5;
- FIG. 8 is a schematic flowchart of an embodiment of step S143 in FIG. 5;
- FIG. 9 is a schematic flowchart of an embodiment of the three-dimensional reconstruction-based interaction method of the present disclosure.
- FIG. 10 is a schematic flowchart of an embodiment of a three-dimensional reconstruction-based measurement method of the present disclosure
- FIG. 11 is a schematic diagram of a framework of an embodiment of a three-dimensional reconstruction apparatus of the present disclosure.
- FIG. 12 is a schematic diagram of the framework of an embodiment of the three-dimensional reconstruction-based interaction device of the present disclosure.
- FIG. 13 is a schematic diagram of a framework of an embodiment of a three-dimensional reconstruction-based measurement device of the present disclosure
- FIG. 14 is a schematic diagram of a framework of an embodiment of an electronic device of the present disclosure.
- FIG. 15 is a schematic diagram of a framework of an embodiment of a computer-readable storage medium of the present disclosure.
- system and "network” are often used interchangeably herein.
- the term “at least one of” is only an association relationship to describe related objects, which means that there can be three kinds of relationships, for example, at least one of A and B can mean that A exists alone, and A and B exist at the same time. B, there are three cases of B alone.
- the character “/” in this document generally indicates that the related objects are an “or” relationship.
- “multiple” herein means two or more than two
- 3D reconstruction is an important problem in the field of computer vision and augmented reality, and it plays an important role in applications such as augmented reality on mobile platforms, games, and 3D printing.
- AR effects of real objects are to be realized on mobile platforms, such as skeleton drive, users are usually required to quickly reconstruct real objects in 3D. Therefore, 3D object scanning and reconstruction technology has a wide range of needs in the field of augmented reality on mobile platforms.
- the present disclosure proposes a three-dimensional reconstruction and related interaction, measurement method, and related devices and equipment, by acquiring multiple frames of images to be processed obtained by scanning a target to be reconstructed by a camera device; Each frame of the to-be-processed image belongs to the target pixel point of the target to be reconstructed and its camera pose parameters; the image data of each frame of the to-be-processed image is divided into corresponding data sets in turn; the image data of the data set and the data whose time sequence is located before it are used The image data and pose optimization parameters of the set are used to determine the pose optimization parameters of the data set; the pose optimization parameters of the data set are used to adjust the camera pose parameters of the to-be-processed images to which the image data included in the data set belongs; The image data of the image to be processed is reconstructed to obtain a three-dimensional model of the target to be reconstructed.
- the pose optimization parameters of each data set can be determined based on the pose optimization parameters of the previous data set, so the pose optimization parameters of each data set are used for the images contained in the data set.
- the preset three-dimensional reconstruction method and the adjusted camera pose parameters of the to-be-processed image are used to The image data of the image to be processed is reconstructed, and the effect of the obtained 3D model of the target to be reconstructed can be effectively improved, and the error elimination of the camera pose parameters in the unit of the data set can reduce the amount of calculation, which is beneficial to reduce the calculation load. .
- the execution subject of the three-dimensional reconstruction method, the interactive method of three-dimensional reconstruction, and the measurement method of three-dimensional reconstruction may be an electronic device, wherein the electronic device may be a smart phone, a desktop computer, a tablet computer, a notebook computer, a smart speaker, a digital assistant, an augmented reality (augmented reality, AR)/virtual reality (VR) devices, smart wearable devices and other types of physical devices. It can also be software running on physical devices, such as applications, browsers, etc.
- the operating system running on the physical device may include, but is not limited to, an Android system, an Apple system (IOS Input Output System, IOS), linux, windows, and the like.
- FIG. 1 is a schematic flowchart of an embodiment of a three-dimensional reconstruction method of the present disclosure. , which can include the following steps:
- Step S11 acquiring multiple frames of images to be processed obtained by scanning the target to be reconstructed by the imaging device.
- the camera device may include, but is not limited to, mobile terminals such as mobile phones and tablet computers.
- the steps in the method embodiments of the present disclosure may be performed by a mobile terminal, or may be performed by a processing device such as a microcomputer connected to a camera device with a scanning and shooting function.
- the imaging device may include a color camera capable of sensing visible light and a depth camera capable of sensing the depth of the object to be reconstructed, such as a structured light depth camera.
- a structured light depth camera such as a structured light depth camera.
- Objects to be reconstructed may include, but are not limited to: people, animals, objects (such as statues, furniture, etc.).
- the 3D model of the statue can be finally obtained by scanning the statue.
- the 3D model of the statue can be further rendered and skeleton bound.
- the target to be reconstructed may be determined according to actual application requirements, and is not limited here.
- Step S12 Using each frame of the to-be-processed image and the calibration parameters of the imaging device, determine the target pixels and camera pose parameters of each frame of the to-be-processed image belonging to the target to be reconstructed.
- the calibration parameters may include internal parameters of the imaging device.
- the calibration parameters may include the internal parameters of the color camera; when the imaging device includes a depth camera, or includes a color camera and a depth camera, it can be deduced by analogy. , and no more examples will be given here.
- the internal parameters may include but are not limited to: camera focal length, camera principal point coordinates.
- the internal parameters may be represented in the form of a matrix.
- the internal parameter K of the color camera may be represented as:
- f x , f y represent the focal length of the color camera
- c x , cy represent the principal point coordinates of the color camera.
- the internal parameters of the depth camera It can be deduced in the same way, and no examples are given here.
- the calibration parameters may also include external parameters between the depth camera and the color camera of the imaging device, which are used to represent the transformation from the world coordinate system to the camera coordinate system.
- the external parameters may include a 3*3 rotation matrix R and a 3*1 translation matrix T. Using the rotation matrix R to multiply the coordinate point P world in the world coordinate system to the left, and summing it with the translation matrix T, the corresponding coordinate point P camera in the camera coordinate system of the coordinate point P world in the world coordinate system can be obtained.
- a pre-trained image segmentation model (for example, Unet model) can be used to perform image segmentation on the image to be processed, so as to obtain target pixels belonging to the target to be reconstructed in the to-be-processed image; in another implementation scenario, further The target to be reconstructed can be placed in an environment with a large color difference with it.
- the target to be reconstructed when the target to be reconstructed is a milky white gypsum statue, the target to be reconstructed can be placed in a black environment for scanning, so that the image to be processed belongs to the environment color.
- the pixels of the target color to be reconstructed are marked as invalid, and the pixels belonging to the target color to be reconstructed are marked as valid, and the size of the connected domain formed by the pixels marked as valid is compared, and the largest connected domain is determined.
- the pixels of the reconstructed target when the target to be reconstructed is a milky white gypsum statue, the target to be reconstructed can be placed in a black environment for scanning, so that the image to be processed belongs to the environment color.
- the pixels of the target color to be reconstructed are marked as invalid, and the pixels belonging to the target color to be reconstructed are marked as valid, and the size of the connected domain formed by the pixels marked as valid is compared, and the largest connected domain is determined.
- the camera device In order to obtain a complete 3D model of the target to be reconstructed, the camera device needs to scan the target to be reconstructed in different poses, so the camera pose parameters used for shooting different images to be processed may be different, so in order to eliminate the camera pose parameter error , so as to improve the effect of subsequent 3D reconstruction, it is necessary to first determine the camera pose parameters of each frame of the image to be processed.
- the target pixels of each frame of the image to be processed belonging to the target to be reconstructed can be used and its previous frame to be processed image belongs to the target pixel of the target to be reconstructed
- the internal parameter K of the camera device constructs the objective function of the relative pose parameter ⁇ T, and uses the ICP (Iterative Closest Point, iterative closest point) algorithm to minimize the objective function, so as to obtain the relative pose parameter ⁇ T, where the relative position
- the pose parameter ⁇ T is the relative parameter of the camera pose parameter T t of each frame of the image to be processed relative to the camera pose parameter T t-1 of the preceding frame of the image to be processed.
- the objective function of the relative pose parameter ⁇ T can be referred to the following formula:
- ⁇ is the weight
- d is the depth data Project to color data The depth value of the rear pixel pi . Therefore, in the above formula, w( ⁇ , p i ) can represent the theoretical corresponding pixel point in the three-dimensional space after the pixel point p i of the current frame is transformed to its previous frame by using the relative pose parameter ⁇ T and the internal parameter K.
- the square sum error E geo between the z coordinate value w( ⁇ , p i ) z of the corresponding pixel in the three-dimensional space is also smaller, so minimizing the above objective function E icp can accurately obtain the relative pose parameter ⁇ T,
- the accuracy of the camera pose parameters can be improved.
- the relative pose parameter ⁇ T After obtaining the relative pose parameter ⁇ T between the camera pose parameter T t of each frame of the image to be processed relative to the camera pose parameter T t -1 of the previous frame of the image to be processed, the relative pose parameter ⁇ The inverse of T (ie ⁇ T -1 ) is left-multiplied by the camera pose parameter T t-1 of the image to be processed in the previous frame to obtain the camera pose parameter T t of the image to be processed in the current frame.
- its camera pose parameters can be initialized as a unit matrix.
- the unit matrix is the main pair of A square matrix in which all elements on the corner are 1 and all other elements are 0.
- the scanning of the to-be-processed image and the determination of the target pixel point and the camera pose parameters can also be performed at the same time, that is, after a frame of the to-be-processed image is scanned and obtained, the image to be processed that has just been scanned is scanned. The image is used to determine the target pixel points and the camera pose parameters. At the same time, the next frame of the image to be processed is obtained by scanning, so that the 3D reconstruction of the target to be reconstructed can be performed in real time and online.
- Step S13 According to a preset division strategy, sequentially divide the image data of each frame of the image to be processed into a corresponding data set, wherein the image data at least includes target pixels.
- the maximum number of frames (for example, 8 frames, 9 frames, 10 frames, etc.) of the to-be-processed image to which the image data that each data set can accommodate may be set, so that in the current
- the number of frames of the to-be-processed image to which the image data included in the data set belongs reaches the maximum number of frames, a new data set is created, and the undivided image data of the to-be-processed image continues to be divided into the newly created data set, This cycle continues until the scan is complete.
- the image data of the to-be-processed images that have similar poses can also be divided into the same data set, which is not detailed here. limited.
- it is also possible to determine the pose difference between the to-be-processed image to which the image data belongs and the to-be-processed image of the previous frame for example, the camera orientation angle difference, Whether the camera position distance
- the to-be-processed image to be divided can also be ignored, and the division operation of the image data of the next frame of the to-be-processed image is processed.
- there may be image data belonging to the same image to be processed between adjacent data sets for example, there may be image data belonging to two frames of the same image to be processed between adjacent data sets, or, adjacent data There may also be image data belonging to three identical frames of images to be processed between the sets, which is not limited here.
- the image data of each frame of the image to be processed may only include target pixels belonging to the target to be reconstructed (eg, target pixels in depth data, target pixels in color data); in another implementation scenario , the image data of each frame of the image to be processed may also include pixels that do not belong to the target to be reconstructed.
- the image data divided into the data set may also be the image data of the entire image to be processed. In this case, the image data also The position coordinates of the target pixel can be included, so that the target pixel can be found later.
- FIG. 2 is a schematic state diagram of an embodiment of the three-dimensional reconstruction method of the present disclosure.
- the target to be reconstructed is a portrait plaster sculpture
- each frame of the to-be-processed image 21 may include color data 22 and depth data 23, and the target pixels belonging to the target to be reconstructed are obtained.
- the image data 24 are sequentially divided into corresponding data sets 25 .
- Step S14 Determine the pose optimization parameters of each data set by sequentially using the image data of each data set, and the image data and pose optimization parameters of the data set whose time sequence is located before it.
- the image data of each data set and the image data of the data set before it in time sequence can be used to determine the spatial transformation parameter T icp between the two, so that the spatial transformation parameter between the two can be used.
- T icp , and their respective pose optimization parameters T frag to construct an objective function about the pose optimization parameter T frag , and then solve the objective function to obtain its pose optimization parameter T frag and the pose of the data set whose time sequence is located before it
- the parameters are optimized, so that the pose optimization parameter T frag of the previous data set can be updated.
- the pose optimization parameter T frag of the data set before it in time sequence is considered, that is , between the data set and the pose optimization parameters of the data set before it. They are related to each other, and with the continuous generation of new data sets, the pose optimization parameters of the previous data sets can also be continuously updated, and thus loop to the last data set, so that the final pose optimization parameters of each data set can be obtained, Therefore, the accumulated error can be effectively eliminated.
- the pose optimization parameters of the first data set may be initialized as an identity matrix.
- the pose optimization parameters of the previous data set can be calculated, and the pose optimization parameters of the related data set can be updated, and so on until the end of the scan , to obtain the final pose optimization parameters of each data set, which can help to balance the amount of calculation, and thus help to reduce the calculation load.
- the camera device is a mobile terminal such as a mobile phone or a tablet computer
- the timing sequence may represent the overall shooting sequence of the images to be processed in the data set.
- Other situations can be deduced by analogy, and no examples are given here.
- the image data in the image data set in the data set 25 can also be sequentially Map to 3D space to get 3D point cloud corresponding to each data set.
- the camera pose parameter T t of the image to be processed to which the image data belongs and the internal parameter K of the imaging device can be used to map the image data to a three-dimensional space to obtain a three-dimensional point cloud.
- Three-dimensional homogeneous get the pixel coordinates, and then use the inverse of the camera pose parameter T t Multiply the pixel coordinates after homogeneous with the inverse K -1 of the internal parameter K to obtain a three-dimensional point cloud in three-dimensional space.
- the pose of the dataset can be used to optimize the inverse of the parameter T frag Left-multiply the 3D point cloud for dynamic adjustment.
- the camera pose parameters of the data set can also be used to adjust the corresponding 3D point cloud.
- the three-dimensional point cloud may be marked with a preset color (eg, green), which is not limited herein.
- Step S15 Using the pose optimization parameters of each data set, adjust the camera pose parameters of the to-be-processed image to which the image data included in the data set belongs.
- the inverse of the pose optimization parameter T frag of each data set can be used
- the camera pose parameter T t of the to-be-processed image to which the image data contained therein belongs is left-multiplied, so as to realize the adjustment of the camera pose parameter.
- the sequence of data set A that has been divided into data set A includes image data 01 (belonging to the image to be processed 01 ), image data 02 (belonging to the image to be processed 02 ), and image data 03 (belonging to the image to be processed 03 ), so it is possible to Use the inverse of the pose optimization parameter T frag of dataset A Left-multiply the camera pose parameter T t of the image to be processed 01 , the camera pose parameter T t of the image to be processed 02 , and the camera pose parameter T t of the image to be processed 03 , thereby realizing the image contained in the data set A. Adjustment of the camera pose parameters of the to-be-processed image to which the data belongs.
- the adjacent data set B includes image data 03 (belonging to the to-be-processed image 03) and image data 04 (belonging to the to-be-processed image 04), so when the pose of the data set A is used Optimizing the inverse of the parameter T frag When left-multiplying the camera pose parameter T t of the image to be processed 01, the camera pose parameter T t of the image to be processed 02, and the camera pose parameter T t of the image to be processed 03, then when the image contained in the data set B is When adjusting the camera pose parameters of the to-be-processed image to which the data belongs, the inverse of the pose optimization parameter T frag of the data set B can be used.
- the pose optimization parameters of each data set use the pose optimization parameters of each data set to analyze the images to be processed to which the image data included in the data set belongs.
- the camera pose parameters 26 are adjusted to obtain the adjusted camera pose parameters 27 .
- the adjusted camera pose parameters 27 of the data set can also be used to adjust the corresponding three-dimensional point cloud 28, so that the user can feel the three-dimensional point cloud dynamic adjustment.
- Step S16 Using the preset three-dimensional reconstruction method and the adjusted camera pose parameters of the to-be-processed image, perform reconstruction processing on the image data of the to-be-processed image to obtain a three-dimensional model of the to-be-reconstructed target.
- the preset three-dimensional reconstruction method may include, but is not limited to: a TSDF (Truncated Signed Distance Function, based on a truncated signed distance function) reconstruction method and a Poisson reconstruction method.
- the TSDF reconstruction method is a method for calculating the latent potential surface in the 3D reconstruction, and details are not repeated here.
- the core idea of Poisson reconstruction is that the three-dimensional point cloud represents the surface position of the object to be reconstructed, and its normal vector represents the direction of inside and outside. By implicitly fitting an indicator function derived from an object, a smooth object surface estimation can be obtained. , and details are not repeated here.
- the above steps can be used to reconstruct the 3D model of the target to be reconstructed in real time, and superimposed and rendered at the same position and angle as the currently captured image frame, so that the to-be-reconstructed model can be displayed to the user.
- the main 3D model of the target may also be printed by a three-dimensional printer, so as to obtain a physical model corresponding to the target to be reconstructed.
- the pose optimization parameters of each data set can be determined based on the pose optimization parameters of the previous data set, so the pose optimization parameters of each data set are used for the image data contained in the data set.
- the preset three-dimensional reconstruction method and the adjusted camera pose parameters of the image to be processed are used to The image data of the processed image is reconstructed, and the effect of the obtained 3D model of the target to be reconstructed can be effectively improved, and the error elimination of the camera pose parameters in the unit of the data set can reduce the amount of calculation, thereby helping to reduce the calculation load.
- FIG. 3 is a schematic flowchart of an embodiment of step S12 in FIG. 1 .
- FIG. 3 is a schematic flowchart of the determination process of the target pixel point in FIG. 1, which may include the following steps:
- Step S121 Obtain the angle between the normal vector of each pixel included in the depth data after alignment with the color data and the gravitational direction of the image to be processed.
- each frame of the image to be processed includes color data It and depth data depth data Projection to color data It yields depth data Dt after alignment .
- the depth data can be converted by formula (6) 2D image coordinates of pixels in Use its depth value d t to convert to a three-dimensional homogeneous coordinate P:
- the internal parameters of the depth camera in the imaging device are used After back-projecting the three-dimensional homogeneous coordinate P to the three-dimensional space, use the rotation matrix R and translation matrix t of the depth camera and the color camera to perform rigid transformation, and then use the internal parameter K of the color camera to project to the two-dimensional plane, and obtain the same color data.
- the pixel coordinate P' of the object in the color data is a three-dimensional coordinate.
- its depth value based on formula (8) that is, its third value P'[2 ] is divided by its first value and second value, respectively, to obtain the two-dimensional coordinate x t of the pixel point coordinate P' of the object in the color data:
- a preset floating point number (for example, 0.5) can also be added to the result of the above division, which will not be repeated here.
- a plane In three-dimensional space, a plane can be determined by any three points that are not on the same line, so that a vector perpendicular to the plane can be obtained, so the normal vector of each pixel can be determined by two adjacent pixels. plane, and then solve for the plane perpendicular to the plane.
- a plurality of adjacent pixels for example, eight adjacent pixels
- Each pixel determines a plane in the three-dimensional space, and solves a vector perpendicular to the plane, and finally calculates the average of multiple vectors as the normal vector of each pixel.
- the pixel point x t as an example, according to its depth value d_t, its three-dimensional homogeneous coordinates can be obtained, and then the inverse K -1 of the internal parameter K is multiplied by the three-dimensional homogeneous coordinates, and the pixel point x t can be back projected into the three-dimensional space.
- the three-dimensional point P x of the pixel point x t is arranged in a counterclockwise order in the 8 neighborhood pixels of the pixel point x t in the 3*3 window, and back-projected to the three-dimensional space respectively to obtain the corresponding three-dimensional point, denoted as ⁇ P 0 , P 1 , P 2 , P 3 , 3, P 7 ⁇ , then the three-dimensional normal vector N x of the pixel point x t can be expressed as
- ⁇ represents the cross product
- % represents the remainder.
- 1% 8 represents the remainder of 1 divided by 8, which is 1, and other situations can be deduced by analogy, and no examples will be given here.
- the angle between the normal vector and the direction of gravity can be calculated by using the cosine formula, which is not repeated here.
- Step S122 Projecting each pixel in the three-dimensional space to the direction of gravity to obtain the height value of each pixel in the three-dimensional space.
- the step of obtaining the angle between the normal vector of each pixel point and the gravitational direction of the image to be processed in the above-mentioned step S121, and the step of obtaining the height value of each pixel point in the three-dimensional space in the step S122, can be according to the order. Sequential execution may also be performed simultaneously, which is not limited here.
- Step S123 Analyze the height values of the pixel points whose included angles satisfy the preset angle condition to obtain the plane height of the object to be reconstructed.
- the preset angle condition may include that the angle between the normal vector of the pixel point and the gravity direction of the image to be processed is less than or equal to a preset angle threshold (for example, 15 degrees, 10 degrees, etc.)
- a preset angle threshold for example, 15 degrees, 10 degrees, etc.
- screening is performed according to the preset angle condition to obtain the pixel point that meets the condition, and then the height value of each pixel point in the three-dimensional space obtained from the aforementioned step S122 , query the height values of the pixel points that satisfy the above-mentioned preset angle conditions, wherein the height values of the pixel points that satisfy the above-mentioned preset angle conditions can be regarded as a height set, and then perform cluster analysis on the height values in the height set , to obtain the plane height of the object to be reconstructed, so that only the height value can be used to obtain the plane height of the object to be reconstructed, which can reduce the calculation load.
- a preset angle threshold for example, 15 degrees, 10 degrees
- a random sampling consensus algorithm (Random Sample Consensus, RANSAC) can be used to cluster the height set, and each time a height value, the current plane height, can be randomly selected, and statistics related to the plane
- RANSAC Random Sample Consensus
- a preset drop range for example, 2 cm
- the minimum value is selected and the corresponding number of inliers is greater than a predetermined threshold. Set the threshold candidate height as the final plane height.
- Step S124 Use the plane height to screen out the target pixels belonging to the object to be reconstructed in the color data.
- the pixels whose height value is greater than the plane height can be screened, and then the pixels corresponding to the screened pixels can be queried in the color data as candidate pixels, and the maximum connected domain formed by the candidate pixels in the color data can be determined.
- the candidate pixels in the maximum connected domain are regarded as the target pixels belonging to the target to be reconstructed.
- the target pixels belonging to the target to be reconstructed in each frame of the to-be-processed image can be automatically identified in combination with the direction of gravity, reducing the computational load of 3D reconstruction and avoiding user intervention, thus improving user experience.
- FIG. 4 is a schematic flowchart of an embodiment of step S13 in FIG. 1 .
- 4 is a schematic flowchart of an embodiment of dividing image data of each frame of images to be processed into corresponding data sets. Can include the following steps:
- Step S131 successively take each frame of the image to be processed as the current image to be processed.
- the image data of a certain frame of the image to be processed when divided, it can be used as the current image to be processed.
- Step S132 When dividing the image data of the current image to be processed, it is judged that the last data set in the existing data set meets the preset overflow condition, if yes, go to step S133, otherwise go to step S134.
- the existing data sets include: data set A, data set B, and data set C. Among them, data set C is created the latest, and data set C can be used as the last data set.
- the preset overflow condition may include any of the following:
- the number of frames of the image to be processed corresponding to the image data contained in the data set is greater than or equal to the preset frame number threshold (for example, 8 frames, 9 frames, 10 frames, etc.); any image data in the end data set belongs to the to-be-processed
- the distance between the camera position of the image and the camera position of the current image to be processed is greater than a preset distance threshold (for example, 20 cm, 25 cm, 30 cm, etc.); the camera of the to-be-processed image to which any image data in the end data set belongs
- the difference between the facing angle and the camera facing angle of the current image to be processed is greater than a preset angle threshold (eg, 25 degrees, 30 degrees, 35 degrees, etc.).
- the camera orientation angle and camera position can be calculated according to the camera pose parameters of the image to be processed.
- the camera pose parameter T t can be determined by the matrix Representation, that is, the camera pose parameters include a rotation matrix R and a translation matrix t, and the camera position can be expressed as:
- T represents the transpose of the matrix.
- the third row vector of R can be represented as the camera facing angle direction.
- Step S133 Obtain the image data of the latest multi-frame images to be processed in the final data set, and store it in a newly created data set as a new final data set, and divide the image data of the current to-be-processed image into the new final data gather.
- the end data set C includes image data 05 (belonging to the image to be processed 05), image data 06 (belonging to the image to be processed 06), image data 07 (belonging to the image to be processed 07), image data 08 (belonging to the image to be processed 08), and image data 09 (belonging to the image to be processed 09), the image data of the image to be processed 07 to the image to be processed 09 can be obtained, or Obtain the image data of the to-be-processed image 08 to the to-be-processed image 09, which is not limited here, and store the acquired image data in a newly created data set.
- the image data is stored in the data set D.
- the data set D includes: image data 07 (belonging to the to-be-processed image 07 ), image data 08 (belonging to the to-be-processed image 08 ), and image data 09 (belonging to the to-be-processed image 08 ) image 09), and take the data set D as a new final data set, and divide the image data 10 (belonging to the image 10 to be processed) into the data set D.
- the end data set may also not meet the preset overflow condition, and the following step S134 may be performed in this case.
- Step S134 Divide the image data of the current image to be processed into an end data set.
- the image data of the current image to be processed is divided, if the last data set in the existing data set satisfies the preset overflow condition, the latest multi-frame to-be-processed image in the end data set is obtained.
- the image data is stored in a newly created data set as a new end data set, so there are multiple frames of the same image data of the image to be processed between adjacent data sets, which is conducive to improving the alignment between adjacent data sets effect, which is beneficial to improve the effect of 3D reconstruction.
- FIG. 5 is a schematic flowchart of an embodiment of step S14 in FIG. 1 .
- 5 is a schematic flowchart of an embodiment of determining the pose optimization parameters of the data set, which may include the following steps:
- Step S141 Take each data set as the current data set in turn, and select at least one data set whose time sequence is located before the current data set as a candidate data set.
- data set B when determining the pose optimization parameters of data set B, data set B can be used as the current data set, and when determining the pose of data set C
- the data set C when optimizing parameters, can be used as the current data set.
- a new data set that is, the pose optimization parameters of the previous data set of the newly created data set can be determined.
- a new data set D is newly created. At this time, the data set C can be used as the current data set, and its pose optimization parameters are determined.
- FIG. 6 is a schematic flowchart of an embodiment of step S141 in FIG. 5 , which may include the following steps:
- Step S61 constructing a bag-of-words model by using the preset image features of the image data in the current data set and the data set whose time sequence is located before it.
- the preset image features can include ORB (Oriented FAST and Rotated Brief) image features, which can quickly create feature vectors for key points in the image data, and the feature vectors can be used to identify the target to be reconstructed in the image data.
- ORB Oriented FAST and Rotated Brief
- Fast and Brief They are the feature detection algorithm and the vector creation algorithm respectively, and details are not repeated here.
- the bag of words model is a simplified expression model under natural language processing and information retrieval. Each preset image feature in the bag of words model is independent, and details are not repeated here.
- the previous data set can be used as the current data set, and the preset image features of the image data in the current data set can be extracted and added to the bag-of-words model. In this way, the bag-of-words model can be incrementally expanded.
- there is duplicate image data between the current data set and its previous data set so when extracting the preset image features of the image data in the current data set, the duplicated image data with the previous data set is not identical. Then perform feature extraction.
- Step S62 Select the image data of the to-be-processed image at a preset time sequence in the current data set as the image data to be matched.
- the preset time sequence may include the first position, the middle position, and the last position.
- the data set C includes image data 05 (belonging to the image to be processed 05 ), image data 06 (belonging to image 06 to be processed), image data 07 (belonging to image 07 to be processed), image data 08 (belonging to image 08 to be processed), and image data 09 (belonging to image 09 to be processed), you can select the first image to be processed 05
- the image data 05, the image data 07 of the middle image 07 to be processed, and the image data 09 of the last image 09 to be processed are used as the image data to be matched, and other implementation scenarios can be deduced by analogy, and will not be exemplified here.
- the preset timing can also be set as the first position, the 1/4 timing position, the 1/2 timing position, the 3/4 timing position, and the last position, which is not limited here.
- Step S63 From the preset range of the bag-of-words model, query the preset image features whose similarity score between the preset image features of the image data to be matched is greater than a preset similarity threshold.
- the preset range may include preset image features of the image data whose data set is not adjacent to the current data set and is not included in the current data set. Still taking the data set A, data set B and data set C in the foregoing embodiment as an example, when the current data set is the data set C, the preset range may be the preset image features belonging to the data set A and the data set B. .
- the preset similarity threshold may be a preset score value, for example, 0.018, 0.019, 0.020, etc., which is not limited herein.
- the maximum score value score adj in the similarity scores between each image data in the data set adjacent to the current data set and the image data to be matched may also be obtained, and the pre-calculation of the maximum score value score adj A multiple (eg, 1.5 times, 2 times, 2.5 times) is set as the preset similarity threshold.
- the preset multiple of the maximum score value score adj and any of the above preset score values can be used as the preset similarity threshold, that is, the query can be made from the preset range of the bag-of-words model.
- the similarity score score loop between the preset image features of the image data to be matched is greater than the preset multiple of the maximum score value score adj , and the preset image feature of any one of the above preset score values, which is not limited here. .
- Step S64 The data set where the image data to which the queried preset image feature belongs, and the data set adjacent to the current data set are used as the candidate data set.
- the data set C and the data set D are queried by using the image data to be matched in the first position, and the data set D and the data set E are queried by using the image data to be matched in the middle position, Using the to-be-matched image data at the last position, the data set E and the data set F are queried, and the data sets C to F and the data set G can be used as the candidate data sets of the current data set H.
- a preset number for example, 2, 3, etc.
- the data sets adjacent to the current data set are used as candidate data sets.
- the three data sets with the largest similarity score score loops and the data set G adjacent to the current data set can be selected from the data sets C to F as the candidate data sets.
- Step S142 Using the image data of the current data set and the image data of the candidate data set, determine the spatial transformation parameters between the current data set and the candidate data set.
- FIG. 7 is a schematic flowchart of an embodiment of step S142 in FIG. 5, which may include follow the steps below:
- Step S71 Search for a set of image data to be matched that satisfies a preset matching condition in the candidate data set and the current data set.
- the preset matching condition may include that the difference between the camera orientation angles of the to-be-processed images to which the image data to be matched belongs is the smallest, wherein, for each candidate data set, a group that satisfies the preset matching can be searched from the current data set and the candidate data set.
- the image data to be matched belonging to the current data set may be denoted as I cur
- the image data to be matched belonging to the candidate data set may be denoted as I similar .
- Step S72 Based on the preset image features extracted from each set of image data to be matched, obtain matching pixel pairs between each set of image data to be matched.
- the preset image features for example, ORB image features
- the preset image features can be matched and screened to obtain the matching pixels between I cur and I similar , for ease of description, can be respectively recorded as p cur and p similar .
- RANSAC algorithm reference may be made to the relevant steps in the foregoing embodiments, which will not be repeated here.
- Step S73 Map the pixels belonging to the current data set in the matching pixel pair to the three-dimensional space to obtain the first three-dimensional matching point, and map the pixels belonging to the candidate data set in the matching pixel pair to the three-dimensional space to obtain the second three-dimensional matching point. 3D matching points.
- p cur and p similar can be converted into three-dimensional homogeneous coordinates respectively, and then the three-dimensional homogeneous coordinates of p cur and p similar can be left-multiplied by the inverse K -1 of the internal parameter K to obtain the first three-dimensional matching points P cur and The second three-dimensional matching point P similar .
- Step S74 Align the first three-dimensional matching point and the second three-dimensional matching point to obtain a spatial transformation parameter.
- first three-dimensional matching point and the second three-dimensional matching point may be aligned in three-dimensional space, so that the degree of coincidence between the two is as large as possible, so as to obtain the space transformation parameter between the two.
- a first pose transformation parameter between the first three-dimensional matching point and the second three-dimensional matching point may be obtained, wherein the first pose may be constructed by using the first three-dimensional matching point and the second three-dimensional matching point. Transform the objective function of the parameters, and then use SVD (Singular Value Decomposition, singular value decomposition) or non-offline optimization to solve the objective function, and obtain the first pose transformation parameter T pcd :
- the first three-dimensional matching point can also be positioned by using the first pose transformation parameter T pcd and a preset pose transformation parameter (eg, identity matrix). pose optimization to obtain the first optimized matching point and the second optimized matching point respectively, wherein the first pose transformation parameter T pcd and the preset pose transformation parameter can be used to multiply the first three-dimensional matching point P cur to the left respectively, so as to obtain respectively
- the first optimal matching point and the second optimal matching point can be respectively recorded as and Then calculate the second three-dimensional matching point P similar and the first optimized matching point respectively Second optimal matching point and select the pose transformation parameter adopted by the optimized matching point with a higher coincidence degree as the second pose transformation parameter, which can be denoted as T select for convenience of description.
- the first optimal matching point when calculating the second three-dimensional matching point P similar and the first optimal matching point, the first optimal matching point can be searched within a preset range (for example, a range of 5 cm) of each second three-dimensional matching point P similar If it can be found, the second three-dimensional matching point P similar is marked as valid, otherwise, it can be marked as invalid. After all the second three-dimensional matching points P similar are searched, the second three-dimensional matching point P similar marked as valid is calculated.
- a preset range for example, a range of 5 cm
- the ratio of the number to the total number of the second three-dimensional matching points P similar that is, the second three-dimensional matching point P similar and the first optimized matching point
- the degree of coincidence between the second three-dimensional matching point P similar and the second optimal matching point The degree of overlap between them can be deduced by analogy, which is not repeated here.
- the second pose transformation parameter T select may be used as an initial value, and a preset alignment method (for example, the point-to-normal ICP method) can be used to match the first three-dimensional matching point P cur and the second three-dimensional matching point P similar are aligned to obtain a spatial transformation parameter between the current data set and the candidate data set, which is denoted as T icp for convenience of description.
- a preset alignment method for example, the point-to-normal ICP method
- T icp the spatial transformation parameter between the current data set and each candidate data set can be obtained.
- Step S143 At least use the pose optimization parameters of the candidate data set and the spatial transformation parameters between the current data set and the candidate data set to obtain the pose optimization parameters of the current data set, and at least update the pose optimization parameters of the candidate data set .
- the above-mentioned spatial transformation parameters T icp may be screened, wherein the current data set and each candidate can be obtained from the Among the spatial transformation parameters T icp between the data sets, the spatial transformation parameters that meet the preset screening conditions are selected for use in solving the pose optimization parameters of the current data set.
- the preset screening conditions may include: a candidate data set related to the spatial transformation parameter T icp is adjacent to the current data set, or an optimized matching point obtained by performing pose optimization on the first three-dimensional matching point P cur using the spatial transformation parameter T icp ,
- the degree of coincidence with the second three-dimensional matching point P similar is greater than a predetermined threshold of coincidence degree (eg, 60%, 65%, 70%, etc.).
- a predetermined threshold of coincidence degree eg, 60%, 65%, 70%, etc.
- the pose optimization parameters of the candidate data set and the spatial transformation between the current data set and the candidate data set can be used to construct the objective function of the pose optimization parameters of the current data set, and the objective function of the current data set can be obtained by solving the objective function.
- the pose optimization parameters are updated, and at least the pose optimization parameters of the candidate data set are updated.
- the previous data set of the newly created data set is respectively used as the current data set, and the pose optimization parameters can be obtained while scanning the target to be reconstructed and creating the data set, which can help to balance the calculation amount and reduce the Calculate the load, and realize 3D reconstruction of the target to be reconstructed in real time and online.
- FIG. 8 is a schematic flowchart of an embodiment of step S143 in FIG. 5 . Can include the following steps:
- Step S81 Take the two data sets corresponding to the respective spatial transformation parameters related to the current data set and the candidate data set whose time sequence is located before it as a data set pair.
- the data sets C to F and the data set G are the candidate data sets of the current data set H
- the spatial transformation parameters are The corresponding candidate data set C and the current data set H are regarded as a pair of data set pairs
- the spatial transformation parameters are
- the corresponding candidate data set D and the current data set H are regarded as a pair of data set pairs
- the spatial transformation parameters are
- the corresponding candidate data set E and the current data set H are regarded as a pair of data set pairs
- the spatial transformation parameters are
- the corresponding candidate data set F and the current data set H are regarded as a pair of data set pairs
- the spatial transformation parameters are The corresponding candidate data set G and the current data set H are regarded as a pair of data sets.
- each data set before the current data set H also has corresponding spatial transformation parameters.
- data set B there may be spatial transformation parameters between data set A and data set A
- the data set B and the data set A can be regarded as a data set pair.
- data set C there can be spatial transformation parameters between the data set A and the data set B respectively, so the data set C and the data set can be respectively used.
- A is regarded as a data set pair
- data set C and data set B are regarded as a data set pair, and so on, and will not be exemplified here.
- Step S82 Using the spatial transformation parameters of each data set pair and the respective pose optimization parameters to construct an objective function related to the pose optimization parameters.
- the objective function can be expressed as:
- i and j respectively represent the numbers of the data sets included in each data set pair (for example, letters such as C, D, E, etc., or can also be represented by Arabic numerals such as 1, 2, and 3) , represents the spatial transformation parameters between each data set pair, respectively represent the pose optimization parameters of each data set for the data set contained in it, and f( ) represents the optimization formula, which can be expressed as:
- Step S83 Solve the objective function by using a preset solving method, and obtain the pose optimization parameters of the data set included in the data set corresponding to the current data set and the candidate data set whose time sequence is located before it.
- the pose optimization parameters of the data sets included in each data set pair can be obtained.
- the pose optimization parameters of the current data set H can be obtained, and the further optimized pose optimization parameters of the data sets C to G can be obtained, as well as the current data set.
- Pose optimization parameters after further optimization of the data set before H When a new data set I is introduced and the spatial transformation parameters related to it are obtained, by constructing the objective function, the pose optimization parameters of the data set I can be obtained, and the pose optimization after further optimization of the previous data set can be obtained. Parameters, such a cycle, can further help eliminate the cumulative error of the pose.
- the image data of the current data set and the image data of the candidate book set are utilized. , determine the spatial transformation parameters between the current data set and the candidate data set, and then use at least the pose optimization parameters of the candidate data set and the spatial transformation parameters between the current data set and the candidate data set to obtain the pose of the current data set.
- Optimizing the parameters, and at least updating the pose optimization parameters of the candidate data set can help eliminate the error of the camera pose parameters accumulated during the scanning process, and reduce the amount of data used to calculate the pose optimization parameters, thereby reducing the computational load. .
- FIG. 9 is a schematic flowchart of an embodiment of an interaction method based on 3D reconstruction of the present disclosure. , which can include the following steps:
- Step S91 Obtain a three-dimensional model of the target to be reconstructed.
- the three-dimensional model may be obtained through the steps in any of the foregoing three-dimensional reconstruction method embodiments, and reference may be made to the aforementioned three-dimensional reconstruction method embodiments, which will not be repeated here.
- Step S92 constructing a three-dimensional map of the scene where the camera device is located by using a preset visual inertial navigation method, and acquiring current pose information of the camera device in the three-dimensional map.
- the preset visual inertial navigation method can include SLAM (Simultaneous Localization and Mapping, real-time positioning and map construction).
- SLAM Simultaneous Localization and Mapping, real-time positioning and map construction.
- SLAM Simultaneous Localization and Mapping, real-time positioning and map construction.
- the 3D model in order to realize the dynamic interaction with the 3D model, can also be bound with bones.
- Bone binding refers to setting up a skeleton system for the 3D model, so that it can move according to the established rules at the skeleton joints, such as , the three-dimensional model is a four-legged animal such as a cow, a sheep, etc., after the three-dimensional model is bound with bones, its bone joints can move according to the established rules of the four-legged animal.
- Step S93 Based on the pose information, display the three-dimensional model in the scene image currently captured by the imaging device.
- the pose information may include the position and orientation of the camera device. For example, when the pose information of the camera device indicates that it is facing the ground, the scene image currently captured by the camera device can display the top of the 3D model; or, when the pose information of the camera device indicates that the camera device is facing an acute angle with the ground , the side of the 3D model can be displayed in the scene image currently captured by the camera device.
- the skeleton after the skeleton is bound to the 3D model, it can also accept the driving instructions input by the user, so that the 3D model can move according to the driving instructions input by the user. For example, if the 3D model is a sheep, the user can drive it to lower its head and walk. Wait, there is no limitation here.
- the three-dimensional model is a person or other objects, it can be deduced in the same way, and will not be exemplified one by one here.
- the above solution based on the pose information of the camera device in the three-dimensional map of the scene, displays the three-dimensional model of the target to be reconstructed in the currently captured scene image, which can realize the geometric consistency fusion of the virtual object and the real scene, and because The three-dimensional model is obtained by the three-dimensional reconstruction method in the first aspect, so the effect of three-dimensional reconstruction can be improved, and the effect of geometrically consistent fusion of virtual and reality can be improved, which is beneficial to improve user experience.
- FIG. 10 is a schematic flowchart of an embodiment of a three-dimensional reconstruction-based measurement method of the present disclosure. , which can include the following steps:
- Step S1010 Obtain a three-dimensional model of the target to be reconstructed.
- the three-dimensional model may be obtained through the steps in any of the foregoing three-dimensional reconstruction method embodiments, and reference may be made to the aforementioned three-dimensional reconstruction method embodiments, which will not be repeated here.
- Step S1020 Receive a plurality of measurement points set by the user on the three-dimensional model.
- the number of measurement points can be two, three, four, etc., which is not limited here.
- the user taking the object to be reconstructed as an example of a plaster portrait, the user can set measurement points respectively in the centers of the two eyes of the three-dimensional model 29 , or can also set measurement points in the root and the person of the three-dimensional model 29 respectively, or , and the measurement points can also be set in the center of the two eyes of the three-dimensional model 29 and in the person, which will not be listed one by one here.
- Step S1030 Acquire the distances between the multiple measurement points, and obtain the distances between the positions on the target to be reconstructed corresponding to the multiple measurement points.
- the distance between the centers of the two eyes of the three-dimensional model 29 can be obtained, or, by obtaining the three-dimensional model 29
- the distance between the mountain root and the human middle can be obtained by obtaining the plaster portrait corresponding to the distance between the mountain root and the human middle.
- the above solution obtains the distance between the multiple measurement points by receiving the multiple measurement points set by the user on the 3D model, and then obtains the distance between the positions corresponding to the multiple measurement points on the target to be reconstructed, so as to satisfy the The measurement requirements for objects in the real scene, and because the 3D model is obtained by using the 3D reconstruction method in the first aspect, the effect of the 3D reconstruction can be improved, thereby improving the measurement accuracy.
- FIG. 11 is a schematic frame diagram of an embodiment of a three-dimensional reconstruction apparatus 1100 of the present disclosure.
- the three-dimensional reconstruction device 1100 includes an image acquisition part 1110, a first determination part 1120, a data division part 1130, a second determination part 1140, a parameter adjustment part 1150, and a model reconstruction part 1160, and the image acquisition part 1110 is configured to acquire the object to be reconstructed by scanning the imaging device The obtained multi-frame images to be processed;
- the first determination part 1120 is configured to use each frame of the to-be-processed image and the calibration parameters of the imaging device to determine the target pixels of each frame of the to-be-processed image belonging to the target to be reconstructed and its camera pose parameters;
- data The dividing part 1130 is configured to divide the image data of each frame of images to be processed into corresponding data sets in turn according to a preset dividing strategy, wherein the image data at least includes target pixels; the second determining part 1140 sequentially utilizes the images of each data set data
- the second determining part 1140 includes a data set selection sub-part, configured to sequentially regard each data set as the current data set, and select at least one data set located before the current data set as a candidate data set
- the first The second determination part 1140 further includes a spatial transformation parameter sub-part, configured to use the image data of the current data set and the image data of the candidate data set to determine the spatial transformation parameters between the current data set and the candidate data set
- the second determination part 1140 also It includes a pose optimization parameter subsection, configured to use at least the pose optimization parameters of the candidate data set and the spatial transformation parameters between the current data set and the candidate data set to obtain the pose optimization parameters of the current data set, and at least update all the The pose optimization parameters of the candidate dataset are described.
- the pose optimization parameter subsection includes a data set pair section, configured to treat two data sets corresponding to respective spatial transformation parameters related to the current data set and the data set temporally before it, as a data set
- the pose optimization parameter subsection also includes an objective function construction part, which is configured to use the spatial transformation parameters of each data set pair and the respective pose optimization parameters to construct an objective function about the pose optimization parameters
- the pose optimization The parameter sub-section also includes an objective function solving part, which is configured to solve the objective function by using a preset solving method, and obtain the current data set and the data set corresponding to the data set whose time sequence is located before it. The pose optimization of the included data set parameter.
- the spatial transformation parameter subsection includes an image data search section configured to search for a set of image data to be matched that satisfies a preset matching condition in the candidate data set and the current data set, and the spatial transformation parameter subsection further includes a matching
- the pixel point selection part is configured to obtain matching pixel point pairs between each group of image data to be matched based on preset image features extracted from each group of image data to be matched
- the spatial transformation parameter sub-part further includes a three-dimensional space mapping part is configured to map the pixels belonging to the current data set in the matching pixel pair to the three-dimensional space to obtain the first three-dimensional matching point, and map the pixels belonging to the candidate data set in the matching pixel pair to the three-dimensional space to obtain the second three-dimensional matching point.
- the three-dimensional matching point and the spatial transformation parameter subsection further includes a three-dimensional matching point alignment part, which is configured to perform alignment processing on the first three-dimensional matching point and the second three-dimensional matching point to obtain the spatial transformation parameters.
- the 3D matching point alignment section includes a first pose transformation parameter subsection configured to obtain a first pose transformation parameter between the first 3D matching point and the second 3D matching point, and the 3D matching point aligning section It also includes a three-dimensional matching point optimization sub-section, configured to use the first pose transformation parameter and the preset pose transformation parameter to perform pose optimization on the first three-dimensional matching point, and obtain the first optimized matching point and the second optimized matching point respectively.
- the three-dimensional matching point alignment part also includes a second pose transformation parameter sub-section, configured to calculate the degree of coincidence between the second three-dimensional matching point and the first optimal matching point and the second optimal matching point, and select a higher degree of coincidence
- the pose transformation parameters adopted by the optimized matching points of the The first three-dimensional matching point and the second three-dimensional matching point are aligned to obtain the spatial transformation parameters between the current data set and the candidate data set.
- the spatial transformation parameter subsection further includes a transformation parameter screening section, configured to select spatial transformation parameters that meet preset parameter screening conditions from the spatial transformation parameters between the current data set and each candidate data set; wherein , the preset parameter screening conditions include any one of the following: the candidate data set related to the spatial transformation parameter is adjacent to the current data set; the optimized matching point obtained by performing pose optimization on the first three-dimensional matching point by using the spatial transformation parameter The coincidence degree between the three-dimensional matching points is greater than a predetermined coincidence degree threshold.
- the data set selection subsection includes a bag-of-words model construction section configured to construct a bag-of-words model using preset image features of the image data in the current data set and the data set temporally located before it, and the data set selection The subsection also includes an image data part to be matched, and is configured to select image data whose image to be processed belongs to a preset time sequence in the current data set, as the image data to be matched, and the data set selection subsection also includes an image feature query part, It is configured to query the preset image features whose similarity score between the preset image features of the image data to be matched is greater than a preset similarity threshold from the preset range of the bag-of-words model, and the data set selection subsection also includes candidate images.
- the data set part is configured to use the data set where the image data to which the queried preset image feature belongs and the data set adjacent to the current data set are located as candidate data sets, wherein the preset range includes the data set and the data set to which they belong.
- the current data set is not adjacent and is not included in the preset image features of the image data in the current data set.
- the data set selection subsection further includes a maximum similarity score value acquisition section, configured to acquire the similarity score between each image data in the data set adjacent to the current data set and the image data to be matched.
- the maximum score value, the data set selection subsection also includes a preset similarity threshold value determination part, configured to use either a preset multiple of the maximum score value or a preset score value as the preset similarity threshold value.
- the data dividing part 1130 includes a current image to be processed sub-part configured to sequentially regard each frame of the image to be processed as the current image to be processed, and the data dividing part 1130 further includes a data processing sub-part configured to When the image data of the image to be processed is divided, if the last data set in the existing data set satisfies the preset overflow condition, the image data of the latest multi-frame to-be-processed images in the last data set is obtained, and stored in a newly created As a new end data set, the image data of the current image to be processed is divided into a new end data set.
- the preset overflow condition includes any one of the following: the frame number of the image to be processed corresponding to the image data included in the end data set is greater than or equal to a preset frame number threshold; any image data in the end data set The distance between the camera position of the to-be-processed image to which it belongs and the camera position of the current to-be-processed image is greater than the preset distance threshold; the camera orientation angle of the to-be-processed image to which any image data in the end data set belongs and the camera of the current to-be-processed image The difference between the orientation angles is greater than a preset angle threshold; wherein, the camera position and the camera orientation angle are calculated by using the camera pose parameters of the image to be processed.
- each frame of the image to be processed includes color data and depth data
- the first determining part 1120 includes an included angle obtaining sub-part, configured to obtain the depth data after alignment with the color data for each pixel included in the pixel.
- the angle between the normal vector and the gravitational direction of the image to be processed, the first determining part 1120 also includes a height acquisition sub-part, configured to project each pixel in the three-dimensional space to the gravitational direction, and obtain each pixel in the three-dimensional space.
- the first determination part 1120 also includes a height analysis sub-part, configured to analyze the height value of the pixel points whose included angle satisfies the preset angle condition, to obtain the plane height of the object to be reconstructed, and the first determination part 1120 also includes The pixel screening subsection is configured to use the plane height to screen the target pixel points belonging to the object to be reconstructed in the color data.
- the height analysis subsection includes a height set acquisition section, configured to use the height values of pixels whose included angles satisfy a preset angle condition as a height set, and the height analysis subsection includes a height cluster analysis section, configured to In order to perform cluster analysis on the height values in the height set, the plane height of the object to be reconstructed is obtained.
- the three-dimensional reconstruction apparatus 1100 further includes a three-dimensional mapping part, configured to sequentially map the image data in each data set to a three-dimensional space to obtain a three-dimensional point cloud corresponding to each data set, and the three-dimensional reconstruction apparatus 1100 further It includes a point cloud adjustment part, which is configured to use the pose optimization parameters of each data set to adjust the corresponding 3D point cloud.
- FIG. 12 is a schematic diagram of a framework of an embodiment of a three-dimensional reconstruction-based interaction apparatus 1200 of the present disclosure.
- the interactive device 1200 based on three-dimensional reconstruction includes a model acquisition part 1210, a mapping positioning part 1220 and a display interactive part 1230.
- the model acquisition part 1210 is configured to acquire a three-dimensional model of the object to be reconstructed, wherein the three-dimensional model is obtained by using any of the above three-dimensional reconstruction devices Obtained by the three-dimensional reconstruction device in the embodiment;
- the mapping and positioning part 1220 is configured to use a preset visual inertial navigation method to construct a three-dimensional map of the scene where the camera device is located, and obtain the current pose information of the camera device in the three-dimensional map;
- display interactive Section 1230 is configured to display the three-dimensional model in the scene image currently captured by the camera device based on the pose information.
- FIG. 13 is a schematic frame diagram of an embodiment of a three-dimensional reconstruction-based measurement device 1300 of the present disclosure.
- the measurement device 1300 based on 3D reconstruction includes a model acquisition part 1310, a display interaction part 1320 and a distance acquisition part 1330.
- the model acquisition part 1310 is configured to acquire a 3D model of the object to be reconstructed, wherein the 3D model is implemented by using any of the above 3D reconstruction devices
- the display interaction part 1320 is configured to receive a plurality of measurement points set by the user on the three-dimensional model; the distance acquisition part 1330 is configured to obtain the distance between the plurality of measurement points, and obtain the corresponding objects on the object to be reconstructed. The distance between the positions of multiple measurement points.
- FIG. 14 is a schematic diagram of a framework of an embodiment of an electronic device 1400 of the present disclosure.
- the electronic device 1400 includes a memory 1410 and a processor 1420 that are coupled to each other, and the processor 1420 is configured to execute program instructions stored in the memory 1410 to implement the steps in any of the foregoing three-dimensional reconstruction method embodiments, or to implement any of the foregoing three-dimensional reconstruction method embodiments.
- the electronic device may include a mobile terminal such as a mobile phone and a tablet computer, or the electronic device may also be a data processing device (such as a microcomputer) connected with a camera device, which is not limited herein.
- the processor 1420 may also be referred to as a CPU (Central Processing Unit, central processing unit).
- the processor 1420 may be an integrated circuit chip with signal processing capability.
- the processor 1420 may also be a general-purpose processor, a digital signal processor (Digital Signal Processor, DSP), an application-specific integrated circuit (Application Specific Integrated Circuit, ASIC), a field-programmable gate array (Field-Programmable Gate Array, FPGA) or other Programmable logic devices, discrete gate or transistor logic devices, discrete hardware components.
- a general purpose processor may be a microprocessor or the processor may be any conventional processor or the like.
- the processor 1420 may be jointly implemented by an integrated circuit chip.
- the above solution can improve the effect of 3D reconstruction and reduce the computational load of 3D reconstruction.
- FIG. 15 is a schematic diagram of a framework of an embodiment of the disclosed computer-readable storage medium 1500 .
- the computer-readable storage medium 1500 stores program instructions 1501 that can be executed by the processor, and the program instructions 1501 are used to implement the steps in any of the foregoing three-dimensional reconstruction method embodiments, or to implement any of the foregoing three-dimensional reconstruction-based interactive method embodiments. steps, or implement the steps in any of the foregoing three-dimensional reconstruction-based measurement method embodiments.
- the above solution can improve the effect of 3D reconstruction and reduce the computational load of 3D reconstruction.
- the disclosed method and apparatus may be implemented in other manners.
- the device implementations described above are only illustrative.
- the division of modules or units is only a logical function division. In actual implementation, there may be other divisions.
- units or components may be combined or integrated. to another system, or some features can be ignored, or not implemented.
- the shown or discussed mutual coupling or direct coupling or communication connection may be through some interfaces, indirect coupling or communication connection of devices or units, which may be in electrical, mechanical or other forms.
- Units described as separate components may or may not be physically separated, and components shown as units may or may not be physical units, that is, may be located in one place, or may be distributed over network units. Some or all of the units may be selected according to actual needs to achieve the purpose of the solution in this implementation manner.
- each functional unit in each embodiment of the present disclosure may be integrated into one processing unit, or each unit may exist physically alone, or two or more units may be integrated into one unit.
- the above-mentioned integrated units may be implemented in the form of hardware, or may be implemented in the form of software functional units.
- the integrated unit if implemented as a software functional unit and sold or used as a stand-alone product, may be stored in a computer-readable storage medium.
- the technical solutions of the present disclosure can be embodied in the form of software products in essence, or the part that contributes to the prior art, or all or part of the technical solutions, and the computer software product is stored in a storage medium , including several instructions for causing a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor (processor) to execute all or part of the steps of the methods of the various embodiments of the present disclosure.
- the aforementioned storage medium includes: U disk, mobile hard disk, Read-Only Memory (ROM, Read-Only Memory), Random Access Memory (RAM, Random Access Memory), magnetic disk or optical disk and other media that can store program codes .
- multiple frames of images to be processed obtained by scanning the target to be reconstructed by the camera device are obtained; each frame of the to-be-processed image and the calibration parameters of the camera device are used to determine the target pixels of each frame of the to-be-processed image belonging to the target to be reconstructed and Its camera pose parameters; sequentially divide the image data of each frame of the image to be processed into the corresponding data set; use the image data of the data set and the image data and pose optimization parameters of the data set before it in time sequence to determine the data set.
- Pose optimization parameters use the pose optimization parameters of the data set to adjust the camera pose parameters of the to-be-processed image to which the image data contained in the data set belongs; reconstruct the image data of the to-be-processed image to obtain the object to be reconstructed 3D model.
- the above solution can improve the effect of 3D reconstruction and reduce the computational load of 3D reconstruction.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Computer Graphics (AREA)
- Geometry (AREA)
- Software Systems (AREA)
- Image Analysis (AREA)
- Length Measuring Devices By Optical Means (AREA)
- Image Processing (AREA)
Abstract
Description
Claims (34)
- 一种三维重建方法,包括:获取摄像器件扫描待重建目标得到的多帧待处理图像;利用每帧所述待处理图像和所述摄像器件的标定参数,确定每帧所述待处理图像属于所述待重建目标的目标像素点及其相机位姿参数;按照预设划分策略,依次将各帧所述待处理图像的图像数据划分至对应的数据集合,其中,所述图像数据至少包括所述目标像素点;依次利用各个所述数据集合的图像数据,及时序位于其之前的数据集合的图像数据和位姿优化参数,确定每一所述数据集合的位姿优化参数;利用每一所述数据集合的位姿优化参数,对包含于所述数据集合内的图像数据所属的所述待处理图像的相机位姿参数进行调整;利用预设三维重建方式和所述待处理图像的调整后的相机位姿参数,对所述待处理图像的图像数据进行重建处理,得到所述待重建目标的三维模型。
- 根据权利要求1所述的三维重建方法,其中,所述依次利用各个所述数据集合的图像数据,及时序位于其之前的数据集合的图像数据和位姿优化参数,确定每一所述数据集合的位姿优化参数包括:依次将每一所述数据集合作为当前数据集合,并选取至少一个时序位于所述当前数据集合之前的数据集合,作为候选数据集合;利用所述当前数据集合的图像数据和所述候选数据集合的图像数据,确定所述当前数据集合和所述候选数据集合之间的空间变换参数;至少利用所述候选数据集合的位姿优化参数,以及所述当前数据集合与所述候选数据集合之间的空间变换参数,获得所述当前数据集合的位姿优化参数,并至少更新所述候选数据集合的位姿优化参数。
- 根据权利要求2所述的三维重建方法,其中,所述至少利用所述候选数据集合的位姿优化参数,以及所述当前数据集合与所述候选数据集合之间的空间变换参数,获得所述当前数据集合的位姿优化参数,并至少更新所述候选数据集合的位姿优化参数包括:将分别与所述当前数据集合以及时序位于其之前的所述候选数据集合相关的各个空间变换参数所对应的两个数据集合,作为一数据集合对;利用各个所述数据集合对的空间变换参数,以及各自的位姿优化参数,构建一关于所述位姿优化参数的目标函数;利用预设求解方式对所述目标函数进行求解,得到所述当前数据集合及时序位于其之前的所述候选数据集合各自对应的数据集合对所包含的数据集合的位姿优化参数。
- 根据权利要求2所述的三维重建方法,其中,所述利用所述当前数据集合的图像数据和所述候选数据集合的图像数据,确定所述当前数据集合和所述候选数据集合之间的空间变换参数包括:在所述候选数据集合和所述当前数据集合中搜索一组满足预设匹配条件的待匹配图像数据;基于从每组所述待匹配图像数据中提取得到的预设图像特征,得到每组所述待匹配图像数据之间的匹配像素点对;将所述匹配像素点对中属于所述当前数据集合的像素点映射至三维空间,得到第一三维匹配点,并将所述匹配像素点对中属于所述候选数据集合的像素点映射至所述三维空间,得到第二三维匹配点;将所述第一三维匹配点和所述第二三维匹配点进行对齐处理,得到所述空间变换参数。
- 根据权利要求4所述的三维重建方法,其中,所述将所述第一三维匹配点和所述第二三维匹配点进行对齐处理,得到所述空间变换参数包括:获取所述第一三维匹配点和所述第二三维匹配点之间的第一位姿变换参数;利用所述第一位姿变换参数和预设位姿变换参数,对所述第一三维匹配点进行位姿优化,分别得到第一优化匹配点和第二优化匹配点;计算所述第二三维匹配点分别与所述第一优化匹配点、所述第二优化匹配点之间的重合度,并选取所述重合度较高的优化匹配点所采用的位姿变换参数,作为第二位姿变换参数;以所述第二位姿变换参数作为初始值,利用预设对齐方式将所述第一三维匹配点和所述第二三维匹配点进行对齐处理,得到所述当前数据集合与所述候选数据集合之间的空间变换参数。
- 根据权利要求4所述的三维重建方法,其中,所述利用所述当前数据集合的图像数据和所述候选数据集合的图像数据,确定所述当前数据集合和所述候选数据集合之间的空间变换参数之后,以及所述至少利用所述候选数据集合的位姿优化参数,以及所述当前数据集合与所述候选数据集合之间的空间 变换参数,获得所述当前数据集合的位姿优化参数之前,所述方法还包括:从所述当前数据集合与各个所述候选数据集合之间的空间变换参数中,选取符合预设参数筛选条件的空间变换参数;其中,所述预设参数筛选条件包括以下任一者:所述空间变换参数相关的所述候选数据集合与所述当前数据集合相邻;利用所述空间变换参数对所述第一三维匹配点进行位姿优化得到的优化匹配点,与所述第二三维匹配点之间的重合度大于一预设重合度阈值。
- 根据权利要求2所述的三维重建方法,其中,所述选取至少一个时序位于所述当前数据集合之前的数据集合,作为候选数据集合包括:利用所述当前数据集合及时序位于其之前的数据集合中的图像数据的预设图像特征,构建词袋模型;选取所属的待处理图像位于所述当前数据集合中的预设时序处的图像数据,作为待匹配图像数据;从所述词袋模型的预设范围中,查询与所述待匹配图像数据的预设图像特征之间的相似度评分大于一预设相似度阈值的预设图像特征;将查询到的预设图像特征所属的图像数据所在的数据集合,以及与所述当前数据集合相邻的数据集合,作为所述候选数据集合;其中,所述预设范围包括所属的数据集合与所述当前数据集合不相邻,且不包含于所述当前数据集合中的图像数据的预设图像特征。
- 根据权利要求7所述的三维重建方法,其中,所述从所述词袋模型的预设范围中,查询与所述待匹配图像数据的预设图像特征之间的相似度评分大于一预设相似度阈值的预设图像特征之前,所述方法还包括:获取与所述当前数据集合相邻的数据集合中各个所述图像数据与所述待匹配图像数据之间的相似度评分中的最大评分值;将所述最大评分值的预设倍数和一预设评分值中的任一者作为所述预设相似度阈值。
- 根据权利要求1所述的三维重建方法,其中,所述按照预设划分策略,依次将各帧所述待处理图像的图像数据划分至对应的数据集合包括:依次将各帧所述待处理图像作为当前待处理图像;在对当前待处理图像的图像数据进行划分时,若已有的所述数据集合中的末尾数据集合满足预设溢出条件,则获取所述末尾数据集合中最新的多帧所述待处理图像的图像数据,并存入一新创建的所述数据集合,作为新的所述末尾数据集合,将所述当前待处理图像的图像数据划分至新的所述末尾数据集合。
- 根据权利要求9所述的三维重建方法,其中,所述预设溢出条件包括以下任一者:所述末尾数据集合中包含的所述图像数据所对应的所述待处理图像的帧数大于或等于预设帧数阈值;所述末尾数据集合中任一所述图像数据所属的待处理图像的相机位置与所述当前待处理图像的相机位置之间的距离大于预设距离阈值;所述末尾数据集合中任一所述图像数据所属的待处理图像的相机朝向角度与所述当前待处理图像的相机朝向角度之间的差异大于预设角度阈值;其中,所述相机位置和所述相机朝向角度是利用所述待处理图像的相机位姿参数计算得到的。
- 根据权利要求1至10任一项所述的三维重建方法,其中,每帧所述待处理图像包括色彩数据和深度数据,所述利用每帧所述待处理图像和所述摄像器件的标定参数,确定每帧所述待处理图像属于所述待重建目标的目标像素点包括:获取与所述色彩数据对齐之后的深度数据中所包含的每一像素点的法向量与所述待处理图像的重力方向之间的夹角;将所述每一像素点在三维空间投影至所述重力方向,得到所述每一像素点在所述三维空间的高度值;对所述夹角满足预设角度条件的像素点的高度值进行分析,得到所述待重建目标的平面高度;利用所述平面高度,筛选所述色彩数据中属于所述待重建物体的目标像素点。
- 根据权利要求11所述的三维重建方法,其中,所述对所述夹角满足预设角度条件的像素点的高度值进行分析,得到所述待重建目标的平面高度包括:将所述夹角满足预设角度条件的所述像素点的高度值,作为一高度集合;对所述高度集合中的高度值进行聚类分析,得到所述待重建目标的平面高度。
- 根据权利要求1至12任一项所述的三维重建方法,其中,所述依次利用各个所述数据集合的图像数据,及时序位于其之前的数据集合的图像数据和位姿优化参数,确定每一所述数据集合的位姿优化参数之后,所述方法还包括:依次将每一所述数据集合中的图像数据映射至三维空间,得到与每一所述数据集合对应的三维点 云;采用每一所述数据集合的所述位姿优化参数将与其对应的所述三维点云进行调整。
- 一种基于三维重建的交互方法,包括:获取待重建目标的三维模型,其中,所述三维模型是利用权利要求1至13任一项所述的三维重建方法得到的;利用预设视觉惯导方式,构建摄像器件所在场景的三维地图,并获取所述摄像器件当前在所述三维地图中的位姿信息;基于所述位姿信息,在所述摄像器件当前拍摄到的场景图像中显示所述三维模型。
- 一种基于三维重建的测量方法,包括:获取待重建目标的三维模型,其中,所述三维模型是利用权利要求1至13任一项所述的三维重建方法得到的;接收用户在所述三维模型上设置的多个测量点;获取所述多个测量点之间的距离,得到所述待重建目标上对应于所述多个测量点的位置之间的距离。
- 一种三维重建装置,包括:图像获取部分,配置为获取摄像器件扫描待重建目标得到的多帧待处理图像;第一确定部分,配置为利用每帧所述待处理图像和所述摄像器件的标定参数,确定每帧所述待处理图像属于所述待重建目标的目标像素点及其相机位姿参数;数据划分部分,配置为按照预设划分策略,依次将各帧所述待处理图像的图像数据划分至对应的数据集合,其中,所述图像数据至少包括所述目标像素点;第二确定部分,依次利用各个所述数据集合的图像数据,及时序位于其之前的数据集合的图像数据和位姿优化参数,确定每一所述数据集合的位姿优化参数;参数调整部分,配置为利用每一所述数据集合的位姿优化参数,对包含于所述数据集合内的图像数据所属的待处理图像的相机位姿参数进行调整;模型重建部分,配置为利用预设三维重建方式和所述待处理图像的调整后的相机位姿参数,对所述待处理图像的图像数据进行重建处理,得到所述待重建目标的三维模型。
- 根据权利要求16所述的三维重建装置,其中,所述第二确定部分包括:数据集合选取子部分,配置为依次将每一所述数据集合作为当前数据集合,并选取至少一个时序位于所述当前数据集合之前的数据集合,作为候选数据集合;空间变换参数子部分,配置为利用所述当前数据集合的图像数据和所述候选数据集合的图像数据,确定所述当前数据集合和所述候选数据集合之间的空间变换参数;位姿优化参数子部分,配置为至少利用所述候选数据集合的位姿优化参数,以及所述当前数据集合与所述候选数据集合之间的空间变换参数,获得所述当前数据集合的位姿优化参数,并至少更新所述候选数据集合的位姿优化参数。
- 根据权利要求17所述的三维重建装置,其中,所述位姿优化参数子部分包括:数据集合对部分,配置为将分别与所述当前数据集合以及时序位于其之前的所述候选数据集合相关的各个空间变换参数所对应的两个数据集合,作为一数据集合对;目标函数构建部分,配置为利用各个所述数据集合对的空间变换参数,以及各自的位姿优化参数,构建一关于所述位姿优化参数的目标函数;目标函数求解,配置为利用预设求解方式对所述目标函数进行求解,得到所述当前数据集合及时序位于其之前的所述候选数据集合各自对应的数据集合对所包含的数据集合的位姿优化参数。
- 根据权利要求17所述的三维重建装置,其中,所述空间变换参数子部分包括:图像数据搜索部分,配置为在所述候选数据集合和所述当前数据集合中搜索一组满足预设匹配条件的待匹配图像数据;匹配像素点选取部分,配置为基于从每组所述待匹配图像数据中提取得到的预设图像特征,得到每组所述待匹配图像数据之间的匹配像素点对;三维空间映射部分,配置为将所述匹配像素点对中属于所述当前数据集合的像素点映射至三维空间,得到第一三维匹配点,并将所述匹配像素点对中属于所述候选数据集合的像素点映射至所述三维空间,得到第二三维匹配点;三维匹配点对齐部分,配置为将所述第一三维匹配点和所述第二三维匹配点进行对齐处理,得到所述空间变换参数。
- 根据权利要求19所述的三维重建装置,其中,三维匹配点对齐部分包括:第一位姿变换参数子部分,配置为获取所述第一三维匹配点和所述第二三维匹配点之间的第一位姿变换参数;三维匹配点优化子部分,配置为利用所述第一位姿变换参数和预设位姿变换参数,对所述第一三维匹配点进行位姿优化,分别得到第一优化匹配点和第二优化匹配点;第二位姿变换参数子部分,配置为计算所述第二三维匹配点分别与所述第一优化匹配点、所述第二优化匹配点之间的重合度,并选取所述重合度较高的优化匹配点所采用的位姿变换参数,作为第二位姿变换参数;空间变换参数子部分,配置为以所述第二位姿变换参数作为初始值,利用预设对齐方式将所述第一三维匹配点和所述第二三维匹配点进行对齐处理,得到所述当前数据集合与所述候选数据集合之间的空间变换参数。
- 根据权利要求19所述的三维重建装置,其中,所述空间变换参数子部分还包括:变换参数筛选单元,配置为在利用所述当前数据集合的图像数据和所述候选数据集合的图像数据,确定所述当前数据集合和所述候选数据集合之间的空间变换参数之后,以及所述至少利用所述候选数据集合的位姿优化参数,以及所述当前数据集合与所述候选数据集合之间的空间变换参数,获得所述当前数据集合的位姿优化参数之前,从所述当前数据集合与各个所述候选数据集合之间的空间变换参数中,选取符合预设参数筛选条件的空间变换参数;其中,所述预设参数筛选条件包括以下任一者:所述空间变换参数相关的所述候选数据集合与所述当前数据集合相邻;利用所述空间变换参数对所述第一三维匹配点进行位姿优化得到的优化匹配点,与所述第二三维匹配点之间的重合度大于一预设重合度阈值。
- 根据权利要求16所述的三维重建装置,其中,所述数据集合选取子部分包括:词袋模型构建单元,配置为利用所述当前数据集合及时序位于其之前的数据集合中的图像数据的预设图像特征,构建词袋模型;待匹配图像数据单元,配置为选取所属的待处理图像位于所述当前数据集合中的预设时序处的图像数据,作为待匹配图像数据;图像特征查询单元,配置为从所述词袋模型的预设范围中,查询与所述待匹配图像数据的预设图像特征之间的相似度评分大于一预设相似度阈值的预设图像特征;候选数据集合单元,配置为将查询到的预设图像特征所属的图像数据所在的数据集合,以及与所述当前数据集合相邻的数据集合,作为所述候选数据集合;其中,所述预设范围包括所属的数据集合与所述当前数据集合不相邻,且不包含于所述当前数据集合中的图像数据的预设图像特征。
- 根据权利要求22所述的三维重建装置,其中,所述数据集合选取子部分还包括:最大相似度评分值获取单元,配置为在从所述词袋模型的预设范围中,查询与所述待匹配图像数据的预设图像特征之间的相似度评分大于一预设相似度阈值的预设图像特征之前,获取与所述当前数据集合相邻的数据集合中各个所述图像数据与所述待匹配图像数据之间的相似度评分中的最大评分值;预设相似度阈值确定单元,配置为将所述最大评分值的预设倍数和一预设评分值中的任一者作为所述预设相似度阈值。
- 根据权利要求16所述的三维重建装置,其中,所述数据划分部分包括:当前待处理图像确定子部分,配置为依次将各帧所述待处理图像作为当前待处理图像;数据处理子部分,配置为在对当前待处理图像的图像数据进行划分时,若已有的所述数据集合中的末尾数据集合满足预设溢出条件,则获取所述末尾数据集合中最新的多帧所述待处理图像的图像数据,并存入一新创建的所述数据集合,作为新的所述末尾数据集合,将所述当前待处理图像的图像数据划分至新的所述末尾数据集合。
- 根据权利要求24所述的三维重建装置,其中,所述预设溢出条件包括以下任一者:所述末尾数据集合中包含的所述图像数据所对应的所述待处理图像的帧数大于或等于预设帧数阈值;所述末尾数据集合中任一所述图像数据所属的待处理图像的相机位置与所述当前待处理图像的相机位置之间的距离大于预设距离阈值;所述末尾数据集合中任一所述图像数据所属的待处理图像的相机朝向角度与所述当前待处理图像的相机朝向角度之间的差异大于预设角度阈值;其中,所述相机位置和所述相机朝向角度是利用所述待处理图像的相机位姿参数计算得到的。
- 根据权利要求16至26任一项所述的三维重建装置,其中,每帧所述待处理图像包括色彩数据和深度数据;第一确定部分包括:夹角获取子部分,配置为获取与所述色彩数据对齐之后的深度数据中所包含的每一像素点的法向量与所述待处理图像的重力方向之间的夹角;高度获取子部分,配置为将所述每一像素点在三维空间投影至所述重力方向,得到所述每一像素点在所述三维空间的高度值;高度分析子部分,配置为对所述夹角满足预设角度条件的像素点的高度值进行分析,得到所述待重建目标的平面高度;像素筛选子部分,配置为利用所述平面高度,筛选所述色彩数据中属于所述待重建物体的目标像素点。
- 根据权利要求26所述的三维重建装置,其中,所述高度分析子部分包括:高度集合获取单元,配置为将所述夹角满足预设角度条件的所述像素点的高度值,作为一高度集合;高度聚类分析单元,配置为对所述高度集合中的高度值进行聚类分析,得到所述待重建目标的平面高度。
- 根据权利要求16至27任一项所述的三维重建装置,其中,所述三维重建装置1100还包括:三维映射部分,配置为依次将每一所述数据集合中的图像数据映射至三维空间,得到与每一所述数据集合对应的三维点云;点云调整部分,配置为采用每一所述数据集合的所述位姿优化参数将与其对应的所述三维点云进行调整。
- 一种基于三维重建的交互装置,其特征在于,包括:模型获取部分,配置为获取待重建目标的三维模型,其中,所述三维模型是利用权利要求16所述的三维重建装置得到的;建图定位部分,配置为利用预设视觉惯导方式,构建摄像器件所在场景的三维地图,并获取所述摄像器件当前在所述三维地图中的位姿信息;显示交互部分,配置为基于所述位姿信息,在所述摄像器件当前拍摄到的场景图像中显示所述三维模型。
- 一种基于三维重建的测量装置,其特征在于,包括:模型获取部分,配置为获取待重建目标的三维模型,其中,所述三维模型是利用权利要求16所述的三维重建装置得到的;显示交互部分,配置为接收用户在所述三维模型上设置的多个测量点;距离获取部分,配置为获取所述多个测量点之间的距离,得到所述待重建目标上对应于所述多个测量点的位置之间的距离。
- 一种电子设备,其特征在于,包括相互耦接的存储器和处理器,所述处理器用于执行所述存储器中存储的程序指令,以实现权利要求1至13任一项所述的三维重建方法,或实现权利要求14所述的基于三维重建的交互方法,或实现权利要求15所述的基于三维重建的测量方法。
- 一种计算机可读存储介质,其上存储有程序指令,其特征在于,所述程序指令被处理器执行时实现权利要求1至13任一项所述的三维重建方法,或实现权利要求14所述的基于三维重建的交互方法,或实现权利要求15所述的基于三维重建的测量方法。
- 一种计算机程序,包括计算机可读代码,在所述计算机可读代码在电子设备中运行,被所述电子设备中的处理器执行的情况下,实现权利要求1至13任一项所述的三维重建方法,或实现权利要求14所述的基于三维重建的交互方法,或实现权利要求15所述的基于三维重建的测量方法。
- 一种计算机程序产品,当其在计算机上运行时,使得计算机执行如权利要求1至13任一项所述的三维重建方法,或执行权利要求14所述的基于三维重建的交互方法,或执行权利要求15所述的基于三维重建的测量方法。
Priority Applications (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
KR1020237025998A KR20230127313A (ko) | 2021-01-11 | 2021-06-28 | 3차원 재구성 및 관련 인터랙션, 측정 방법 및 관련장치, 기기 |
JP2023513719A JP7453470B2 (ja) | 2021-01-11 | 2021-06-28 | 3次元再構成及び関連インタラクション、測定方法及び関連装置、機器 |
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110031502.0A CN112767538B (zh) | 2021-01-11 | 2021-01-11 | 三维重建及相关交互、测量方法和相关装置、设备 |
CN202110031502.0 | 2021-01-11 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2022147976A1 true WO2022147976A1 (zh) | 2022-07-14 |
Family
ID=75701311
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2021/102882 WO2022147976A1 (zh) | 2021-01-11 | 2021-06-28 | 三维重建及相关交互、测量方法和相关装置、设备 |
Country Status (4)
Country | Link |
---|---|
JP (1) | JP7453470B2 (zh) |
KR (1) | KR20230127313A (zh) |
CN (1) | CN112767538B (zh) |
WO (1) | WO2022147976A1 (zh) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115661371A (zh) * | 2022-12-14 | 2023-01-31 | 深圳思谋信息科技有限公司 | 三维对象建模方法、装置、计算机设备及存储介质 |
CN115690693A (zh) * | 2022-12-13 | 2023-02-03 | 山东鲁旺机械设备有限公司 | 一种建筑吊篮的智能监控系统及监控方法 |
CN116758157A (zh) * | 2023-06-14 | 2023-09-15 | 深圳市华赛睿飞智能科技有限公司 | 一种无人机室内三维空间测绘方法、系统及存储介质 |
CN116863087A (zh) * | 2023-06-01 | 2023-10-10 | 中国航空油料集团有限公司 | 基于数字孪生的航油信息显示方法、装置和可读存储介质 |
CN117168313A (zh) * | 2023-11-03 | 2023-12-05 | 武汉工程大学 | 基于光栅投影三维重建的相位误差模型校正方法及系统 |
CN117476509A (zh) * | 2023-12-27 | 2024-01-30 | 联合富士半导体有限公司 | 一种用于半导体芯片产品的激光雕刻装置及控制方法 |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112767538B (zh) * | 2021-01-11 | 2024-06-07 | 浙江商汤科技开发有限公司 | 三维重建及相关交互、测量方法和相关装置、设备 |
CN113450417A (zh) * | 2021-05-12 | 2021-09-28 | 深圳市格灵精睿视觉有限公司 | 标定参数优化方法、装置、设备及存储介质 |
CN113240656B (zh) * | 2021-05-24 | 2023-04-07 | 浙江商汤科技开发有限公司 | 视觉定位方法及相关装置、设备 |
CN115222799B (zh) * | 2021-08-12 | 2023-04-11 | 达闼机器人股份有限公司 | 图像重力方向的获取方法、装置、电子设备及存储介质 |
CN116051723B (zh) * | 2022-08-03 | 2023-10-20 | 荣耀终端有限公司 | 集束调整方法及电子设备 |
CN116704152B (zh) * | 2022-12-09 | 2024-04-19 | 荣耀终端有限公司 | 图像处理方法和电子设备 |
CN116486008B (zh) * | 2023-04-12 | 2023-12-12 | 荣耀终端有限公司 | 一种三维重建方法、显示方法及电子设备 |
CN117152399A (zh) * | 2023-10-30 | 2023-12-01 | 长沙能川信息科技有限公司 | 基于变电站的模型制作方法、装置、设备和存储介质 |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108986037A (zh) * | 2018-05-25 | 2018-12-11 | 重庆大学 | 基于半直接法的单目视觉里程计定位方法及定位系统 |
CN110910493A (zh) * | 2019-11-29 | 2020-03-24 | 广州极飞科技有限公司 | 三维重建方法、装置及电子设备 |
US10733718B1 (en) * | 2018-03-27 | 2020-08-04 | Regents Of The University Of Minnesota | Corruption detection for digital three-dimensional environment reconstruction |
CN112767538A (zh) * | 2021-01-11 | 2021-05-07 | 浙江商汤科技开发有限公司 | 三维重建及相关交互、测量方法和相关装置、设备 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9892552B2 (en) * | 2015-12-15 | 2018-02-13 | Samsung Electronics Co., Ltd. | Method and apparatus for creating 3-dimensional model using volumetric closest point approach |
CN108537876B (zh) * | 2018-03-05 | 2020-10-16 | 清华-伯克利深圳学院筹备办公室 | 三维重建方法、装置、设备及存储介质 |
CN109166149B (zh) * | 2018-08-13 | 2021-04-02 | 武汉大学 | 一种融合双目相机与imu的定位与三维线框结构重建方法与系统 |
-
2021
- 2021-01-11 CN CN202110031502.0A patent/CN112767538B/zh active Active
- 2021-06-28 WO PCT/CN2021/102882 patent/WO2022147976A1/zh active Application Filing
- 2021-06-28 JP JP2023513719A patent/JP7453470B2/ja active Active
- 2021-06-28 KR KR1020237025998A patent/KR20230127313A/ko active Search and Examination
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US10733718B1 (en) * | 2018-03-27 | 2020-08-04 | Regents Of The University Of Minnesota | Corruption detection for digital three-dimensional environment reconstruction |
CN108986037A (zh) * | 2018-05-25 | 2018-12-11 | 重庆大学 | 基于半直接法的单目视觉里程计定位方法及定位系统 |
CN110910493A (zh) * | 2019-11-29 | 2020-03-24 | 广州极飞科技有限公司 | 三维重建方法、装置及电子设备 |
CN112767538A (zh) * | 2021-01-11 | 2021-05-07 | 浙江商汤科技开发有限公司 | 三维重建及相关交互、测量方法和相关装置、设备 |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115690693A (zh) * | 2022-12-13 | 2023-02-03 | 山东鲁旺机械设备有限公司 | 一种建筑吊篮的智能监控系统及监控方法 |
CN115661371A (zh) * | 2022-12-14 | 2023-01-31 | 深圳思谋信息科技有限公司 | 三维对象建模方法、装置、计算机设备及存储介质 |
CN116863087A (zh) * | 2023-06-01 | 2023-10-10 | 中国航空油料集团有限公司 | 基于数字孪生的航油信息显示方法、装置和可读存储介质 |
CN116863087B (zh) * | 2023-06-01 | 2024-02-02 | 中国航空油料集团有限公司 | 基于数字孪生的航油信息显示方法、装置和可读存储介质 |
CN116758157A (zh) * | 2023-06-14 | 2023-09-15 | 深圳市华赛睿飞智能科技有限公司 | 一种无人机室内三维空间测绘方法、系统及存储介质 |
CN116758157B (zh) * | 2023-06-14 | 2024-01-30 | 深圳市华赛睿飞智能科技有限公司 | 一种无人机室内三维空间测绘方法、系统及存储介质 |
CN117168313A (zh) * | 2023-11-03 | 2023-12-05 | 武汉工程大学 | 基于光栅投影三维重建的相位误差模型校正方法及系统 |
CN117168313B (zh) * | 2023-11-03 | 2024-01-23 | 武汉工程大学 | 基于光栅投影三维重建的相位误差模型校正方法及系统 |
CN117476509A (zh) * | 2023-12-27 | 2024-01-30 | 联合富士半导体有限公司 | 一种用于半导体芯片产品的激光雕刻装置及控制方法 |
CN117476509B (zh) * | 2023-12-27 | 2024-03-19 | 联合富士半导体有限公司 | 一种用于半导体芯片产品的激光雕刻装置及控制方法 |
Also Published As
Publication number | Publication date |
---|---|
CN112767538B (zh) | 2024-06-07 |
JP2023540917A (ja) | 2023-09-27 |
CN112767538A (zh) | 2021-05-07 |
JP7453470B2 (ja) | 2024-03-19 |
KR20230127313A (ko) | 2023-08-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022147976A1 (zh) | 三维重建及相关交互、测量方法和相关装置、设备 | |
US20210232924A1 (en) | Method for training smpl parameter prediction model, computer device, and storage medium | |
CN111598998B (zh) | 三维虚拟模型重建方法、装置、计算机设备和存储介质 | |
WO2021175050A1 (zh) | 三维重建方法和三维重建装置 | |
CN108509848B (zh) | 三维物体的实时检测方法及系统 | |
US11928800B2 (en) | Image coordinate system transformation method and apparatus, device, and storage medium | |
WO2020001168A1 (zh) | 三维重建方法、装置、设备和存储介质 | |
Stoll et al. | Fast articulated motion tracking using a sums of gaussians body model | |
WO2019205852A1 (zh) | 确定图像捕捉设备的位姿的方法、装置及其存储介质 | |
WO2021057743A1 (zh) | 地图融合方法及装置、设备、存储介质 | |
WO2015139574A1 (zh) | 一种静态物体重建方法和系统 | |
WO2015135323A1 (zh) | 一种摄像机跟踪方法及装置 | |
US20170330375A1 (en) | Data Processing Method and Apparatus | |
CN111625667A (zh) | 一种基于复杂背景图像的三维模型跨域检索方法及系统 | |
JP2019096113A (ja) | キーポイントデータに関する加工装置、方法及びプログラム | |
TWI785588B (zh) | 圖像配準方法及其相關的模型訓練方法、設備和電腦可讀儲存媒體 | |
US20240046557A1 (en) | Method, device, and non-transitory computer-readable storage medium for reconstructing a three-dimensional model | |
US7200269B2 (en) | Non-rigid image registration using distance functions | |
US20200057778A1 (en) | Depth image pose search with a bootstrapped-created database | |
WO2022247126A1 (zh) | 视觉定位方法、装置、设备、介质及程序 | |
CN114627244A (zh) | 三维重建方法及装置、电子设备、计算机可读介质 | |
WO2022142049A1 (zh) | 地图构建方法及装置、设备、存储介质、计算机程序产品 | |
WO2023078135A1 (zh) | 三维建模方法和装置、计算机可读存储介质及计算机设备 | |
JP2002520969A (ja) | 動き画像からの自動化された3次元シーン走査方法 | |
CN114638921A (zh) | 动作捕捉方法、终端设备及存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 21917022 Country of ref document: EP Kind code of ref document: A1 |
|
ENP | Entry into the national phase |
Ref document number: 2023513719 Country of ref document: JP Kind code of ref document: A |
|
ENP | Entry into the national phase |
Ref document number: 20237025998 Country of ref document: KR Kind code of ref document: A |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21917022 Country of ref document: EP Kind code of ref document: A1 |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 21917022 Country of ref document: EP Kind code of ref document: A1 |