CN112950787A - Target object three-dimensional point cloud generation method based on image sequence - Google Patents
Target object three-dimensional point cloud generation method based on image sequence Download PDFInfo
- Publication number
- CN112950787A CN112950787A CN202110225992.8A CN202110225992A CN112950787A CN 112950787 A CN112950787 A CN 112950787A CN 202110225992 A CN202110225992 A CN 202110225992A CN 112950787 A CN112950787 A CN 112950787A
- Authority
- CN
- China
- Prior art keywords
- image
- image sequence
- kpm
- gray
- pixel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 41
- 239000011159 matrix material Substances 0.000 claims description 38
- 230000011218 segmentation Effects 0.000 claims description 20
- 238000004364 calculation method Methods 0.000 claims description 10
- 238000013519 translation Methods 0.000 claims description 6
- 238000000354 decomposition reaction Methods 0.000 claims description 4
- 238000003709 image segmentation Methods 0.000 claims description 4
- 230000000694 effects Effects 0.000 claims description 3
- 238000001914 filtration Methods 0.000 claims description 3
- 238000009499 grossing Methods 0.000 claims description 3
- 230000001131 transforming effect Effects 0.000 claims description 3
- 238000011161 development Methods 0.000 abstract description 7
- 238000001514 detection method Methods 0.000 abstract description 3
- 238000000605 extraction Methods 0.000 abstract description 3
- 238000012545 processing Methods 0.000 abstract description 3
- 230000018109 developmental process Effects 0.000 description 5
- 238000005516 engineering process Methods 0.000 description 3
- 238000004422 calculation algorithm Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000009825 accumulation Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 230000004927 fusion Effects 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000004088 simulation Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T17/00—Three dimensional [3D] modelling, e.g. data description of 3D objects
- G06T17/20—Finite element generation, e.g. wire-frame surface description, tesselation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F17/00—Digital computing or data processing equipment or methods, specially adapted for specific functions
- G06F17/10—Complex mathematical operations
- G06F17/16—Matrix or vector computation, e.g. matrix-matrix or matrix-vector multiplication, matrix factorization
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/136—Segmentation; Edge detection involving thresholding
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10028—Range image; Depth image; 3D point clouds
Abstract
The invention discloses a target object three-dimensional point cloud generating method based on an image sequence, which is characterized in that on the basis of a segmented image, the segmented target image sequence is subjected to feature point detection, extraction and matching and error matching pair reduction to obtain an effective matching pair, and then three-dimensional point cloud reconstruction is carried out on the effective matching pair to obtain a target object three-dimensional point cloud model with higher quality and precision. The method can effectively improve the efficiency of the three-dimensional modeling process, improve the quality of the three-dimensional point cloud model, reduce the probability of occurrence of processing errors in each link, is suitable for development of intelligent robot development, and has important engineering value and theoretical guidance significance for the efficiency of intelligent robot operation, especially underwater robot operation.
Description
Technical Field
The invention belongs to the technical field of computer vision simulation, and particularly relates to a method for three-dimensional reconstruction of a shot continuous image sequence.
Background
With the progress of image processing technology and the development of three-dimensional sensors, the demands of three-dimensional application scenes such as intelligent mobile robots, unmanned driving, AR and the like are rapidly increased, and the three-dimensional reconstruction technology is required. Different from two-dimensional images, the three-dimensional model has stronger sense of reality, can present more information of people, is also different from drawing modeling, and the three-dimensional scene reconstruction technology has higher requirements on reflecting the original scene situation in real time more quickly and better, wherein the three-dimensional modeling is realized only by an image sequence, although the economic cost is reduced, the three-dimensional model is difficult to ensure the quality, the precision and the accuracy of the three-dimensional model construction, while the reconstruction of continuous multiple images needs to consider the error accumulation generated by observation at different angles, and a process of optimizing or balancing the three-dimensional model exists. Therefore, in order to realize the real-scene three-dimensional model reconstruction quickly and well, the research on the three-dimensional reconstruction method based on the image sequence has important engineering value and practical significance.
At present, there are many methods for improving the three-dimensional reconstruction efficiency and precision of an image sequence, and the methods can be roughly divided into two types: one is that the image is divided to obtain a target image and a background image to respectively carry out three-dimensional reconstruction, and finally the reconstruction of an image three-dimensional model is realized through the splicing and fusion of point clouds; and the other type is that the three-dimensional reconstruction is directly carried out on the scene image, and then the obtained point cloud model segmentation is carried out on the optimization model.
Disclosure of Invention
The invention aims to solve the problems of low quality and poor accuracy of the existing three-dimensional point cloud model, and provides a target object three-dimensional point cloud generation method based on an image sequence.
The purpose of the invention is realized as follows:
a target object three-dimensional point cloud generation method based on an image sequence comprises the following steps:
acquiring a source image sequence, and calculating a gray level image sequence of the source image sequence;
performing threshold segmentation on each image in the gray image sequence, determining an optimal segmentation threshold, segmenting the gray image sequence according to the optimal segmentation threshold, reserving pixels larger than the optimal segmentation threshold as target object pixels to obtain a target gray image sequence, and correspondingly reserving the target object pixels in the source image sequence to obtain the target image sequence;
detecting characteristic points of the target image sequence, extracting and storing the characteristic points, and calculating the module value and the direction of the characteristic points;
and (3) an effective quasi-matching pair in the component target image sequence, wherein the module values and the directions of two characteristic points in the effective quasi-matching pair are equal, and the module values and the directions are recordedA qth valid aligned pair of the pth image and the (p + 1) th image in the target image sequence; deleting the mismatching pairs from the effective quasi-matching pairs to obtain effective matching pairs;
calculating a basic matrix F between the p image and the p +1 image according to the obtained effective matching pairp(ii) a Acquiring an internal parameter matrix K of the camera; according to the obtained basic matrix FpComputing the eigenmatrix EpFor intrinsic matrix EpCalculating an extrinsic parameter rotation matrix R of the camera between the p-th image and the p + 1-th image by singular value decompositionpTranslation matrix Tp:
According to a rotation matrix RpTranslation matrix TpCalculating point cloud data of effective matching pairs between the p image and the p +1 imagePosition:
wherein X ' ═ (X ', y ', 1) is the position of the feature point in the p +1 th image in the valid matching pair;
and calculating all the effective matching pairs to form a point cloud data set, transforming the obtained point cloud data set to the same coordinate system, and eliminating repeated data points to generate the three-dimensional point cloud of the target object.
The calculation method of the gray level image sequence of the source image sequence comprises the following steps:
performing gray value calculation on each pixel point of each image in the source image sequence:
Gray=0.299R+0.587G+0.114B (2)
gray is a Gray value, and R, G, B are R, G, B three channel values of the pixel points respectively;
and storing the gray value of each pixel point in the pixel point of a new image sequence corresponding to the pixel point of the source image sequence, and taking the new image sequence as a gray image sequence.
The determination method of the optimal segmentation threshold comprises the following steps:
(a1) counting the number of pixels corresponding to each gray scale from 0 to 255 in the gray image sequence, and taking mu (t) (t is 0,1,2, … and 255) as the number of pixels corresponding to the gray value t;
Wherein p [ mu (t) ] is a probability function of the number of pixels mu (t) with the gray value of t;
(a3) calculating the pixel C of the target image of the object to be segmented in the image0Average gray level mu of1:
Where Pr (i | C0) indicates that the number of pixels having a gradation value i accounts for the pixel C of the divided object target image0Probability of (P)iThe probability that the pixel with the gray value i accounts for the total number of the pixels is obtained;
(a4) calculating the pixel C of the target image of the object to be segmented in the image0Number of pixels ofProportion omega1:
Wherein N is1Is a target image C0The number of pixels, Sum is the total number of image pixels;
(a5) calculating a background image pixel C in an image1Average gray level mu of2:
Where Pr (i | C1) indicates that the number of pixels with a gray level value of i accounts for background image pixel C1Probability of (P)iThe probability that the pixel with the gray value i accounts for the total number of the pixels is obtained;
(a6) calculating a background image pixel C in an image1Is the ratio omega of the number of pixels2:
Wherein N is2As a background image C1The number of pixels;
(a7) calculating an optimal segmentation threshold tbestSaid optimal segmentation threshold tbestTo maximize the value of t for the inter-class variance g (t),
wherein, the value of alpha belongs to (0,1) is adjusted according to each image segmentation binarization effect graph.
The detection method of the characteristic points comprises the following steps:
each image pixel value f (x, y) in the target grayscale image sequence is processed as follows:
(b1) the image pixel value f (x, y) is subjected to Gaussian filtering G under different Gaussian smoothing parameters sigmaσAnd (3) calculating:
(b2) DoG response image DoG f that calculates image pixel value f (x, y):
DoGf(x,y)=DoG*f(x,y)=g1(x,y)-g2(x,y) (10)
(b3) repeating the steps (b1) and (b2) to obtain Dogf1(x,y)、DoGf2(x,y)、DoGf3(x,y);
For Dogf2All values in (x, y) are detected if a certain DoGf2(x, y) in DogfiIf the value of (x + n, y + m) (i is 1,2, 3; n is-1, 0, 1; and m is-1, 0,1) is maximum or minimum, marking the pixel as a characteristic point, and recording the characteristic pointThe position of the characteristic point is (x, y) for the kth characteristic point of the mth image in the target image sequence.
And setting the module value of the characteristic point as M, wherein the calculation method comprises the following steps:
L(x,y,σ)=G(x,y,σ)*I(x,y) (12)
σ is an autonomously selected parameter and remains unchanged in all feature points.
Let the direction of the feature point be θ, the calculation method is:
θ(x,y)=tan-1((L(x,y+1)-L(x,y-1))/(L(x+1,y)-L(x-1,y))) (13)。
the method for deleting the mismatching pairs comprises the following steps:
(c1) from valid quasi-matching pairs KPMq' inSelecting the first M non-collinear sample data, wherein M is more than or equal to 4, calculating an optimal homography matrix according to the formula, and marking the optimal homography matrix as a model M;
wherein (x, y), (x ', y') respectively represent valid quasi-matching pairs KPMq' the positions of two pairs of characteristic points in the equation, s is a scale parameter, and h is usually given as33Normalizing the matrix by 1;
(c2) calculating an effective quasi-matching pair KPM based on the model Mq' projection error of all data in e:
if e is less than 0.75, retaining the data, otherwise, removing the data;
(c3) let optimal valid matching pair KPM'best,KPM′bestInitial value of (1) ═ KPM1', if the current valid quasi-matching is to KPMq'the number of elements is greater than the optimal effective matching pair KPM'bestAnd then KPM'best=KPMq', update the iteration number kk at the same time:
wherein, p is confidence coefficient, and takes value of 0.995, and w is KPM'bestNumber of elements/KPMq' number of elements, number of initial iterations kk0Middle KPM'bestThe number of the elements is 1;
(c4) if the iteration times are more than kk, exiting the loop and recording KPMqFor efficient matching of KPMq=KPM′best(ii) a Otherwise, the number of iterations is increased by 1, and the above step (c3) is repeated.
The internal parameter matrix of the camera is as follows:
wherein f isx,fyIs the focal length, x0、y0Is the coordinate of the principal point, and s is the inclination parameter of the coordinate axis.
Basis matrix F between p image and p +1 imagepThe calculation method comprises the following steps:
wherein X is K-1(x,y,1),X′=K-1(x ', y', 1) is an effective matching pairThe origin is located at the center of the image.
Compared with the prior art, the invention has the beneficial effects that:
the invention provides a target object three-dimensional point cloud generating method based on an image sequence, which takes target object three-dimensional point cloud generation based on the image sequence as a research object, deeply analyzes various process collocation of three-dimensional modeling and advantages and disadvantages of an algorithm, and designs the target object image sequence three-dimensional point cloud generating method based on image segmentation. The invention is suitable for development of intelligent robot development, and has important engineering value and theoretical guidance significance for the operation of the intelligent robot, especially the operation efficiency of the underwater robot.
Drawings
FIG. 1 is a flow chart of a method for generating a three-dimensional point cloud of a target object based on an image sequence;
FIG. 2 is a flow chart of a target object thresholding method based on an image sequence;
FIG. 3 is a schematic diagram of a feature point extraction algorithm model.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings and specific embodiments.
As shown in fig. 1, the method for generating a three-dimensional point cloud of a target object based on an image sequence specifically includes the following steps:
1. acquiring a source image sequence from shooting equipment, and recording the source image sequence as S1; calculating a R, G, B three-channel value of each pixel point of each image in the source image sequence S1 according to the following formula (1), obtaining a Gray value Gray of each pixel point, storing the Gray value into the pixel points of a new image sequence corresponding to the pixel points of the source image sequence, and recording the new image sequence as a Gray image sequence S2.
Gray=0.299R+0.587G+0.114B (1)
2. Each image in the grayscale image sequence S2 is threshold-segmented to determine an optimal segmentation threshold.
The flow chart of the threshold segmentation is shown in fig. 2, and the specific steps are as follows:
the number of pixels corresponding to each gradation level is counted from 0 to 255, and μ (t) (t is 0,1,2, …,255) is the number of pixels corresponding to the gradation value t.
(a1) Counting the number of pixels corresponding to each gray scale from 0 to 255 in the gray image sequence, and taking mu (t) (t is 0,1,2, … and 255) as the number of pixels corresponding to the gray value t;
Wherein p [ mu (t) ] is a probability function of the number of pixels mu (t) with the gray value of t; (ii) a
(a3) Calculating the pixel C of the target image of the object to be segmented in the image0Average gray level mu of1:
Where Pr (i | C0) indicates that the number of pixels having a gradation value i accounts for the pixel C of the divided object target image0Probability of (P)iIs the probability of the pixel with the gray value i occupying the total number of pixels.
(a4) Calculating the pixel C of the target image of the object to be segmented in the image0Is the ratio omega of the number of pixels1:
Wherein N is1Is a target image C0The number of pixels, Sum is the total number of image pixels;
(a5) calculating the average gray level μ of the background image pixel C1 in the image2:
Where Pr (i | C1) indicates that the number of pixels with a gray level value of i accounts for background image pixel C1Probability of (P)iIs the probability of the pixel with the gray value i occupying the total number of pixels.
(a6) Calculating a background image pixel C in an image1The ratio of the number of pixels:
wherein N is2As a background image C1The number of pixels;
(a7) calculating an optimal segmentation threshold tbestSaid optimal segmentation threshold tbestTo maximize the value of t for the inter-class variance g (t),
g(t)=ω1 α-2(ω1μ-μ(t))2+ω2 α-2(μ(ω2-1)+μ(t))2 (7)
wherein, the value of alpha belongs to (0,1) is adjusted according to each image segmentation binarization effect graph.
According to the optimal segmentation threshold tbestThe gray image sequence S2 is divided to be less than tbestThe pixel of (A) is a background pixel C1Greater than tbestIs the target object pixel C0At this time, the target object pixel C0Reserving to obtain an object target gray level image sequence SAblackAnd correspondingly reserving pixel points in the source image sequence S1 to obtain a segmented target image sequence SA.
3. The method for detecting, extracting and storing the feature points of the target image sequence SA includes the following steps, and a flowchart of the method for extracting the feature points is shown in fig. 3:
for target gray image sequence SAblackIs processed as follows:
(b1) the image pixel value f (x, y) is subjected to Gaussian filtering G under different Gaussian smoothing parameters sigmaσAnd (3) calculating:
(b2) DoG response image DoG f that calculates image pixel value f (x, y):
DoGf(x,y)=DoG*f(x,y)=g1(x,y)-g2(x,y) (9)
(b3) repeating the steps (b1) and (b2) to obtain Dogf1(x,y)、DoGf2(x,y)、DoGf3(x,y);
For Dogf2All values in (x, y) are detected if a certain DoGf2(x, y) in DogfiIf the value of (x + n, y + m) (i is 1,2, 3; n is-1, 0, 1; and m is-1, 0,1) is maximum or minimum, marking the pixel as a characteristic point, and recording the characteristic pointThe position of the characteristic point is (x, y) for the kth characteristic point of the mth image in the target image sequence.
L(x,y,σ)=G(x,y,σ)*I(x,y) (11)
θ(x,y)=tan-1((L(x,y+1)-L(x,y-1))/(L(x+1,y)-L(x-1,y))) (12)
4. for the characteristic points in the target image sequence SAAre matched in pairs, i.e.Anda pair of feature points KP with the middle module value M and the direction theta equal are effectively matched and recordedFor the q (q ═ 1,2,3, …) th valid quasi-matching pair of the p-th image and the p + 1-th image (p ═ 1,2,3, …) in the target image sequence SA, the mis-matching pair is deleted from the valid quasi-matching pair, resulting in a valid matching pair.
The specific steps for deleting the mismatching pairs are as follows:
(c1) from valid quasi-matching pairs KPMqSelecting the previous M non-collinear sample data, wherein M is more than or equal to 4, calculating an optimal homography matrix according to the formula, and marking the optimal homography matrix as a model M;
wherein (x, y), (x ', y') respectively represent valid quasi-matching pairs KPMq' the positions of two pairs of characteristic points in the equation, s is a scale parameter, and h is usually given as33Normalizing the matrix by 1;
(c2) calculating an effective quasi-matching pair KPM based on the model Mq' projection error of all data in e:
if e is less than 0.75, retaining the data, otherwise, removing the data;
(c3) let optimal valid matching pair KPM'best,KPM′bestInitial value of (1) ═ KPM1', if the current valid quasi-matching is to KPMq'the number of elements is greater than the optimal effective matching pair KPM'bestAnd then KPM'best=KPMq', update the iteration number kk at the same time:
wherein, p is confidence coefficient, and takes value of 0.995, and w is KPM'bestNumber of elements/KPMq' number of elements, number of initial iterations kk0Middle KPM'bestThe number of the elements is 1;
(c4) if the iteration times are more than kk, exiting the loop and recording KPMqFor efficient matching of KPMq=KPM′best(ii) a Otherwise, the number of iterations is increased by 1, and the above step (c3) is repeated.
5. Obtaining internal parameters of the camera by referring to relevant information of the camera or using a camera calibration method, wherein an internal parameter matrix K comprises:
wherein f isx,fyFor focal length, x is generally equal to0、y0Is the principal point coordinate (relative to the imaging plane) and s is the coordinate axis tilt parameter, ideally 0.
According to the obtained effective matching pairsCalculating a basis matrix F between the p-th image and the p + 1-th image by equation (16)p。
Wherein X is K-1(x,y,1),X′=K-1(x ', y', 1) is an effective matching pairThe origin is located at the center of the image.
According to the obtained basic matrix FpCalculating an extrinsic parameter rotation matrix R of the camera between the p-th image and the p + 1-th image by equation (18)pTranslation matrix Tp。
Fp=C-TEpC-1,Ep=Tp×Rp(18)
Where C is the vector of the position of the camera center in the world coordinate system given by itself, often let C equal the camera coordinate system of the p-th image, EpFor the intrinsic matrix of the camera between the p-th image and the p + 1-th image, by matching the intrinsic matrix EpThe exterior of the camera between the p-th image and the p + 1-th image can be calculated by Singular Value Decomposition (Singular Value Decomposition)Partial parameter rotation matrix RpTranslation matrix Tp。
6. Calculating the effective matching pair between the p-th image and the p + 1-th image by the formula (19)Point cloud data ofPosition:
will effectively match with the pairAll calculated and recorded as a point cloud data set Xp(p ═ 1,2,3, …). The obtained point cloud data set XpAnd transforming to the same coordinate system, and eliminating repeated data points to complete the generation of the three-dimensional point cloud of the target object.
In summary, the following steps: the invention discloses a target object three-dimensional point cloud generating method based on an image sequence, which is characterized in that on the basis of a segmented image, the segmented target image sequence is subjected to feature point detection, extraction and matching and error matching pair reduction to obtain an effective matching pair, and then the effective matching pair is subjected to three-dimensional point cloud reconstruction to obtain a target object three-dimensional point cloud model with higher quality and precision. The method can effectively improve the efficiency of the three-dimensional modeling process, improve the quality of the three-dimensional point cloud model, reduce the probability of occurrence of processing errors in each link, is suitable for development of intelligent robot development, and has important engineering value and theoretical guidance significance for the efficiency of intelligent robot operation, especially underwater robot operation.
Claims (9)
1. A target object three-dimensional point cloud generating method based on an image sequence is characterized by comprising the following steps:
acquiring a source image sequence, and calculating a gray level image sequence of the source image sequence;
performing threshold segmentation on each image in the gray image sequence, determining an optimal segmentation threshold, segmenting the gray image sequence according to the optimal segmentation threshold, reserving pixels larger than the optimal segmentation threshold as target object pixels to obtain a target gray image sequence, and correspondingly reserving the target object pixels in the source image sequence to obtain the target image sequence;
detecting characteristic points of the target image sequence, extracting and storing the characteristic points, and calculating the module value and the direction of the characteristic points;
and (3) an effective quasi-matching pair in the component target image sequence, wherein the module values and the directions of two characteristic points in the effective quasi-matching pair are equal, and the module values and the directions are recordedA qth valid aligned pair of the pth image and the (p + 1) th image in the target image sequence; deleting the mismatching pairs from the effective quasi-matching pairs to obtain effective matching pairs;
calculating a basic matrix F between the p image and the p +1 image according to the obtained effective matching pairp(ii) a Acquiring an internal parameter matrix K of the camera; according to the obtained basic matrix FpComputing the eigenmatrix EpFor intrinsic matrix EpCalculating an extrinsic parameter rotation matrix R of the camera between the p-th image and the p + 1-th image by singular value decompositionpTranslation matrix Tp:
According to a rotation matrix RpTranslation matrix TpCalculating point cloud data of effective matching pairs between the p image and the p +1 imagePosition:
wherein X ' ═ (X ', y ', 1) is the position of the feature point in the p +1 th image in the valid matching pair;
and calculating all the effective matching pairs to form a point cloud data set, transforming the obtained point cloud data set to the same coordinate system, and eliminating repeated data points to generate the three-dimensional point cloud of the target object.
2. The method for generating a three-dimensional point cloud of a target object based on an image sequence according to claim 1, wherein the calculation method of the gray-scale image sequence of the source image sequence is as follows:
performing gray value calculation on each pixel point of each image in the source image sequence:
Gray=0.299R+0.587G+0.114B (2)
gray is a Gray value, and R, G, B are R, G, B three channel values of the pixel points respectively;
and storing the gray value of each pixel point in the pixel point of a new image sequence corresponding to the pixel point of the source image sequence, and taking the new image sequence as a gray image sequence.
3. The method for generating a three-dimensional point cloud of a target object based on an image sequence according to claim 1 or 2, wherein the optimal segmentation threshold is determined by:
(a1) counting the number of pixels corresponding to each gray scale from 0 to 255 in the gray image sequence, and taking mu (t) (t is 0,1,2, … and 255) as the number of pixels corresponding to the gray value t;
Wherein p [ mu (t) ] is a probability function of the number of pixels mu (t) with the gray value of t;
(a3) calculating the pixel C of the target image of the object to be segmented in the image0Average gray level mu of1:
Where Pr (i | C0) indicates that the number of pixels having a gradation value i accounts for the pixel C of the divided object target image0Probability of (P)iThe probability that the pixel with the gray value i accounts for the total number of the pixels is obtained;
(a4) calculating the pixel C of the target image of the object to be segmented in the image0Is the ratio omega of the number of pixels1:
Wherein N is1Is a target image C0The number of pixels, Sum is the total number of image pixels;
(a5) calculating a background image pixel C in an image1Average gray level mu of2:
Where Pr (i | C1) indicates that the number of pixels with a gray level value of i accounts for background image pixel C1Probability of (P)iThe probability that the pixel with the gray value i accounts for the total number of the pixels is obtained;
(a6) calculating a background image pixel C in an image1Is the ratio omega of the number of pixels2:
Wherein N is2As a background image C1The number of pixels;
(a7) calculating an optimal segmentation threshold tbestSaid optimal segmentation threshold tbestTo maximize the value of t for the inter-class variance g (t),
wherein, the value of alpha belongs to (0,1) is adjusted according to each image segmentation binarization effect graph.
4. The method for generating a three-dimensional point cloud of a target object based on an image sequence according to claim 1 or 2, wherein the method for detecting the feature points comprises:
each image pixel value f (x, y) in the target grayscale image sequence is processed as follows:
(b1) the image pixel value f (x, y) is subjected to Gaussian filtering G under different Gaussian smoothing parameters sigmaσAnd (3) calculating:
(b2) DoG response image DoG f that calculates image pixel value f (x, y):
DoGf(x,y)=DoG*f(x,y)=g1(x,y)-g2(x,y) (10)
(b3) repeating the steps (b1) and (b2) to obtain Dogf1(x,y)、DoGf2(x,y)、DoGf3(x,y);
For Dogf2All values in (x, y) are detected if a certain DoGf2(x, y) in DogfiIf the value of (x + n, y + m) (i is 1,2, 3; n is-1, 0, 1; and m is-1, 0,1) is maximum or minimum, marking the pixel as a characteristic point, and recording the characteristic pointThe position of the characteristic point is (x, y) for the kth characteristic point of the mth image in the target image sequence.
5. The method for generating a three-dimensional point cloud of a target object based on an image sequence according to claim 4, wherein the module value of the feature point is M, and the calculation method comprises:
L(x,y,σ)=G(x,y,σ)*I(x,y) (12)
σ is an autonomously selected parameter and remains unchanged in all feature points.
6. The method for generating a three-dimensional point cloud of a target object based on an image sequence according to claim 5, wherein the direction of the feature point is θ, and the calculation method comprises:
θ(x,y)=tan-1((L(x,y+1)-L(x,y-1))/(L(x+1,y)-L(x-1,y))) (13)。
7. the method for generating the three-dimensional point cloud of the target object based on the image sequence according to claim 1 or 2, wherein the method for deleting the mismatching pairs comprises:
(c1) from valid quasi-matching pairs KPMqSelecting the previous M non-collinear sample data, wherein M is more than or equal to 4, calculating an optimal homography matrix according to the formula, and marking the optimal homography matrix as a model M;
wherein (x, y), (x ', y') respectively represent valid quasi-matching pairs KPMq' the positions of two pairs of feature points in the equation, s is a scale parameter, and h is usually given33Normalizing the matrix by 1;
(c2) calculating an effective quasi-matching pair KPM based on the model Mq' projection error of all data in e:
if e is less than 0.75, retaining the data, otherwise, removing the data;
(c3) let optimal valid matching pair KPM'best,KPM′bestInitial value of (1) ═ KPM1', if the current valid quasi-matching is to KPMq'the number of elements is greater than the optimal effective matching pair KPM'bestAnd then KPM'best=KPMq', update the iteration number kk at the same time:
wherein, p is confidence coefficient, and takes value of 0.995, and w is KPM'bestNumber of elements/KPMq' number of elements, number of initial iterations kk0Middle KPM'bestThe number of the elements is 1;
(c4) if the iteration times are more than kk, exiting the loop and recording KPMqFor efficient matching of KPMq=KPM′best(ii) a Otherwise, the number of iterations is increased by 1, and the above step (c3) is repeated.
8. The method of generating a three-dimensional point cloud of a target object based on an image sequence according to claim 1 or 2, wherein the internal parameter matrix of the camera is:
wherein f isx,fyIs the focal length, x0、y0Is the coordinate of the principal point, and s is the inclination parameter of the coordinate axis.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110225992.8A CN112950787B (en) | 2021-03-01 | 2021-03-01 | Target object three-dimensional point cloud generation method based on image sequence |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110225992.8A CN112950787B (en) | 2021-03-01 | 2021-03-01 | Target object three-dimensional point cloud generation method based on image sequence |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112950787A true CN112950787A (en) | 2021-06-11 |
CN112950787B CN112950787B (en) | 2022-08-02 |
Family
ID=76246962
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110225992.8A Active CN112950787B (en) | 2021-03-01 | 2021-03-01 | Target object three-dimensional point cloud generation method based on image sequence |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112950787B (en) |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116071520A (en) * | 2023-03-31 | 2023-05-05 | 湖南省水务规划设计院有限公司 | Digital twin water affair simulation test method |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106157373A (en) * | 2016-07-27 | 2016-11-23 | 中测高科(北京)测绘工程技术有限责任公司 | A kind of construction three-dimensional model building method and system |
CN110517348A (en) * | 2019-07-23 | 2019-11-29 | 西安电子科技大学 | Target object three-dimensional point cloud method for reconstructing based on display foreground segmentation |
-
2021
- 2021-03-01 CN CN202110225992.8A patent/CN112950787B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106157373A (en) * | 2016-07-27 | 2016-11-23 | 中测高科(北京)测绘工程技术有限责任公司 | A kind of construction three-dimensional model building method and system |
CN110517348A (en) * | 2019-07-23 | 2019-11-29 | 西安电子科技大学 | Target object three-dimensional point cloud method for reconstructing based on display foreground segmentation |
Non-Patent Citations (4)
Title |
---|
ZHUHAITAO 等: "A feature-based Section Curve Reconstructing Strategy", 《KEY ENGINEERING MATERIALS》 * |
朱冰琳等: "基于机器视觉的大田植株生长动态三维定量化研究", 《农业机械学报》 * |
王晖等: "城市园林建设中花卉色彩图像匹配仿真研究", 《计算机仿真》 * |
王欢 等: "利用Kinect深度信息的三维点云配准方法研究", 《计算机工程与应用》 * |
Cited By (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN116071520A (en) * | 2023-03-31 | 2023-05-05 | 湖南省水务规划设计院有限公司 | Digital twin water affair simulation test method |
Also Published As
Publication number | Publication date |
---|---|
CN112950787B (en) | 2022-08-02 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109784333B (en) | Three-dimensional target detection method and system based on point cloud weighted channel characteristics | |
CN110108258B (en) | Monocular vision odometer positioning method | |
CN109740665B (en) | Method and system for detecting ship target with occluded image based on expert knowledge constraint | |
CN108597009B (en) | Method for detecting three-dimensional target based on direction angle information | |
CN110555908B (en) | Three-dimensional reconstruction method based on indoor moving target background restoration | |
CN111862201A (en) | Deep learning-based spatial non-cooperative target relative pose estimation method | |
CN111696196B (en) | Three-dimensional face model reconstruction method and device | |
CN113052109A (en) | 3D target detection system and 3D target detection method thereof | |
CN109325444B (en) | Monocular texture-free three-dimensional object posture tracking method based on three-dimensional geometric model | |
CN111797688A (en) | Visual SLAM method based on optical flow and semantic segmentation | |
CN113065546A (en) | Target pose estimation method and system based on attention mechanism and Hough voting | |
CN113850865A (en) | Human body posture positioning method and system based on binocular vision and storage medium | |
CN113888461A (en) | Method, system and equipment for detecting defects of hardware parts based on deep learning | |
CN112541423A (en) | Synchronous positioning and map construction method and system | |
CN112950787B (en) | Target object three-dimensional point cloud generation method based on image sequence | |
CN115082498A (en) | Robot grabbing pose estimation method, device, equipment and storage medium | |
CN112767478B (en) | Appearance guidance-based six-degree-of-freedom pose estimation method | |
CN116977328B (en) | Image quality evaluation method in active vision of vehicle bottom robot | |
CN112053385B (en) | Remote sensing video shielding target tracking method based on deep reinforcement learning | |
CN112686202A (en) | Human head identification method and system based on 3D reconstruction | |
CN116843738A (en) | Tree dumping risk assessment system and method based on TOF depth camera | |
CN116862984A (en) | Space pose estimation method of camera | |
CN116643291A (en) | SLAM method for removing dynamic targets by combining vision and laser radar | |
CN111709269A (en) | Human hand segmentation method and device based on two-dimensional joint information in depth image | |
CN113592947B (en) | Method for realizing visual odometer by semi-direct method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
EE01 | Entry into force of recordation of patent licensing contract |
Application publication date: 20210611 Assignee: Enshobo (Guangzhou) Technology Co.,Ltd. Assignor: HARBIN ENGINEERING University Contract record no.: X2023980048395 Denomination of invention: A 3D point cloud generation method for target objects based on image sequences Granted publication date: 20220802 License type: Common License Record date: 20231127 |
|
EE01 | Entry into force of recordation of patent licensing contract |