CN104778721A - Distance measuring method of significant target in binocular image - Google Patents
Distance measuring method of significant target in binocular image Download PDFInfo
- Publication number
- CN104778721A CN104778721A CN201510233157.3A CN201510233157A CN104778721A CN 104778721 A CN104778721 A CN 104778721A CN 201510233157 A CN201510233157 A CN 201510233157A CN 104778721 A CN104778721 A CN 104778721A
- Authority
- CN
- China
- Prior art keywords
- mrow
- msup
- msub
- image
- point
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 44
- 239000011159 matrix material Substances 0.000 claims abstract description 68
- 230000000007 visual effect Effects 0.000 claims abstract description 21
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 19
- 238000012545 processing Methods 0.000 claims abstract description 14
- 238000005259 measurement Methods 0.000 claims abstract description 12
- 238000000605 extraction Methods 0.000 claims abstract description 8
- 238000005295 random walk Methods 0.000 claims abstract description 6
- 239000013598 vector Substances 0.000 claims description 17
- 238000003384 imaging method Methods 0.000 claims description 16
- 230000011218 segmentation Effects 0.000 claims description 15
- 238000010606 normalization Methods 0.000 claims description 13
- 238000004364 calculation method Methods 0.000 claims description 11
- 238000012216 screening Methods 0.000 claims description 9
- 230000002093 peripheral effect Effects 0.000 claims description 8
- 238000003708 edge detection Methods 0.000 claims description 7
- 238000010586 diagram Methods 0.000 claims description 6
- 238000001914 filtration Methods 0.000 claims description 6
- 238000001514 detection method Methods 0.000 claims description 5
- 238000003709 image segmentation Methods 0.000 claims description 4
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 claims description 3
- 230000003287 optical effect Effects 0.000 claims description 3
- 238000007781 pre-processing Methods 0.000 claims description 3
- 238000012887 quadratic function Methods 0.000 claims description 3
- 238000012217 deletion Methods 0.000 claims description 2
- 230000037430 deletion Effects 0.000 claims description 2
- 230000005764 inhibitory process Effects 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims description 2
- 230000004927 fusion Effects 0.000 claims 1
- 238000000638 solvent extraction Methods 0.000 abstract 1
- 238000000691 measurement method Methods 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
Landscapes
- Image Analysis (AREA)
Abstract
The invention relates to a distance measuring method of a significant target in a binocular image. The distance measuring method of the significant target in the binocular image aims to solve the problem that an existing target distance measuring method is low in the processing speed. The method includes the steps that step 1, significant feature extraction is conducted on the binocular image through a visual significance model, and a seed point and a background point are marked; step 2, a weighed graph is established for the binocular image; step 3, a significant target in the binocular image is partitioned through a random walk image partitioning algorithm by means of the seed point and the background point in step 1 and the weighed graph in step 2; step 4, key point matching is conducted on the significant target separately through an SIFT algorithm; step 5, the significant target distance is worked out by applying a parallax matrix K' worked out in step 4 into a binocular distance measuring model. The distance measuring method can be applied to distance measurement of the significant target of the image in front of the vision in the intelligent automobile running process.
Description
Technical Field
The invention relates to a distance measuring method for a target in a binocular image, in particular to a distance measuring method for a significant target in the binocular image, and belongs to the technical field of image processing.
Background
The distance information is mainly applied to providing safety judgment for a control system of an automobile in traffic image processing. In the research process of intelligent automobiles, the traditional target measurement method is to measure the distance of a target by using radar or laser with a specific wavelength. Compared with radar and laser, the visual sensor has the advantages of price and wider visual angle. And the specific content of the target can be judged while measuring the target distance by using the vision sensor.
However, the current traffic image information is relatively complex, the traditional target distance measurement algorithm is difficult to obtain an ideal result in a complex image, and because a salient target in the image cannot be found but is detected globally, the processing speed is low, a lot of irrelevant data is added, and the algorithm cannot meet the requirements of practical application.
Disclosure of Invention
The invention aims to provide a distance measuring method for a significant target in a binocular image, which aims to solve the problem that the existing target distance measuring method is low in processing speed.
The invention relates to a distance measuring method of a significant target in a binocular image, which is realized according to the following steps: the method comprises the following steps of firstly, extracting the saliency characteristics of a binocular image by using a visual saliency model, marking out seed points and background points, and specifically comprising the following steps:
the method comprises the following steps of firstly, extracting the saliency characteristics of a binocular image by using a visual saliency model, marking out seed points and background points, and specifically comprising the following steps:
firstly, preprocessing is carried out, edge detection is carried out on a binocular image, and an edge map of the binocular image is generated; performing saliency feature extraction on the binocular image by using a visual saliency model to generate a saliency feature map;
step three, finding out a pixel point with the maximum gray value in the image according to the saliency characteristic map, and marking the pixel point as a seed point; traversing pixels in a 25 multiplied by 25 window with the seed point as the center, and finding out pixel points with the gray value of less than 0.1 and the farthest distance from the seed point as background points;
step two, establishing a weighted graph for the binocular image;
establishing a weighted graph for the binocular image by using a classical Gaussian weight function:
wherein, WijRepresents the weight, g, between vertex i and vertex jiDenotes the brightness of the vertex i, gjRepresenting the brightness of the vertex j, beta is a free parameter, and e is a natural base number;
the laplace matrix L of the weighted graph is found by:
wherein L isijElements of the Laplace matrix L corresponding to vertices i to j, diIs the sum of the vertex i and the surrounding point weights, di=∑Wij;
Thirdly, segmenting the salient objects in the binocular image by using the seed points and the background points in the first step and the weighted graph in the second step through a random walk image segmentation algorithm;
step three, dividing pixel points of the binocular image into two sets according to the seed points and the background points marked in the step one, namely a marked point set VMAnd the unmarked point set VULaplace matrix L according to VMAnd VUPreferentially arranging the marked points and then arranging the non-marked points; wherein said L is divided into LM、LU、B、BTAnd four parts, representing the Laplace matrix as follows:
wherein L isMLaplace matrix, L, from mark point to mark pointULaplace matrix from unmarked point to unmarked point, B and BTLaplace matrixes from mark points to non-mark points and from non-mark points to mark points respectively;
step two, solving a combined Dirichlet integral Dx according to the Laplace matrix and the mark points;
the combined dirichlet integral formula is as follows:
wherein x is a probability matrix from a top point to a mark point in the weighted graph, and xiAnd xjThe probabilities from the vertexes i and j to the mark points are respectively;
according to the set V of the mark pointsMAnd the unmarked point set VUDividing x into xMAnd xUTwo moieties, xMSet of marked points VMCorresponding probability matrix, xUFor unmarked point set VUA corresponding probability matrix; decomposing formula (4) into:
for the mark point s, m is setsIf any vertex i is s, thenOtherwiseFor D [ x ]u]For xUAnd (3) differentiating to obtain a solution of the minimum value of the formula (5), namely the Dirichlet probability value of the mark point s:
wherein,representing the probability that the vertex i reaches the mark point s for the first time;
from a combined Dirichlet integralPerforming threshold segmentation according to equation (7) to generate a segmentation map:
wherein s isiThe pixel size of a corresponding position of a certain vertex i in the segmentation graph is obtained;
the pixel point with the brightness of 1 in the segmentation graph is represented as a saliency target in the image, and the pixel point with the brightness of 0 is the background;
thirdly, multiplying the segmentation image by the pixels corresponding to the original image to generate a target image, namely extracting the segmented saliency target, wherein the formula is as follows:
ti=si·Ii (8)
wherein, tiIs the gray value of a certain vertex I of the target graph TiA gray value corresponding to the position I of the input image I (sigma);
fourthly, independently matching key points of the significant targets through an SIFT algorithm;
establishing a Gaussian pyramid for the target image, calculating the difference between every two filtered images to obtain a DOG image, wherein the DOG image is defined as D (x, y, sigma), and the formula is as follows:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*T(x,y)
(9)
=C(x,y,kσ)-C(x,y,σ)
wherein,the method comprises the following steps that (1) a Gaussian function with a variable scale is adopted, p and q represent the dimension of a Gaussian template, (x and y) represent the positions of pixel points in a Gaussian pyramid image, sigma is a scale space factor of the image, k represents a specific scale value, and C (x, y and sigma) is defined as the convolution of G (x, y and sigma) and a target image T (x and y), namely C (x, y and sigma) is G (x, y and sigma) T (x and y);
step two, solving extreme points in adjacent DOG images, determining the positions and scales of the extreme points as key points by fitting a three-dimensional quadratic function, and performing stability detection on the key points according to a Hessian matrix to eliminate edge response, wherein the method specifically comprises the following steps:
(one) solving a curve fitting D (X) of DOG in the scale space by performing Taylor expansion:
where X ═ (X, y, σ)TD is curve fitting, and the offset formula (11) of the extreme point is obtained by deriving the formula (10) and making it 0:
to remove the extreme points of low contrast, equation (11) is substituted into equation (10) to obtain equation (12):
if the value of the formula (12) is greater than 0.03, the extreme point is reserved, and the accurate position and scale of the extreme point are obtained, otherwise, the extreme point is discarded;
secondly, unstable key points are eliminated through Hessian matrix screening at the key points;
calculating curvature by using the ratio between characteristic values of the Hessian matrix;
judging edge points according to the curvatures of the key point neighborhoods;
the curvature ratio is set to be 10, if the curvature ratio is larger than 10, deleting the curvature ratio, otherwise, keeping the curvature ratio, and keeping the curvature ratio as a stable key point;
step four, using the pixel of the 16 × 16 window of the key point neighborhood to assign a direction parameter for each key point;
for the key points detected in the DOG image, the magnitude and direction of the gradient are calculated as follows:
(13)
θ(x,y)=tan-1((C(x,y+1)-C(x,y-1))/(C(x+1,y)-C(x-1,y)))
wherein C is the scale space where the key point is located, m is the gradient size of the key point, and theta is the gradient direction of the solved point; with the key point as the center, a 16 x 16 neighborhood is defined in the surrounding area, the gradient size and the gradient direction of the pixel point are solved, and the gradient of the point in the neighborhood is counted by using a histogram; the abscissa of the histogram is the direction, divide 360 degrees into 36, each is a item in 10 degrees corresponding histogram, the ordinate of the histogram is the magnitude of gradient, add for the magnitude of the point of the corresponding gradient direction, its sum is the magnitude of the ordinate; the main direction is defined as the interval direction with the maximum gradient size hm, and the interval with the gradient size above 08 × hm is used as the auxiliary direction of the main direction to enhance the matching stability;
fourthly, establishing descriptors to express local characteristic information of key points
Firstly, rotating coordinates around a key point to be the direction of the key point;
then, a 16 × 16 window around the keypoint is selected, the neighborhood is divided into 16 4 × 4 small windows, the size and direction of the corresponding gradient in the 4 × 4 small windows are calculated, the gradient information of each small window is counted by using a histogram of 8 bins, and the descriptor of the 16 × 16 window around the keypoint is calculated by a gaussian weighting algorithm according to the following formula:
wherein h is a descriptor, (a, b) is the position of the key point in the Gaussian pyramid image, and mgTaking the gradient size of the key point, namely the gradient size of the main direction of the histogram in the fourth step and the third step, d is the side length of the window, namely 16, (x, y) is the position of the pixel point in the Gaussian pyramid image, (x'Y') is the new coordinate of the pixel in the neighborhood of the direction of rotating the coordinate to the keypoint, the calculation formula for the new coordinate is as follows:
θgthe gradient direction of the key point;
calculating the characteristic vector of 128 key points by a 16-by-16 windowIs denoted by H ═ H1,h2,h3,...,h128) Normalizing the feature vector, and recording the normalized feature vector as LgThe normalization formula is as follows:
wherein L isg=(l1,l2,...,li,...,l128) Feature vectors for the keypoints after normalization, li1,2, 3.. is some normalized vector;
matching the key points in the binocular image by using Euclidean distance of the feature vectors of the key points as judgment measurement of similarity of the key points in the binocular image, wherein the coordinate information of the key points which are matched with each other is used as a group of key information;
step four, screening the generated matching key points;
calculating the horizontal parallax of each pair of key points to generate a parallax matrix, wherein the parallax matrix is defined as Kn={k1,k2...knN is the logarithm of the match, k1、k2、knAs a single matching point disparity;
determining parallaxMedian k of the matrixmAnd obtaining a reference parallax matrix, which is marked as Kn', the formula is as follows:
Kn'={k1-km,k2-km,...,kn-km} (17)
setting parallax threshold value to be 3, and setting Kn' deletion of corresponding parallaxes larger than the threshold value in the ' sequence to obtain the final visual matrix result K ', K1'、k2'、kn'All the parallax errors are the parallax errors of the screened correct matching points, n' is the logarithm of the final correct matching, and the formula is as follows:
K'={k1',k2',...,kn'} (18)
substituting the parallax matrix K' obtained in the step four into a binocular ranging model to obtain a significant target distance;
the focal lengths of the two identical imaging systems are separated by J along the horizontal direction, the two optical axes are parallel to the horizontal plane, and the image plane is parallel to the vertical plane;
suppose a target point M (X, Y, Z) in the scene, Pl (X) at the left and right imaging points respectively1,y1) And Pr (x)2,y2),x1,y1And x2,y2Coordinates of Pl and Pr in the vertical plane of the image, respectively, and the parallax in the binocular model is defined as k ═ Pl-Pr | -x2-x1And l, obtaining a distance formula according to a triangular similarity relation, wherein X, Y and Z are coordinates of a horizontal axis, a vertical axis and a longitudinal axis in a space coordinate system:
wherein dx 'represents the physical distance of each pixel in the horizontal axis direction of the imaged negative film, f is the focal length of the imaging system, Z is the distance from the target point M to the connecting line of the two imaging centers, the parallax matrix obtained in the step four is taken into a formula (19), and the corresponding distance matrix Z' ═ Z is obtained according to the physical information of the binocular model1,z2,...,zn'},z1,z2,zn'The distance of the significant target is calculated for single matching parallax, and finally the average value of the distance matrix is calculated, namely the distance Z of the significant target in the binocular imagefThe formula is as follows:
the invention has the beneficial effects that:
1. the invention adopts a method of simulating a human visual system to extract the region interested by human eyes, and the significance target extracted by an algorithm is basically consistent with the detection result of the human eyes, so that the significance target can be automatically identified as the human eyes by extraction.
2. The invention automatically completes the distance measurement of the saliency target without manually selecting the saliency target.
3. The method matches the same target, thereby ensuring that the parallax results of key point matching are similar, effectively screening out wrong matching points, ensuring that the matching accuracy is close to 100 percent, ensuring that the relative error of the parallax is less than 2 percent, and increasing the accuracy of distance measurement.
4. The method has less matching information, can effectively reduce additional irrelevant calculation, reduces the matching calculation by at least 75 percent, reduces the introduction of irrelevant data, has the utilization rate of the matching data of more than 90 percent, realizes the measurement of the distance of the significant target in a complex image environment, and improves the image processing efficiency.
5. The invention measures the distance of the image significance target in front of the field of vision during the driving of the intelligent automobile, thereby providing key information for the safe driving of the automobile, solving the defect that the traditional image distance measurement can only carry out depth detection on the whole image, and well avoiding the problems of larger error and excessive noise.
6. According to the method, the saliency features of the binocular images are extracted, the saliency target is segmented, the target range is narrowed, the matching time is shortened, the efficiency is improved, the saliency target key points are matched, the parallax is calculated, the distance measurement is further realized, wrong matching key points can be well screened out due to the fact that the target is on a vertical plane, the accuracy is improved, and the method can be used for rapidly identifying the saliency target and accurately measuring the distance of the saliency target.
Drawings
FIG. 1 is a flow chart of the method of the present invention;
FIG. 2 is a flow chart of visual saliency analysis;
FIG. 3 is a flow chart of a random walk algorithm;
FIG. 4 is a SIFT algorithm flow chart;
FIG. 5 shows a binocular measurement system, where X, Y, and Z are defined spatial coordinate systems, M is a point in space, Pl and Pr are imaging points of M on an imaging plane, M is a point in space, and f is the focal length of the imaging system.
Detailed Description
The embodiments of the present invention will be described in further detail with reference to the accompanying drawings.
The first embodiment is as follows: the present embodiment is described below with reference to fig. 1 to 5, and the method of the present embodiment includes the steps of:
the method comprises the following steps of firstly, extracting the saliency characteristics of a binocular image by using a visual saliency model, marking out seed points and background points, and specifically comprising the following steps:
and performing saliency extraction on the binocular image by using a visual saliency model, respectively calculating three saliency characteristics of brightness, color and direction of each pixel point of the binocular image, and normalizing the three saliency characteristics to obtain a weighted saliency map of the image. Each pixel on the saliency map represents a magnitude of saliency for a corresponding location in the image. Finding out the point with the maximum pixel value in the graph, namely the point with the strongest significance, and marking the point as a seed point; and gradually expanding the range around the seed point to find the point with the weakest significance, and marking the point as a background point. The process of extracting image saliency using a visual saliency model is shown in fig. 2.
Firstly, preprocessing is carried out, edge detection is carried out on a binocular image, a visual saliency model is generated, and edge information is important saliency information of the image;
performing saliency feature extraction on the binocular image by using a visual saliency model to generate a saliency feature map;
step three, finding out a pixel point with the maximum brightness in the image according to the saliency characteristic map, and marking the pixel point as a seed point; traversing pixels in a 25 multiplied by 25 window with the seed point as the center, and finding out pixel points with the gray value of less than 0.1 and the farthest distance from the seed point as background points;
step two, establishing a weighted graph for the binocular image;
establishing a weighted graph for the binocular image by using a classical Gaussian weight function, giving a certain weight between each pixel point in the binocular image and the surrounding pixels thereof as an edge through different gray levels of the pixels, and establishing the weighted graph containing the vertex and the edge by taking each pixel point as the vertex;
the method comprises the following steps of using graph theory to regard the whole image as an undirected weighted graph and regard each pixel as a vertex in the weighted graph, wherein the edges of the weighted graph are weighted by the gray value of the pixel, and specifically adopting a classical Gaussian weight function as follows:
wherein, WijRepresents the weight, g, between vertex i and vertex jiRepresenting the brightness, g, of the pixel ijRepresents the luminance of pixel j, β is a free parameter, and e is a natural base number;
the laplace matrix L of the weighted graph is found by:
wherein L isijElements of the Laplace matrix L corresponding to vertices i to j, diIs the sum of the vertex i and the surrounding point weights, di=∑Wij;
Thirdly, segmenting the salient objects in the binocular image by using the seed points and the background points in the first step and the weighted graph in the second step through a random walk image segmentation algorithm;
thirdly, segmenting the salient objects in the binocular image by using the seed points and the background points in the first step and the weighted graph in the second step through a random walk image segmentation algorithm;
step three, dividing pixel points of the binocular image into two sets according to the seed points and the background points marked in the step one, namely a marked point set VMAnd the unmarked point set VULaplace matrix L according to VMAnd VUPreferentially arranging the marked points and then arranging the non-marked points; wherein said L is divided into LM、LU、B、BTAnd four parts, representing the Laplace matrix as follows:
wherein L isMLaplace matrix, L, from mark point to mark pointULaplace matrix from unmarked point to unmarked point, B and BTLaplace matrixes from mark points to non-mark points and from non-mark points to mark points respectively;
step two, solving a combined Dirichlet integral Dx according to the Laplace matrix and the mark points;
the combined dirichlet integral formula is as follows:
wherein x is a probability matrix from a top point to a mark point in the weighted graph, and xiAnd xjThe probabilities from the vertexes i and j to the mark points are respectively;
according to the set V of the mark pointsMAnd the unmarked point set VUDividing x into xMAnd xUTwo moieties, xMSet of marked points VMCorresponding probability matrix, xUFor unmarked point set VUA corresponding probability matrix; decomposing formula (4) into:
setting msIs defined as for the marked point s, if any vertex i is s, thenOtherwiseFor D [ x ]u]For xUAnd (3) differentiating to obtain a solution of the minimum value of the formula (5), namely the Dirichlet probability value of the mark point s:
wherein,representing the probability that the vertex i reaches the mark point s for the first time;
from a combined Dirichlet integralPerforming threshold segmentation according to equation (7) to generate a segmentation map:
wherein s isiThe pixel size of a corresponding position of a certain vertex i in the segmentation graph is obtained;
the pixel point with the brightness of 1 in the segmentation graph is represented as a saliency target in the image, and the pixel point with the brightness of 0 is the background;
thirdly, multiplying the segmentation image by the pixels corresponding to the original image to generate a target image, namely extracting the segmented saliency target, wherein the formula is as follows:
ti=si·Ii(8)
wherein, tiIs the gray value of the corresponding position I of the target map TiA gray value corresponding to the position I of the input image I (sigma);
fourthly, independently matching key points of the significant targets through an SIFT algorithm;
and (3) independently detecting and matching key points of the segmented saliency target through an SIFT algorithm, screening the obtained matching coordinate, and providing a result of the error matching to leave a correct matching result.
The process of matching the binocular image by the SIFT algorithm is shown in fig. 4.
Establishing a Gaussian pyramid for the target image, calculating the difference between every two filtered images to obtain a DOG image, wherein the DOG image is defined as D (x, y, sigma), and the formula is as follows:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*T(x,y)
(9)
=C(x,y,kσ)-C(x,y,σ)
wherein,the method comprises the following steps that (1) a Gaussian function with a variable scale is adopted, p and q represent the dimension of a Gaussian template, (x and y) represent the positions of pixel points in a Gaussian pyramid image, sigma is a scale space factor of the image, k represents a specific scale value, and C (x, y and sigma) is defined as the convolution of G (x, y and sigma) and a target image T (x and y), namely C (x, y and sigma) is G (x, y and sigma) T (x and y);
step two, solving extreme points in adjacent DOG images, determining the positions and scales of the extreme points as key points by fitting a three-dimensional quadratic function, and performing stability detection on the key points according to a Hessian matrix to eliminate edge response, wherein the method specifically comprises the following steps:
the key points are composed of local extreme points of the DOG image, each point on the DOG image is traversed, the gray value of 26 points which are the same as 8 adjacent points of the same scale and adjacent upper and lower 2 multiplied by 9 points is detected for each point, and if the gray value is larger or smaller than the adjacent points around, the gray value is the extreme point.
The extreme points found are not real key points, and in order to improve stability, it is necessary to (a) find the curve fitting d (x) of the scale space DOG by taylor expansion:
where X ═ (X, y, σ)TD is curve fitting, and the offset formula (11) of the extreme point is obtained by deriving the formula (10) and making it 0:
to remove the extreme points of low contrast, equation (11) is substituted into equation (10) to obtain equation (12):
if the value of the formula (12) is greater than 0.03, the extreme point is reserved, and the accurate position and scale of the extreme point are obtained, otherwise, the extreme point is discarded;
secondly, unstable key points are eliminated through Hessian matrix screening at the key points;
calculating curvature by using the ratio between characteristic values of the Hessian matrix;
judging edge points according to the curvatures of the key point neighborhoods;
the curvature ratio is set to be 10, if the curvature ratio is larger than 10, deleting the curvature ratio, otherwise, keeping the curvature ratio, and keeping the curvature ratio as a stable key point;
if the value of equation (12) is greater than 0.03, the extreme point is retained and the exact position (the original position plus the offset after fitting) and scale of the extreme point are obtained, otherwise discarded. To eliminate unstable keypoints, screening was performed by Hessian matrix at the keypoints:
and step three, after the positions and the scales of the key points are determined, assigning a direction to the key points, and defining that the key point descriptors are relative to the direction. Using the pixels of the 16 x 16 window of the key point neighborhood to assign a direction parameter to each key point;
for the key points detected in the DOG image, the magnitude and direction of the gradient are calculated as follows:
(13)
θ(x,y)=tan-1((C(x,y+1)-C(x,y-1))/(C(x+1,y)-C(x-1,y)))
wherein C is the scale space where the key point is located, m is the gradient size of the key point, and theta is the gradient direction of the key point; taking a key point as a center, dividing a neighborhood in a surrounding area, and counting the gradient of a point in the neighborhood by using a histogram;
the abscissa of the histogram is the direction, and 360 degrees are divided into 36 shares, each of which is one of 10 degrees corresponding to the histogram. The ordinate of the histogram is the magnitude of the gradient, and the magnitudes of the points corresponding to the respective gradient directions are added, and the sum is the magnitude of the ordinate. The main direction is defined as the interval direction with the maximum gradient size hm, and other intervals with the height of 08 × hm are used as auxiliary directions of the main direction to enhance the stability of matching.
And step four, after the above stages, each detected key point has three kinds of information of position, direction and scale. And establishing a descriptor for each key point to express the local characteristic information of the key point.
The coordinates around the keypoint are first rotated into the direction of the keypoint. Then a 16 x 16 window around the keypoint is chosen, divided into 16 small windows of 4 x 4 in the neighborhood. In a small 4 × 4 window, the magnitude and direction of its corresponding gradient are calculated. And a histogram of 8 bins is used to count the gradient information for each small window. The descriptor is calculated by gaussian weighting algorithm for a 16 × 16 window around the keypoint as follows:
wherein h is a descriptor, (a, b) is the position of a key point in the Gaussian pyramid image, d is the side length of a window, namely 16, (x, y) is the position of a pixel point in the Gaussian pyramid image, and (x ', y') is a new coordinate of the pixel in a neighborhood which rotates the coordinate into the direction of the key point, and the calculation formula of the new coordinate is as follows:
theta is the direction of the key point.
The feature vectors of 128 key points are obtained by calculating a 16 × 16 window, and are recorded as H ═ (H)1,h2,h3,...,h128) In order to reduce the influence of light, the feature vector is normalized, and the normalized feature vector is recorded as LgThe normalization formula is as follows:
wherein L isg=(l1,l2,l3,...,l128) The feature vectors of the key points after normalization;
after descriptors of key points of two images of the binocular image are generated, matching the key points in the binocular image by using Euclidean distance of feature vectors of the key points as judgment measurement of similarity of the key points in the binocular image, and using coordinate information of the key points which are matched with each other as a group of key information;
step four, screening the generated matching key points in order to avoid the generation of errors to the maximum extent;
because the measuring system is a binocular model, the key points of the significance target are twoIn each image, there is a horizontal plane, and the level difference of each pair of key points is theoretically equal. Therefore, the coordinate horizontal parallax of each pair of key points is obtained to generate a parallax matrix, and the parallax matrix is defined as Kn={k1,k2...knN is the logarithm of the match, k1、k2、knAs a single matching point disparity;
determining the median k of the disparity matrixmAnd obtaining a reference parallax matrix, which is marked as Kn', the formula is as follows:
Kn'={k1-km,k2-km,...,kn-km}
setting parallax threshold value to be 3, and setting KnThe corresponding parallaxes larger than the threshold value in the 'are deleted to obtain a final visual matrix result K', so that the interference caused by mismatching key points is avoided. k is a radical of1'、k2'、kn'All the parallax errors are the parallax errors of the screened correct matching points, n' is the logarithm of the final correct matching, and the formula is as follows:
K'={k1',k2',...,kn'}
substituting the parallax matrix K' obtained in the step four into a binocular ranging model to obtain a significant target distance;
and subtracting the coordinate of the key point matched with the saliency target to obtain the parallax of the saliency target in the binocular image. And (4) substituting the parallax into a binocular ranging model to obtain the significant target distance.
The binocular imaging can acquire two images of the same scene at different visual angles, and the binocular model is as shown in figure 5.
The focal lengths of the two identical imaging systems are separated by a distance B along the horizontal direction, the two optical axes are parallel to the horizontal plane, and the image plane is parallel to the vertical plane;
suppose a point M (X, Y, Z) in the scene, Pl (X) at the left and right imaging points, respectively1,y1) And Pr(x2,y2),x1,y1And x2,y2Coordinates of Pl and Pr in the vertical plane of the image, respectively, and the parallax in the binocular model is defined as k ═ Pl-Pr | -x2-x1And l, obtaining a distance formula according to a triangular similarity relation, wherein X, Y and Z are coordinates of a horizontal axis, a vertical axis and a longitudinal axis in a space coordinate system:
wherein dx represents the physical distance of each pixel in the horizontal axis direction of the imaged negative film, f is the focal length of the imaging system, Z is the distance from the target point M to the connecting line of the two imaging centers, the parallax matrix obtained in the step four is taken into a formula (17), and the corresponding distance matrix Z' ═ Z is obtained according to the physical information of the binocular model1,z2,...,zn'},z1,z2,zn'The distance of the significant target is calculated for single matching parallax, and finally the average value of the distance matrix is calculated, namely the distance Z of the significant target in the binocular imagefThe formula is as follows:
the second embodiment is as follows: the present embodiment will be described below with reference to the accompanying drawings, which are different from the specific embodiments in that: step one, the specific process of performing edge detection on the image is as follows:
performing convolution operation on binocular images by adopting a 2D Gaussian filtering template to eliminate noise interference of the images one by one;
step one, two, respectively calculating gradient amplitude and gradient direction of pixels on the filtered binocular image I (x, y) by utilizing the difference of first-order partial derivatives in the horizontal direction and the vertical direction, wherein the partial derivatives dx and dy in the x direction and the y direction are respectively as follows:
dx=[I(x+1,y)-I(x-1,y)]/2 (21)
dy=[I(x,y+1)-I(x,y-1)]/2 (22)
the gradient magnitude is then:
D'=(dx2+dy2)1/2 (23)
the gradient direction is as follows:
θ'=arctan(dy/dx) (24);
d 'and theta' respectively represent the gradient magnitude and gradient direction of pixels on the filtered binocular image I (x, y);
step three, carrying out non-maximum value inhibition on the gradient, and then carrying out double-threshold processing on the image to generate an edge image; the gray value of the edge point of the edge image is 255, and the gray value of the non-edge point is 0.
The third concrete implementation mode: the present embodiment will be described below with reference to the accompanying drawings, which are different from the first or second embodiments in that: the second step is that the visual saliency model is used for carrying out saliency feature extraction on the binocular image, and the specific process of generating the saliency feature map is as follows:
after the first binocular image edge detection, overlapping the original image and the edge image:
I1(σ)=0.7I(σ)+0.3C(σ) (25)
wherein I (sigma) is the original image of the input binocular image, C (sigma) is the edge image, I1(σ) is the image after the superimposition processing;
calculating nine layers of Gaussian pyramids of the images after superposition processing by adopting a Gaussian difference function, wherein the 0 th layer is an input superposition image, the 1 to 8 layers are respectively formed by adopting Gaussian filtering and reduced order sampling on the previous layer, the size of the layer corresponds to 1/2 to 1/256 of the input image, and the brightness, the color and the direction characteristics are extracted for each layer of the Gaussian pyramid to generate a corresponding brightness pyramid, a corresponding color pyramid and a corresponding direction pyramid;
the formula for extracting the brightness features is as follows:
In=(r+g+b)/3 (26)
wherein r, g, b correspond to red, green, blue three components of the input binocular image color, InIs a brightness characteristic;
the formula for extracting the color features is as follows:
R=r-(g+b)/2 (27)
G=g-(r+b)/2 (28)
B=b-(r+g)/2 (29)
Y=r+g-2(|r-g|+b) (30)
r, G, B, Y correspond to the color components of the superimposed image;
o (σ, ω) is for the luminance characteristic InThe directional characteristic of Gabor function filtering extraction is carried out in the scale direction, omega is the direction of the Gabor function, namely the number of Gaussian pyramid layers, and sigma is the total directional quantity of the Gabor function, wherein sigma belongs to [0,1,2 …,8]],ω∈[0°,45°,90°,135°];
Step two, step three, to solving the three characteristics of luminance, colour and direction of different yardstick of gaussian pyramid carry on the central peripheral contrast to make the difference, specifically:
let the scale c (c ∈ {2,3,4}) be the central scale, and the scale u (u ═ c +, ∈ {3,4}) be the peripheral scale; there are 6 combinations (2-5, 2-6, 3-6, 3-7, 4-7, 4-8) between the central dimension c and the peripheral dimension u in the gaussian pyramid of 9 levels;
the local directional feature contrast, which is the difference between the central and peripheral contrast, is represented by the difference between the feature maps of scale c and scale s as follows:
In(c,u)=|In(c)-In(u)| (31)
RG(c,u)=|(R(c)-G(c))-(G(u)-R(u))| (32)
BY(c,u)=|(B(c)-Y(c))-(Y(u)-B(u))| (33)
O(c,u,ω)=|O(c,ω)-O(u,ω)| (34)
before difference making, interpolation is needed to make the sizes of the two images consistent, and then difference making is carried out;
and step two, four, fusing the feature maps of different features generated by difference through normalization to generate a significant feature map of the input binocular image, specifically comprising the following steps:
firstly, the scale contrast characteristic diagram of each characteristic is normalized and fused to generate a comprehensive characteristic diagram of the characteristic The feature map is normalized for the luminance features,the feature map is normalized for the color features,normalizing the feature map for the directional features; the calculation process is shown in the following formula:
wherein, N (-) represents a normalization calculation function, firstly, for a feature map to be calculated, normalizing the feature value of each pixel in the feature map to a closed region [0,255], then finding out a global maximum significant value A in each normalized feature map, then calculating the average value a of local maximum values in the feature map, and finally multiplying the feature value corresponding to each pixel of the feature by 2 (A-a);
and then, carrying out normalization processing by using the comprehensive characteristic diagram of each characteristic to obtain a final significant characteristic diagram S, wherein the calculation process is as follows:
Claims (3)
1. A method for distance measurement of a salient object in a binocular image, the method comprising the steps of:
the method comprises the following steps of firstly, extracting the saliency characteristics of a binocular image by using a visual saliency model, marking out seed points and background points, and specifically comprising the following steps:
firstly, preprocessing is carried out, edge detection is carried out on a binocular image, and an edge map of the binocular image is generated;
performing saliency feature extraction on the binocular image by using a visual saliency model to generate a saliency feature map;
step three, finding out a pixel point with the maximum gray value in the image according to the saliency characteristic map, and marking the pixel point as a seed point; traversing pixels in a 25 multiplied by 25 window with the seed point as the center, and finding out pixel points with the gray value of less than 0.1 and the farthest distance from the seed point as background points;
step two, establishing a weighted graph for the binocular image;
establishing a weighted graph for the binocular image by using a classical Gaussian weight function:
wherein, WijRepresents the weight, g, between vertex i and vertex jiDenotes the brightness of the vertex i, gjRepresenting the brightness of the vertex j, beta is a free parameter, and e is a natural base number;
the laplace matrix L of the weighted graph is found by:
wherein L isijElements of the Laplace matrix L corresponding to vertices i to j, diIs the sum of the vertex i and the surrounding point weights, <math>
<mrow>
<msub>
<mi>d</mi>
<mi>i</mi>
</msub>
<mo>=</mo>
<mi>Σ</mi>
<msub>
<mi>W</mi>
<mi>ij</mi>
</msub>
<mo>;</mo>
</mrow>
</math>
thirdly, segmenting the salient objects in the binocular image by using the seed points and the background points in the first step and the weighted graph in the second step through a random walk image segmentation algorithm;
step three, dividing pixel points of the binocular image into two sets according to the seed points and the background points marked in the step one, namely a marked point set VMAnd the unmarked point set VULaplace matrix L according to VMAnd VUPreferentially arranging the marked points and then arranging the non-marked points; wherein said L is divided into LM、LU、B、BTAnd four parts, representing the Laplace matrix as follows:
wherein L isMLaplace matrix, L, from mark point to mark pointULaplace matrix from unmarked point to unmarked point, B and BTLaplace matrixes from mark points to non-mark points and from non-mark points to mark points respectively;
step two, solving a combined Dirichlet integral Dx according to the Laplace matrix and the mark points;
the combined dirichlet integral formula is as follows:
wherein x is a probability matrix from a top point to a mark point in the weighted graph, and xiAnd xjThe probabilities from the vertexes i and j to the mark points are respectively;
according to the set V of the mark pointsMAnd the unmarked point set VUDividing x into xMAnd xUTwo moieties, xMSet of marked points VMCorresponding probability matrix, xUFor unmarked point set VUA corresponding probability matrix; decomposing formula (4) into:
for the mark point s, m is setsIf any vertex i is s, thenOtherwiseFor D [ x ]u]For xUAnd (3) differentiating to obtain a solution of the minimum value of the formula (5), namely the Dirichlet probability value of the mark point s:
wherein,representing the probability that the vertex i reaches the mark point s for the first time;
from a combined Dirichlet integralPerforming threshold segmentation according to equation (7) to generate a segmentation map:
wherein s isiThe pixel size of a corresponding position of a certain vertex i in the segmentation graph is obtained;
the pixel point with the brightness of 1 in the segmentation graph is represented as a saliency target in the image, and the pixel point with the brightness of 0 is the background;
thirdly, multiplying the segmentation image by the pixels corresponding to the original image to generate a target image, namely extracting the segmented saliency target, wherein the formula is as follows:
ti=si·Ii (8)
wherein, tiIs the gray value of a certain vertex I of the target graph TiA gray value corresponding to the position I of the input image I (sigma);
fourthly, independently matching key points of the significant targets through an SIFT algorithm;
establishing a Gaussian pyramid for the target image, calculating the difference between every two filtered images to obtain a DOG image, wherein the DOG image is defined as D (x, y, sigma), and the formula is as follows:
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*T(x,y)
(9)
=C(x,y,kσ)-C(x,y,σ)
wherein,the method comprises the following steps that (1) a Gaussian function with a variable scale is adopted, p and q represent the dimension of a Gaussian template, (x and y) represent the positions of pixel points in a Gaussian pyramid image, sigma is a scale space factor of the image, k represents a specific scale value, and C (x, y and sigma) is defined as the convolution of G (x, y and sigma) and a target image T (x and y), namely C (x, y and sigma) is G (x, y and sigma) T (x and y);
step two, solving extreme points in adjacent DOG images, determining the positions and scales of the extreme points as key points by fitting a three-dimensional quadratic function, and performing stability detection on the key points according to a Hessian matrix to eliminate edge response, wherein the method specifically comprises the following steps:
(one) solving a curve fitting D (X) of DOG in the scale space by performing Taylor expansion:
where X ═ (X, y, σ)TD is curve fitting, and the offset formula (11) of the extreme point is obtained by deriving the formula (10) and making it 0:
to remove the extreme points of low contrast, equation (11) is substituted into equation (10) to obtain equation (12):
if the value of the formula (12) is greater than 0.03, the extreme point is reserved, and the accurate position and scale of the extreme point are obtained, otherwise, the extreme point is discarded;
secondly, unstable key points are eliminated through Hessian matrix screening at the key points;
calculating curvature by using the ratio between characteristic values of the Hessian matrix;
judging edge points according to the curvatures of the key point neighborhoods;
the curvature ratio is set to be 10, if the curvature ratio is larger than 10, deleting the curvature ratio, otherwise, keeping the curvature ratio, and keeping the curvature ratio as a stable key point;
step four, using the pixel of the 16 × 16 window of the key point neighborhood to assign a direction parameter for each key point;
for the key points detected in the DOG image, the magnitude and direction of the gradient are calculated as follows:
wherein C is the scale space where the key point is located, m is the gradient size of the key point, and theta is the gradient direction of the solved point; with the key point as the center, a 16 x 16 neighborhood is defined in the surrounding area, the gradient size and the gradient direction of the pixel point are solved, and the gradient of the point in the neighborhood is counted by using a histogram; the abscissa of the histogram is the direction, divide 360 degrees into 36, each is a item in 10 degrees corresponding histogram, the ordinate of the histogram is the magnitude of gradient, add for the magnitude of the point of the corresponding gradient direction, its sum is the magnitude of the ordinate; the main direction is defined as the interval direction with the maximum gradient size hm, and the interval with the gradient size above 08 × hm is used as the auxiliary direction of the main direction to enhance the matching stability;
fourthly, establishing descriptors to express local characteristic information of key points
Firstly, rotating coordinates around a key point to be the direction of the key point;
then, a 16 × 16 window around the keypoint is selected, the neighborhood is divided into 16 4 × 4 small windows, the size and direction of the corresponding gradient in the 4 × 4 small windows are calculated, the gradient information of each small window is counted by using a histogram of 8 bins, and the descriptor of the 16 × 16 window around the keypoint is calculated by a gaussian weighting algorithm according to the following formula:
wherein h is a descriptor, (a, b) is the position of the key point in the Gaussian pyramid image, and mgThe gradient size of the key point, namely the gradient size in the main direction of the histogram in the fourth step and the third step, d is the side length of the window, namely 16, (x, y) is the position of the pixel point in the Gaussian pyramid image, (x ', y') is the new coordinate of the pixel in the neighborhood of the direction of rotating the coordinate into the key point, and the calculation formula of the new coordinate is as follows:
θgthe gradient direction of the key point;
the feature vectors of 128 key points are obtained by calculating a 16 × 16 window, and are recorded as H ═ (H)1,h2,h3,...,h128) Normalizing the feature vector, and recording the normalized feature vector as LgThe normalization formula is as follows:
wherein L isg=(l1,l2,...,li,...,l128) Feature vectors for the keypoints after normalization, li1,2, 3.. is some normalized vector;
matching the key points in the binocular image by using Euclidean distance of the feature vectors of the key points as judgment measurement of similarity of the key points in the binocular image, wherein the coordinate information of the key points which are matched with each other is used as a group of key information;
step four, screening the generated matching key points;
calculating the horizontal parallax of each pair of key points to generate a parallax matrix, wherein the parallax matrix is defined as Kn={k1,k2...knN is the logarithm of the match, k1、k2、knAs a single matching point disparity;
determining the median k of the disparity matrixmAnd obtaining a reference parallax matrix, which is marked as Kn', the formula is as follows:
Kn'={k1-km,k2-km,...,kn-km} (17)
setting parallax threshold value to be 3, and setting Kn' deletion of corresponding parallaxes larger than the threshold value in the ' sequence to obtain the final visual matrix result K ', K1'、k2'、kn'All the parallax errors are the parallax errors of the screened correct matching points, n' is the logarithm of the final correct matching, and the formula is as follows:
K'={k1',k2',...,kn'} (18)
substituting the parallax matrix K' obtained in the step four into a binocular ranging model to obtain a significant target distance;
the focal lengths of the two identical imaging systems are separated by J along the horizontal direction, the two optical axes are parallel to the horizontal plane, and the image plane is parallel to the vertical plane;
suppose a target point M (X, Y, Z) in the scene, Pl (X) at the left and right imaging points respectively1,y1) And Pr (x)2,y2),x1,y1And x2,y2Coordinates of Pl and Pr in the vertical plane of the image, respectively, and the parallax in the binocular model is defined as k ═ Pl-Pr | -x2-x1And l, obtaining a distance formula according to a triangular similarity relation, wherein X, Y and Z are coordinates of a horizontal axis, a vertical axis and a longitudinal axis in a space coordinate system:
wherein dx 'represents the physical distance of each pixel in the horizontal axis direction of the imaged negative film, f is the focal length of the imaging system, Z is the distance from the target point M to the connecting line of the two imaging centers, the parallax matrix obtained in the step four is taken into a formula (19), and the corresponding distance matrix Z' ═ Z is obtained according to the physical information of the binocular model1,z2,...,zn'},z1,z2,zn'The distance of the significant target is calculated for single matching parallax, and finally the average value of the distance matrix is calculated, namely the distance Z of the significant target in the binocular imagefThe formula is as follows:
2. the method for measuring the distance between the salient objects in the binocular images according to claim 1, wherein the specific process of edge detection on the images comprises:
performing convolution operation on binocular images by adopting a 2D Gaussian filtering template to eliminate noise interference of the images one by one;
step one, two, respectively calculating gradient amplitude and gradient direction of pixels on the filtered binocular image I (x, y) by utilizing the difference of first-order partial derivatives in the horizontal direction and the vertical direction, wherein the partial derivatives dx and dy in the x direction and the y direction are respectively as follows:
dx=[I(x+1,y)-I(x-1,y)]/2 (21)
dy=[I(x,y+1)-I(x,y-1)]/2 (22)
the gradient magnitude is then:
D'=(dx2+dy2)1/2 (23)
the gradient direction is as follows:
θ'=arctan(dy/dx) (24);
d 'and theta' respectively represent the gradient magnitude and gradient direction of pixels on the filtered binocular image I (x, y);
step three, carrying out non-maximum value inhibition on the gradient, and then carrying out double-threshold processing on the image to generate an edge image; the gray value of the edge point of the edge image is 255, and the gray value of the non-edge point is 0.
3. The method for measuring the distance between the salient objects in the binocular images according to claim 2, wherein the step two of extracting the salient features of the binocular images by using the visual saliency model comprises the following specific steps:
after the first binocular image edge detection, overlapping the original image and the edge image:
I1(σ)=0.7I(σ)+0.3C(σ) (25)
wherein I (sigma) is the original image of the input binocular image, C (sigma) is the edge image, I1(σ) is the image after the superimposition processing;
calculating nine layers of Gaussian pyramids of the images after superposition processing by adopting a Gaussian difference function, wherein the 0 th layer is an input superposition image, the 1 to 8 layers are respectively formed by adopting Gaussian filtering and reduced order sampling on the previous layer, the size of the layer corresponds to 1/2 to 1/256 of the input image, and the brightness, the color and the direction characteristics are extracted for each layer of the Gaussian pyramid to generate a corresponding brightness pyramid, a corresponding color pyramid and a corresponding direction pyramid;
the formula for extracting the brightness features is as follows:
In=(r+g+b)/3 (26)
wherein r, g, b correspond to red, green, blue three components of the input binocular image color, InIs a brightness characteristic;
the formula for extracting the color features is as follows:
R=r-(g+b)/2 (27)
G=g-(r+b)/2 (28)
B=b-(r+g)/2 (29)
Y=r+g-2(|r-g|+b) (30)
r, G, B, Y correspond to the color components of the superimposed image;
o (sigma, omega) is a directional characteristic obtained by filtering and extracting a Gabor function on the brightness characteristic In a scale direction, omega is the direction of the Gabor function, namely the number of Gaussian pyramid layers, and sigma is the total direction number of the Gabor function, wherein sigma belongs to [0,1,2 …,8], omega belongs to [0 °,45 °,90 °,135 ° ];
step two, step three, to solving the three characteristics of luminance, colour and direction of different yardstick of gaussian pyramid carry on the central peripheral contrast to make the difference, specifically:
let the scale c (c ∈ {2,3,4}) be the central scale, and the scale u (u ═ c +, ∈ {3,4}) be the peripheral scale; there are 6 combinations (2-5, 2-6, 3-6, 3-7, 4-7, 4-8) between the central dimension c and the peripheral dimension u in the gaussian pyramid of 9 levels;
the local directional feature contrast, which is the difference between the central and peripheral contrast, is represented by the difference between the feature maps of scale c and scale s as follows:
In(c,u)=|In(c)-In(u)| (31)
RG(c,u)=|(R(c)-G(c))-(G(u)-R(u))| (32)
BY(c,u)=|(B(c)-Y(c))-(Y(u)-B(u))| (33)
O(c,u,ω)=|O(c,ω)-O(u,ω)| (34)
before difference making, interpolation is needed to make the sizes of the two images consistent, and then difference making is carried out;
and step two, four, fusing the feature maps of different features generated by difference through normalization to generate a significant feature map of the input binocular image, specifically comprising the following steps:
for each timeThe scale of each feature is compared with the feature map to carry out normalization fusion to generate a comprehensive feature map of the feature The feature map is normalized for the luminance features,the feature map is normalized for the color features,normalizing the feature map for the directional features; the calculation process is shown in the following formula:
wherein, N (-) represents a normalization calculation function, firstly, for a feature map to be calculated, normalizing the feature value of each pixel in the feature map to a closed region [0,255], then finding out a global maximum significant value A in each normalized feature map, then calculating the average value a of local maximum values in the feature map, and finally multiplying the feature value corresponding to each pixel of the feature by 2 (A-a);
and then, carrying out normalization processing by using the comprehensive characteristic diagram of each characteristic to obtain a final significant characteristic diagram S, wherein the calculation process is as follows:
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510233157.3A CN104778721B (en) | 2015-05-08 | 2015-05-08 | The distance measurement method of conspicuousness target in a kind of binocular image |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201510233157.3A CN104778721B (en) | 2015-05-08 | 2015-05-08 | The distance measurement method of conspicuousness target in a kind of binocular image |
Publications (2)
Publication Number | Publication Date |
---|---|
CN104778721A true CN104778721A (en) | 2015-07-15 |
CN104778721B CN104778721B (en) | 2017-08-11 |
Family
ID=53620167
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201510233157.3A Active CN104778721B (en) | 2015-05-08 | 2015-05-08 | The distance measurement method of conspicuousness target in a kind of binocular image |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN104778721B (en) |
Cited By (23)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105574928A (en) * | 2015-12-11 | 2016-05-11 | 深圳易嘉恩科技有限公司 | Driving image processing method and first electronic equipment |
CN106023198A (en) * | 2016-05-16 | 2016-10-12 | 天津工业大学 | Hessian matrix-based method for extracting aortic dissection of human thoracoabdominal cavity CT image |
CN106094516A (en) * | 2016-06-08 | 2016-11-09 | 南京大学 | A kind of robot self-adapting grasping method based on deeply study |
CN106780476A (en) * | 2016-12-29 | 2017-05-31 | 杭州电子科技大学 | A kind of stereo-picture conspicuousness detection method based on human-eye stereoscopic vision characteristic |
CN106918321A (en) * | 2017-03-30 | 2017-07-04 | 西安邮电大学 | A kind of method found range using object parallax on image |
CN106920244A (en) * | 2017-01-13 | 2017-07-04 | 广州中医药大学 | A kind of method of background dot near detection image edges of regions |
CN107392929A (en) * | 2017-07-17 | 2017-11-24 | 河海大学常州校区 | A kind of intelligent target detection and dimension measurement method based on human vision model |
CN107423739A (en) * | 2016-05-23 | 2017-12-01 | 北京陌上花科技有限公司 | Image characteristic extracting method and device |
CN107633498A (en) * | 2017-09-22 | 2018-01-26 | 成都通甲优博科技有限责任公司 | Image dark-state Enhancement Method, device and electronic equipment |
CN107644398A (en) * | 2017-09-25 | 2018-01-30 | 上海兆芯集成电路有限公司 | Image interpolation method and its associated picture interpolating device |
CN107730521A (en) * | 2017-04-29 | 2018-02-23 | 安徽慧视金瞳科技有限公司 | The quick determination method of roof edge in a kind of image |
CN108036730A (en) * | 2017-12-22 | 2018-05-15 | 福建和盛高科技产业有限公司 | A kind of fire point distance measuring method based on thermal imaging |
CN108460794A (en) * | 2016-12-12 | 2018-08-28 | 南京理工大学 | A kind of infrared well-marked target detection method of binocular solid and system |
CN108665740A (en) * | 2018-04-25 | 2018-10-16 | 衢州职业技术学院 | A kind of classroom instruction control system of feeling and setting happily blended Internet-based |
CN109300154A (en) * | 2018-11-27 | 2019-02-01 | 郑州云海信息技术有限公司 | A kind of distance measuring method and device based on binocular solid |
WO2019029099A1 (en) * | 2017-08-11 | 2019-02-14 | 浙江大学 | Image gradient combined optimization-based binocular visual sense mileage calculating method |
CN110060240A (en) * | 2019-04-09 | 2019-07-26 | 南京链和科技有限公司 | A kind of tyre contour outline measurement method based on camera shooting |
CN110889866A (en) * | 2019-12-04 | 2020-03-17 | 南京美基森信息技术有限公司 | Background updating method for depth map |
CN112489104A (en) * | 2020-12-03 | 2021-03-12 | 海宁奕斯伟集成电路设计有限公司 | Distance measurement method and device, electronic equipment and readable storage medium |
CN112784814A (en) * | 2021-02-10 | 2021-05-11 | 中联重科股份有限公司 | Posture recognition method for vehicle backing and warehousing and conveying vehicle backing and warehousing guide system |
CN116523900A (en) * | 2023-06-19 | 2023-08-01 | 东莞市新通电子设备有限公司 | Hardware processing quality detection method |
CN117152144A (en) * | 2023-10-30 | 2023-12-01 | 潍坊华潍新材料科技有限公司 | Guide roller monitoring method and device based on image processing |
CN117889867A (en) * | 2024-03-18 | 2024-04-16 | 南京师范大学 | Path planning method based on local self-attention moving window algorithm |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110065790B (en) * | 2019-04-25 | 2021-07-06 | 中国矿业大学 | Method for detecting blockage of coal mine belt transfer machine head based on visual algorithm |
Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008040945A1 (en) * | 2006-10-06 | 2008-04-10 | Imperial Innovations Limited | A method of identifying a measure of feature saliency in a sequence of images |
CN103824284A (en) * | 2014-01-26 | 2014-05-28 | 中山大学 | Key frame extraction method based on visual attention model and system |
-
2015
- 2015-05-08 CN CN201510233157.3A patent/CN104778721B/en active Active
Patent Citations (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2008040945A1 (en) * | 2006-10-06 | 2008-04-10 | Imperial Innovations Limited | A method of identifying a measure of feature saliency in a sequence of images |
CN103824284A (en) * | 2014-01-26 | 2014-05-28 | 中山大学 | Key frame extraction method based on visual attention model and system |
Non-Patent Citations (2)
Title |
---|
刘英哲 等: "H.264中一种基于搜索范围自适应调整的运动估计算法", 《电子与信息学报》 * |
蒋寓文 等: "选择性背景优先的显著性检测模型", 《电子与信息学报》 * |
Cited By (35)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105574928A (en) * | 2015-12-11 | 2016-05-11 | 深圳易嘉恩科技有限公司 | Driving image processing method and first electronic equipment |
CN106023198A (en) * | 2016-05-16 | 2016-10-12 | 天津工业大学 | Hessian matrix-based method for extracting aortic dissection of human thoracoabdominal cavity CT image |
CN107423739A (en) * | 2016-05-23 | 2017-12-01 | 北京陌上花科技有限公司 | Image characteristic extracting method and device |
CN107423739B (en) * | 2016-05-23 | 2020-11-13 | 北京陌上花科技有限公司 | Image feature extraction method and device |
CN106094516A (en) * | 2016-06-08 | 2016-11-09 | 南京大学 | A kind of robot self-adapting grasping method based on deeply study |
CN108460794A (en) * | 2016-12-12 | 2018-08-28 | 南京理工大学 | A kind of infrared well-marked target detection method of binocular solid and system |
CN108460794B (en) * | 2016-12-12 | 2021-12-28 | 南京理工大学 | Binocular three-dimensional infrared salient target detection method and system |
CN106780476A (en) * | 2016-12-29 | 2017-05-31 | 杭州电子科技大学 | A kind of stereo-picture conspicuousness detection method based on human-eye stereoscopic vision characteristic |
CN106920244A (en) * | 2017-01-13 | 2017-07-04 | 广州中医药大学 | A kind of method of background dot near detection image edges of regions |
CN106920244B (en) * | 2017-01-13 | 2019-08-02 | 广州中医药大学 | A kind of method of the neighbouring background dot of detection image edges of regions |
CN106918321A (en) * | 2017-03-30 | 2017-07-04 | 西安邮电大学 | A kind of method found range using object parallax on image |
CN107730521A (en) * | 2017-04-29 | 2018-02-23 | 安徽慧视金瞳科技有限公司 | The quick determination method of roof edge in a kind of image |
CN107730521B (en) * | 2017-04-29 | 2020-11-03 | 安徽慧视金瞳科技有限公司 | Method for rapidly detecting ridge type edge in image |
CN107392929A (en) * | 2017-07-17 | 2017-11-24 | 河海大学常州校区 | A kind of intelligent target detection and dimension measurement method based on human vision model |
CN107392929B (en) * | 2017-07-17 | 2020-07-10 | 河海大学常州校区 | Intelligent target detection and size measurement method based on human eye vision model |
WO2019029099A1 (en) * | 2017-08-11 | 2019-02-14 | 浙江大学 | Image gradient combined optimization-based binocular visual sense mileage calculating method |
CN107633498A (en) * | 2017-09-22 | 2018-01-26 | 成都通甲优博科技有限责任公司 | Image dark-state Enhancement Method, device and electronic equipment |
CN107644398A (en) * | 2017-09-25 | 2018-01-30 | 上海兆芯集成电路有限公司 | Image interpolation method and its associated picture interpolating device |
CN108036730A (en) * | 2017-12-22 | 2018-05-15 | 福建和盛高科技产业有限公司 | A kind of fire point distance measuring method based on thermal imaging |
CN108036730B (en) * | 2017-12-22 | 2019-12-10 | 福建和盛高科技产业有限公司 | Fire point distance measuring method based on thermal imaging |
CN108665740A (en) * | 2018-04-25 | 2018-10-16 | 衢州职业技术学院 | A kind of classroom instruction control system of feeling and setting happily blended Internet-based |
CN109300154A (en) * | 2018-11-27 | 2019-02-01 | 郑州云海信息技术有限公司 | A kind of distance measuring method and device based on binocular solid |
CN110060240A (en) * | 2019-04-09 | 2019-07-26 | 南京链和科技有限公司 | A kind of tyre contour outline measurement method based on camera shooting |
CN110060240B (en) * | 2019-04-09 | 2023-08-01 | 南京链和科技有限公司 | Tire contour measurement method based on image pickup |
CN110889866A (en) * | 2019-12-04 | 2020-03-17 | 南京美基森信息技术有限公司 | Background updating method for depth map |
CN112489104A (en) * | 2020-12-03 | 2021-03-12 | 海宁奕斯伟集成电路设计有限公司 | Distance measurement method and device, electronic equipment and readable storage medium |
CN112489104B (en) * | 2020-12-03 | 2024-06-18 | 海宁奕斯伟集成电路设计有限公司 | Ranging method, ranging device, electronic equipment and readable storage medium |
CN112784814A (en) * | 2021-02-10 | 2021-05-11 | 中联重科股份有限公司 | Posture recognition method for vehicle backing and warehousing and conveying vehicle backing and warehousing guide system |
CN112784814B (en) * | 2021-02-10 | 2024-06-07 | 中联重科股份有限公司 | Gesture recognition method for reversing and warehousing of vehicle and reversing and warehousing guiding system of conveying vehicle |
CN116523900A (en) * | 2023-06-19 | 2023-08-01 | 东莞市新通电子设备有限公司 | Hardware processing quality detection method |
CN116523900B (en) * | 2023-06-19 | 2023-09-08 | 东莞市新通电子设备有限公司 | Hardware processing quality detection method |
CN117152144A (en) * | 2023-10-30 | 2023-12-01 | 潍坊华潍新材料科技有限公司 | Guide roller monitoring method and device based on image processing |
CN117152144B (en) * | 2023-10-30 | 2024-01-30 | 潍坊华潍新材料科技有限公司 | Guide roller monitoring method and device based on image processing |
CN117889867A (en) * | 2024-03-18 | 2024-04-16 | 南京师范大学 | Path planning method based on local self-attention moving window algorithm |
CN117889867B (en) * | 2024-03-18 | 2024-05-24 | 南京师范大学 | Path planning method based on local self-attention moving window algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN104778721B (en) | 2017-08-11 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN104778721B (en) | The distance measurement method of conspicuousness target in a kind of binocular image | |
CN110569704B (en) | Multi-strategy self-adaptive lane line detection method based on stereoscopic vision | |
Awrangjeb et al. | Automatic detection of residential buildings using LIDAR data and multispectral imagery | |
EP2811423B1 (en) | Method and apparatus for detecting target | |
CN109598794B (en) | Construction method of three-dimensional GIS dynamic model | |
CN107424142B (en) | Weld joint identification method based on image significance detection | |
CN104835175B (en) | Object detection method in a kind of nuclear environment of view-based access control model attention mechanism | |
CN109584281B (en) | Overlapping particle layering counting method based on color image and depth image | |
CN110675408A (en) | High-resolution image building extraction method and system based on deep learning | |
CN112801074B (en) | Depth map estimation method based on traffic camera | |
CN107818303B (en) | Unmanned aerial vehicle oil and gas pipeline image automatic contrast analysis method, system and software memory | |
CN104517095B (en) | A kind of number of people dividing method based on depth image | |
CN104063702A (en) | Three-dimensional gait recognition based on shielding recovery and partial similarity matching | |
CN107492107B (en) | Object identification and reconstruction method based on plane and space information fusion | |
CN117036641A (en) | Road scene three-dimensional reconstruction and defect detection method based on binocular vision | |
CN105678318B (en) | The matching process and device of traffic sign | |
CN111126393A (en) | Vehicle appearance refitting judgment method and device, computer equipment and storage medium | |
CN106446785A (en) | Passable road detection method based on binocular vision | |
CN116452852A (en) | Automatic generation method of high-precision vector map | |
EP4287137A1 (en) | Method, device, equipment, storage media and system for detecting drivable space of road | |
Yousri et al. | A deep learning-based benchmarking framework for lane segmentation in the complex and dynamic road scenes | |
Zakharov et al. | Automatic building detection from satellite images using spectral graph theory | |
CN118429524A (en) | Binocular stereoscopic vision-based vehicle running environment modeling method and system | |
CN105005962B (en) | Islands and reefs Remote Sensing Image Matching method based on hierarchical screening strategy | |
Ebrahimikia et al. | True orthophoto generation based on unmanned aerial vehicle images using reconstructed edge points |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
EXSB | Decision made by sipo to initiate substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
TA01 | Transfer of patent application right |
Effective date of registration: 20170717 Address after: 510000, Guangdong, Guangzhou, Guangzhou new Guangzhou knowledge city nine Buddha, Jianshe Road 333, room 245 Applicant after: Guangzhou Xiaopeng Automobile Technology Co. Ltd. Address before: 150001 Harbin, Nangang, West District, large straight street, No. 92 Applicant before: Harbin Institute of Technology |
|
TA01 | Transfer of patent application right | ||
GR01 | Patent grant | ||
GR01 | Patent grant |