CN107392929B - Intelligent target detection and size measurement method based on human eye vision model - Google Patents

Intelligent target detection and size measurement method based on human eye vision model Download PDF

Info

Publication number
CN107392929B
CN107392929B CN201710580068.5A CN201710580068A CN107392929B CN 107392929 B CN107392929 B CN 107392929B CN 201710580068 A CN201710580068 A CN 201710580068A CN 107392929 B CN107392929 B CN 107392929B
Authority
CN
China
Prior art keywords
point
points
eye image
image
left eye
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710580068.5A
Other languages
Chinese (zh)
Other versions
CN107392929A (en
Inventor
李庆武
周亚琴
马云鹏
邢俊
许金鑫
席淑雅
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Changzhou Campus of Hohai University
Original Assignee
Changzhou Campus of Hohai University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Changzhou Campus of Hohai University filed Critical Changzhou Campus of Hohai University
Priority to CN201710580068.5A priority Critical patent/CN107392929B/en
Publication of CN107392929A publication Critical patent/CN107392929A/en
Application granted granted Critical
Publication of CN107392929B publication Critical patent/CN107392929B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/13Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/30Determination of transform parameters for the alignment of images, i.e. image registration
    • G06T7/33Determination of transform parameters for the alignment of images, i.e. image registration using feature-based methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20228Disparity calculation for image-based rendering

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an intelligent target detection and size measurement method based on a human eye vision model. The method comprises the steps of combining human eye significance detection with a binocular vision imaging model, detecting significance edges of a target and feature points constrained by spatial information as seed points, completing target detection by utilizing a multi-feature fusion growth mechanism, and finally completing size measurement according to the binocular vision model. The intelligent target detection and size measurement system based on the human eye vision model can accurately complete the detection and size measurement of various targets, the measurement error rate is lower than 2%, and the actual target detection and measurement requirements are met.

Description

Intelligent target detection and size measurement method based on human eye vision model
Technical Field
The invention relates to an intelligent target detection and size measurement method based on a human eye vision model, and belongs to the field of digital image processing, target detection and size measurement.
Background
The size is one of important characteristics of an object, so the size measurement becomes an important technology for production life, such as workpiece size measurement, large-scale cultural building size measurement, vehicle size measurement and the like. Existing dimensional measurement methods can be classified into the following categories: three-coordinate measuring machine, theodolite measuring system, total station measuring system, articulated coordinate measuring machine, indoor Global Positioning System (GPS), laser tracking measuring system, acoustic ranging system, laser ranging system, manual measurement and machine vision measuring system. However, the conventional target dimension measuring method is easily restricted by factors such as the shape of an object and the background. Such as a three-coordinate measuring machine, a total station, a joint type coordinate machine, manual measurement and laser tracking all belong to contact measurement methods, are not suitable for non-contact objects or materials, and have limited application range; the acoustic ranging and the laser ranging are restricted by the measuring range, and the remote measurement is difficult to carry out. Machine vision has the remarkable advantages of non-contact property, high measuring speed, high measuring precision, strong real-time property and the like, and is one of research hotspots in the field of dimension measurement.
The machine vision measuring method can be divided into a plurality of research directions such as monocular vision, binocular vision, multi-ocular vision, infrared vision, ultraviolet vision, mixed vision and the like according to the number and the types of the vision sensors. Among them, monocular vision is often served in various fields as the simplest and fastest image acquisition method, but the obtained visual information has a large limitation. Monocular vision contains a small amount of information, fails to embody three-dimensional information, has a limited visual field range and is easily influenced by the external environment. Compared with monocular vision, the binocular vision technology simulates a human binocular vision system, obtains three-dimensional information of an environment and a target by utilizing parallax information in left and right eye images, and has the following specific advantages: 1) the human binocular vision system is simulated, and the depth information of the target can be acquired on the basis of the two-dimensional information of the target, so that the space positioning and three-dimensional measurement of the target are completed; 2) the human binocular vision system is suitable for obtaining and processing multi-source images through the evolution of 'competitive selection of things and survival of fittest'.
The existing size measuring method using binocular vision has high precision, but is limited to the improvement of measuring precision, and omits a target detection step. In an actual measurement environment, targets are in different backgrounds, and accurate target detection is an important prerequisite for size measurement. Therefore, it is one of the problems to be solved in the present stage to complete target detection before dimension measurement and complete intelligent target detection and dimension measurement.
Disclosure of Invention
The technical problem to be solved by the invention is as follows: the existing dimension measurement method omits an object detection step and has a narrow application range.
In order to solve the technical characteristics, the invention provides an intelligent target detection and size measurement method based on a human eye vision model, which comprises the following steps:
1) acquiring binocular images to be detected: taking a target as an angle of a complete individual closest to the camera, and shooting a group of left eye images and right eye images by using a calibrated binocular camera to serve as images to be measured;
2) extracting effective significant points: extracting a salient point diagram of the left eye image by using a Gaussian difference algorithm, calculating the mean value of all salient points with the salient values larger than 0, taking the mean value as a screening threshold value, and extracting pixel points with the salient values larger than the threshold value as image salient points;
3) extracting a significant edge: acquiring an edge image of the image by using a Canny edge detection algorithm, calculating the number of salient points on each edge, sequencing all edge lines in a descending order according to the number of the salient points, and reserving the first a% of edges as salient edges; a, taking 10;
4) and (3) registering the left eye image and the right eye image: detecting a matching point pair of the binocular image by using a SURF algorithm, selecting the matching point pair according to Euclidean distance minimization between the characteristic points, screening the characteristic points by using slopes between the matching point pairs, removing abnormal point pairs, and reserving effective matching point pairs;
5) extracting feature points restricted by spatial information: calculating parallax information of all effective matching points, sorting the matching points according to a parallax descending order, and extracting the top c% as a characteristic point restricted by spatial information; c, taking 10;
6) target detection: rapidly growing a target area by utilizing a multi-feature fusion method of significant information, spatial information and color information to complete target detection;
7) size measurement using binocular vision model: the method comprises the steps of obtaining two point pairs with the largest distance in the horizontal direction and the vertical direction in a two-dimensional image of a detected target, calculating three-dimensional coordinates of the two point pairs according to binocular parallax of a binocular vision model and internal and external parameters of a camera, calculating Euclidean distances of the two point pairs in the horizontal direction and the vertical direction to serve as the length and the height of the target, and completing size measurement of the target.
Compared with the prior art, the method has the advantages that the method firstly combines human eye vision saliency detection with a binocular vision three-dimensional information perception model, extracts saliency edges and feature points constrained by spatial information as seed points, then completes target detection by utilizing a multi-feature fusion growth mechanism of saliency, spatial information and color information, and finally calculates the size of a target according to the binocular vision model. The automatic detection and the size measurement of various targets are completed, the application range is greatly improved, the measurement error is lower than 2%, and the actual measurement requirement is met.
Drawings
FIG. 1 is a flow chart of an intelligent target detection and size measurement method;
FIG. 2 is a schematic representation of salient edge extraction;
FIG. 3 is a schematic diagram of feature point extraction constrained by spatial information;
FIG. 4 is a schematic diagram of a multi-feature fusion growth mechanism;
FIG. 5 is a growth strategy diagram;
FIG. 6 is a schematic view of an object dimension measurement method;
fig. 7 is a diagram illustrating the measurement result of the target dimension.
Detailed Description
FIG. 1 is a flow chart of the intelligent target detection and size measurement method of the present invention, which is characterized in that a binocular camera is calibrated by using a Zhang calibration method, then a left and right eye image of a target to be measured is shot by using the calibrated binocular camera with the target as an angle of a complete individual closest to the camera, a Gaussian difference algorithm (DoG) and a Canny algorithm are combined to calculate a saliency edge, a modified SURF algorithm is used to match the left and right eye image, and feature points restricted by spatial information are obtained according to parallax. And taking the significant edge and the characteristic points restricted by the spatial information as seed points, quickly detecting the target by using the color similarity, and calculating the size of the target by using a binocular vision model.
The method comprises the following specific steps in sequence:
(1) and shooting a plurality of groups of binocular images by using a chessboard model of 7 × 7, inputting the images into an MAT L AB calibration work box, completing camera calibration and acquiring internal and external parameters of the binocular camera.
(2) And shooting the target by using a calibrated binocular camera according to the angle of the complete individual with the target closest to the camera, acquiring a left eye image and a right eye image of the target, and transmitting the left eye image and the right eye image to a PC (personal computer) through a USB (universal serial bus) interface for displaying and storing.
(3) As shown in fig. 2(b), a gaussian difference saliency map of the left eye image is extracted. The DoG algorithm can achieve the effects of exciting a local central region and inhibiting surrounding neighborhoods, and accords with the visual characteristics of human eyes, so that the significance of an image can be reflected to a certain degree. The initial gaussian difference saliency map for the left eye image is calculated as follows:
Figure GDA0002453150560000041
wherein σ1And σ2Respectively representing excitation and suppression bandwidths, the value σ being taken in the text1=0.6,σ20.9, I is a grayscale image, symbol
Figure GDA0002453150560000044
Representing that sliding frequency filtering is carried out on the image, and the DoG (x, y) is the significance metric value of the obtained pixel point (x, y);
setting a negative value generated by the saliency metric DoG (x, y) as 0, and setting a saliency mean value as a threshold value T for extraction of a saliency point to obtain a saliency point diagram D (x, y):
Figure GDA0002453150560000042
Figure GDA0002453150560000043
wherein, count (DoG > 0) represents the number of significant points with significant value greater than 0 in DoG, and sum (DoG > 0) represents the sum of significant values greater than 0.
(4) As shown in fig. 2(c), the Canny edge detection algorithm can detect all continuous edges of the image, but cannot highlight salient objects, especially the background and shadow edges interfere with the object detection greatly. In order to ensure the accuracy of target detection, the saliency of the DoG is used for constraining the Canny edge, and the saliency edge of the target is extracted, as shown in fig. 2 (d). The method for obtaining the significant edge comprises the following steps:
a. screening effective edges: the Canny detection result has a plurality of short edges caused by shadow, reflection and other factors, which are not beneficial to the extraction of the target edge, and the extraction of the long edge as the effective edge is beneficial to removing the false detection edge, refining the edge information and reducing the calculated amount, so that the lengths of all edges are sequenced, and 50 percent of the edge boundary with the longest length is reserved as the long edge;
b. calculating the edge significance: counting the number of significant points on each effective boundary as a standard for measuring the edge significance, wherein the number of significant points is in direct proportion to the edge significance;
c. extracting a significant edge: sorting the edges in descending order according to the number of the significant points, and reserving the edge line of the first a% as a significant edge;
(5) the method adopts an improved SURF algorithm to register the left eye image and the right eye image, accurately completes the matching of the random feature points of the left eye and the random feature points of the right eye, uses the SURF algorithm to detect the feature points of the images in a scale space, selects windows with different sizes when filtering the images, extracts the feature points through a Hessian matrix, and defines the Hessian matrix as follows:
Figure GDA0002453150560000051
where σ is the scale of the Gaussian kernel function, Lxx(x, σ) is a second order differential of Gaussian
Figure GDA0002453150560000052
Convolution with the image I (x, y) in the x-direction, Lyy(x, σ) is a second order differential of Gaussian
Figure GDA0002453150560000053
Convolution with the image I (x, y) in the y direction,
Figure GDA0002453150560000054
convolution with image I (x, y) in the xy direction, the formula for which is as follows:
Figure GDA0002453150560000055
g (σ) is a Gaussian kernel function.
a. Selecting a main direction of the characteristic points: in order to ensure the direction invariance of the algorithm, the SURF algorithm assigns a unique main direction for the feature points according to the information of surrounding pixel points. First, with the detected feature point as the center, σ as the step size, and the size of the wavelet as 4 σ, the response of the Harr wavelet in the horizontal and vertical directions in a circular neighborhood with a radius of 6 σ is calculated, where σ represents the scale space in which it is located. And performing Gaussian weighting on the Harr wavelet response taking the current characteristic point as the center by using a Gaussian function with the scale of 2 sigma. The smaller the distance between the specification and the characteristic point is, the smaller the weight of the pixel point is, and a new Harr wavelet response value is obtained. Constructing a fan-shaped window with an angle of 60 degrees by the SURF algorithm, calculating a Harr wavelet response value in the window, continuously rotating and traversing the whole circular region by using the window until the Harr wavelet response in the window reaches the strongest value, and defining the direction pointed by the fan-shaped window at the moment as the main direction of the characteristic point;
b. constructing a block with the size of 20 sigma along the main direction of the current feature point by taking the current feature point as the center, then dividing the region into 16 small regions, and calculating the haar wavelet response in each region with the size of 5 sigma × 5 sigma, wherein each sub-region v is expressed as v (∑ dx, ∑ | dx |, ∑ dy, ∑ | dy |), and finally obtaining a 64-dimensional descriptor of the point, and the feature points of the left-eye image and the right-eye image can be defined as:
Figure GDA0002453150560000061
wherein Pos1 represents the feature point parameter of the left eye image, Pos2 represents the feature point parameter of the right eye image, m and n are the number of feature points in the left eye image and the right eye image respectively, i and j represent the subscripts of the feature points of the left eye image and the right eye image respectively, (x)m',ym') denotes the mth feature point coordinate of the left eye image; (x)n,yn) Representing the coordinates of the nth characteristic point of the right eye image;
c. calculating Euclidean distances of all points in feature point parameters Pos1 and Pos2 of the left eye image and the right eye image, selecting a point with the minimum Euclidean distance as a rough matching point pair, sorting the rough matching point pair according to ascending Euclidean distances, deleting abnormal points, selecting the first K matching point pairs, and defining as:
Pos_K={{(x'1,y'1),(x1,y1)},{(x'2,y'2),(x2,y2)},...,{(x'i,y'i),(xi,yi)},...{(x'K,y'K),(xK,yK)}},1≤i≤K;
d. screening the matched point pairs according to the slopes of the corresponding points in the K matched point pairs Pos _ K, calculating the slopes of all coarse matched point pairs, and calculating the slope of 10-2Keeping all slope values in the order of magnitude, calculating the occurrence frequency of all slopes, selecting the slope with the maximum occurrence frequency as a leading slope, deleting the matching point pairs corresponding to other abnormal slopes, and updating to obtain the H groups of accurate matching point pairs Pos _ KnewThe formula is as follows:
Figure GDA0002453150560000062
(xzi,yzi),(xyi,yyi) Respectively represent a group of matching points to center the left eyeFeature point coordinates of the image and the right eye image.
(6) Extracting the feature points restricted by the spatial information, directly calculating and obtaining the parallaxes of all the matching points through SURF algorithm matching, sequencing the parallaxes in an ascending order, and reserving the first c% matching points with the maximum parallaxes as the feature points restricted by the spatial information. As shown in fig. 3, the white automobile model is a complete target in the image with a space closest to the camera, and the extracted feature points constrained by the space information are marked in the left and right target images. All the extracted feature points in the graph are on the target, so that the probability of segmentation errors is reduced, and generally c is 10.
(7) As shown in fig. 4, a complete target cannot be obtained only from the significant edge, but the significant edge and the feature points constrained by the spatial information can accurately represent the features of the target, and the selection requirement of the growing seed points is being satisfied. Therefore, the invention fuses the obvious edge information, the space information and the color information to carry out rapid growth and finish the target detection. Fig. 5 simulates the process of seed point growth, and a repeated growth area exists in the growth process, that is, two seed points simultaneously determine the area as the area to be grown, and repeated operation reduces the growth efficiency.
In the growth process of the seed set, the similarity judgment is a very important growth or stop index, in order to quantitatively calculate the similarity between the seed set and the region to be grown, a similarity function is selected to calculate the similarity, and if the similarity between the region to be grown and the seed set is greater than a set threshold value, the growth is carried out. Because the HSI color model accords with the visual characteristics of human eyes, in order to reduce the influence of brightness on target detection, the method adopts the HSI color space to calculate the similarity.
Taking the significant edge and the feature point restricted by the spatial information as growth seed points, taking all the growth seed points as initial seed areas, converting the left eye image of the image into an HSI color model, respectively calculating the mean component mean square deviations of the hue (H), the saturation (S) and the bright spot (I) of the area to be grown and the seed areas in the HSI color information, carrying out weighted summation on the mean square deviations, taking the weighted summation value as the difference value between the area to be grown and the seed areas, growing the point if the difference value is lower than a threshold value, combining the point into the corresponding seed areas, updating to obtain new seed areas, and continuously circulating until all the seed areas have no pixels meeting the conditions or reach the edge of the image, and stopping growing.
Taking the significant edge and the characteristic points constrained by the spatial information as initial seed regions, and counting the number of the initial seed points to be N, RiThen it is the ith seed region, i ∈ [1, N],
Figure GDA0002453150560000071
Is a seed region RiAnd each seed region grows according to the similarity with the region to be grown, newly grown pixel points are continuously combined into the corresponding seed region to be used as a new seed region for the next growth, and the growth is stopped until all the seed regions do not meet the condition of the region to be grown.
Region to be grown
Figure GDA0002453150560000073
With the seed region RiSimilarity function of
Figure GDA0002453150560000072
Is defined as:
Figure GDA0002453150560000081
Figure GDA0002453150560000082
k is the number of pixel points in each seed region, the number of the pixel points is continuously increased along with the growth of the seed region, j represents the jth pixel point in the growth region, and Ht、St、ItRespectively the hue, saturation and brightness value in the region to be seeded,
Figure GDA0002453150560000083
respectively being growth areas
Figure GDA0002453150560000085
Internal hue, saturation, and brightness mean;12and3for adjusting the weights of the hue component, saturation and brightness components. Since the mean square error can reflect the difference, the larger the difference is, the larger the influence degree of the component on the image is, so the mean square error of the H, S, I component of the image is used as the mean square error of the image12And3the accuracy of target area detection can be improved.
(8) Given that the base-line distance between the binocular vision cameras is b, the focal length of the cameras is f, and the parallax is represented by d, assuming that the left eye image and the right eye image are already registered, the parallax can be represented by the position difference of the same point in the image pair as d ═ x (x)l-xr)。
Wherein x isl、xrThe abscissa of the matched point in the left eye image and the right eye image respectively, and the spatial coordinate (x) of a certain point P in the left camera coordinate system can be calculated according to the following formulac,yc,zc),(xc,yc,zc) For spatial coordinates of the matching points in the world coordinate system, (x)l,yl) Is the two-dimensional image coordinates of the matching point in the left eye image.
Figure GDA0002453150560000084
Calculating the point pair with the maximum distance between the segmented target in the horizontal and vertical directions on the two-dimensional plane image, such as L in the horizontal direction in FIG. 61And L2H in the vertical direction1And H2And obtaining space coordinates of four points according to a binocular vision principle, and calculating the maximum Euclidean distance in the horizontal direction and the vertical direction as the horizontal width and the vertical height of the target by using the three-dimensional coordinates.
Suppose that these four three-dimensional coordinates are L respectively1(xL1,yL1,zL1)、L2(xL2,yL2,zL2) And H1(xH1,yH1,zH1)、H2(xH2,yH2,zH2) The maximum length and height of the object can be calculated according to the following formula:
Figure GDA0002453150560000091
as shown in FIG. 7, the present invention completes the detection and size measurement of the object, and calculates the length and height of the object with an accuracy of 0.1mm, and displays them on the image.
The present invention has been disclosed in terms of the preferred embodiment, but it is not intended to be limited to the embodiment, and all technical solutions obtained by substituting or converting the equivalent embodiments fall within the scope of the present invention.

Claims (8)

1. An intelligent target detection and size measurement method based on a human eye vision model is characterized by comprising the following steps:
1) acquiring binocular images to be detected: taking a target as an angle of a complete individual closest to the camera, and shooting a group of left eye images and right eye images by using a calibrated binocular camera to serve as images to be measured;
2) extracting effective significant points: extracting a salient point diagram of the left eye image by using a Gaussian difference algorithm, calculating the mean value of all salient points with the salient values larger than 0, taking the mean value as a screening threshold value, and extracting pixel points with the salient values larger than the threshold value as image salient points;
3) extracting a significant edge: acquiring an edge image of the image by using a Canny edge detection algorithm, calculating the number of salient points on each edge, sequencing all edge lines in a descending order according to the number of the salient points, and reserving the first a% of edges as salient edges;
4) and (3) registering the left eye image and the right eye image: detecting a matching point pair of the binocular image by using a SURF algorithm, selecting the matching point pair according to Euclidean distance minimization between the characteristic points, screening the characteristic points by using slopes between the matching point pairs, removing abnormal point pairs, and reserving effective matching point pairs;
5) extracting feature points restricted by spatial information: calculating parallax information of all effective matching points, sorting the matching points according to a parallax descending order, and extracting the top c% as a characteristic point restricted by spatial information;
6) target detection: rapidly growing a target area by utilizing a multi-feature fusion method of significant information, spatial information and color information to complete target detection;
7) size measurement using binocular vision model: the method comprises the steps of obtaining two point pairs with the largest distance in the horizontal direction and the vertical direction in a two-dimensional image of a detected target, calculating three-dimensional coordinates of the two point pairs according to binocular parallax of a binocular vision model and internal and external parameters of a camera, calculating Euclidean distances of the two point pairs in the horizontal direction and the vertical direction to serve as the length and the height of the target, and completing size measurement of the target.
2. The intelligent target detection and size measurement method based on human eye vision model as claimed in claim 1, wherein in the step 2), an initial gaussian difference saliency map of the left eye image is calculated as follows:
Figure FDA0002453150550000011
wherein σ1And σ2Respectively representing excitation bandwidth and suppression bandwidth, I is gray image, symbol
Figure FDA0002453150550000021
Representing that sliding frequency filtering is carried out on the image, and the DoG (x, y) is the significance metric value of the obtained pixel point (x, y);
setting a negative value generated by the saliency metric DoG (x, y) as 0, and setting a saliency mean value as a threshold value T for extraction of a saliency point to obtain a saliency point diagram D (x, y):
Figure FDA0002453150550000022
Figure FDA0002453150550000023
wherein, count (DoG > 0) represents the number of significant points with significant value greater than 0 in DoG, and sum (DoG > 0) represents the sum of significant values greater than 0.
3. The intelligent object detection and dimension measurement method based on human eye vision model as claimed in claim 1, wherein in the step 3), a-10.
4. The intelligent target detection and dimension measurement method based on human eye vision model as claimed in claim 1, wherein in the step 4), the following steps are included:
a. obtaining feature point coordinates of the left eye image and the right eye image and 64-dimensional descriptors of all feature points by using an SURF algorithm, wherein the feature points of the left eye image and the right eye image are defined as follows:
Figure FDA0002453150550000024
wherein Pos1 represents the feature point parameter of the left eye image, Pos2 represents the feature point parameter of the right eye image, m and n are the number of feature points in the left eye image and the right eye image respectively, i and j represent the subscripts of the feature points of the left eye image and the right eye image respectively, (x)m',ym') denotes the mth feature point coordinate of the left eye image; (x)n,yn) Representing the coordinates of the nth characteristic point of the right eye image;
b. calculating Euclidean distances of all points in feature point parameters Pos1 and Pos2 of the left eye image and the right eye image, selecting a point with the minimum Euclidean distance as a rough matching point pair, sorting the rough matching point pair according to ascending Euclidean distances, deleting abnormal points, selecting the first K matching point pairs, and defining as:
Pos_K={{(x′1,y′1),(x1,y1)},{(x′2,y′2),(x2,y2)},...,{(x′i,y′i),(xi,yi)},...{(x′K,y′K),(xK,yK)}},1≤i≤K;
c. screening the matched point pairs according to the slopes of the corresponding points in the K matched point pairs Pos _ K, calculating the slopes of all coarse matched point pairs, and calculating the slope of 10-2Keeping all slope values in the order of magnitude, calculating the occurrence frequency of all slopes, selecting the slope with the maximum occurrence frequency as a leading slope, deleting the matching point pairs corresponding to other abnormal slopes, and updating to obtain the H groups of accurate matching point pairs Pos _ KnewThe formula is as follows:
Figure FDA0002453150550000031
(xzi,yzi),(xyi,yyi) Respectively representing the coordinates of the characteristic points of the left eye image and the right eye image in a group of matching point pairs.
5. The intelligent object detection and dimension measurement method based on human eye vision model as claimed in claim 1, wherein in the step 5), c-10.
6. The intelligent target detection and dimension measurement method based on human eye vision model as claimed in claim 1, characterized in that in the step 6), the salient edges and the feature points restricted by the spatial information are used as growth seed points, all the growing seed points are used as initial seed areas, the left eye image is converted into an HSI color model, the hue, saturation and mean square deviation of the bright point mean value components of the area to be grown and the seed areas in HSI color information are respectively calculated, and the mean square error is weighted and summed, the weighted and summed value is used as the difference value between the area to be grown and the seed area, if the difference value is lower than the threshold value, growing the point, combining the point into the corresponding seed region, updating to obtain a new seed region, and continuously circulating until all seed regions have no qualified pixels or reach the image edge, and stopping growing.
7. The intelligent target detection and size measurement method based on the human eye vision model as claimed in claim 6, wherein a repeated growth area exists in the growth process, that is, two seed points simultaneously judge a certain area as an area to be grown, the repeated operation reduces the growth efficiency, if a certain pixel point is processed by the previous seed point, the pixel point is marked with a 'processed pixel point label', and the next seed point does not repeatedly judge the area.
8. The method as claimed in claim 1, wherein in the step 7), the base distance between the binocular vision cameras is b, the focal length of the cameras is f, the parallax is d, and the parallax is represented by d (x ═ x) which is the difference between the positions of the same point in the image point pairl-xr) Wherein x isl、xrRespectively, the abscissa of the matched point in the left eye image and the right eye image, and the spatial coordinate (x) of a certain point P in the left camera coordinate system is calculated according to the following formulac,yc,zc),(xc,yc,zc) For spatial coordinates of the matching points in the world coordinate system, (x)l,yl) Two-dimensional image coordinates of the matching points in the left eye image;
Figure FDA0002453150550000041
calculating the point pairs with the maximum distance of the segmented target in the horizontal and vertical directions on the two-dimensional plane image, obtaining the space coordinates of four points according to the binocular vision principle, and calculating the maximum Euclidean distance in the horizontal and vertical directions by using the three-dimensional coordinates as the horizontal width and the vertical height of the target;
suppose that the four-point three-dimensional coordinates are L respectively1(xL1,yL1,zL1)、L2(xL2,yL2,zL2) And H1(xH1,yH1,zH1)、H2(xH2,yH2,zH2) The maximum length and height of the object are calculated as follows:
Figure FDA0002453150550000042
CN201710580068.5A 2017-07-17 2017-07-17 Intelligent target detection and size measurement method based on human eye vision model Active CN107392929B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710580068.5A CN107392929B (en) 2017-07-17 2017-07-17 Intelligent target detection and size measurement method based on human eye vision model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710580068.5A CN107392929B (en) 2017-07-17 2017-07-17 Intelligent target detection and size measurement method based on human eye vision model

Publications (2)

Publication Number Publication Date
CN107392929A CN107392929A (en) 2017-11-24
CN107392929B true CN107392929B (en) 2020-07-10

Family

ID=60339282

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710580068.5A Active CN107392929B (en) 2017-07-17 2017-07-17 Intelligent target detection and size measurement method based on human eye vision model

Country Status (1)

Country Link
CN (1) CN107392929B (en)

Families Citing this family (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107950402B (en) * 2017-11-29 2020-10-30 北京伟景智能科技有限公司 Automatic control method of milking device based on binocular vision
CN108381549B (en) * 2018-01-26 2021-12-14 广东三三智能科技有限公司 Binocular vision guide robot rapid grabbing method and device and storage medium
CN109211198B (en) * 2018-08-15 2021-01-01 河海大学常州校区 Intelligent target detection and measurement system and method based on trinocular vision
CN109166153A (en) * 2018-08-21 2019-01-08 江苏德丰建设集团有限公司 Tower crane high altitude operation 3-D positioning method and positioning system based on binocular vision
CN109191533B (en) * 2018-08-21 2021-06-25 江苏德丰建设集团有限公司 Tower crane high-altitude construction method based on fabricated building
CN109242776B (en) * 2018-09-11 2023-04-07 江苏君英天达人工智能研究院有限公司 Double-lane line detection method based on visual system
CN111882621A (en) * 2020-07-22 2020-11-03 武汉大学 Rice thickness parameter automatic measurement method based on binocular image
CN113989489A (en) * 2021-10-22 2022-01-28 珠海格力电器股份有限公司 Distance detection method and system for non-smooth edge
CN114549449B (en) * 2022-02-17 2023-05-12 中国空气动力研究与发展中心超高速空气动力研究所 Fine quantitative identification method for global defects of small-size curved surface component

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003098922A1 (en) * 2002-05-15 2003-11-27 The Board Of Governors For Higher Education, State Of Rhode Island And Providence Plantations An imaging system and method for tracking the motion of an object
CN104778721A (en) * 2015-05-08 2015-07-15 哈尔滨工业大学 Distance measuring method of significant target in binocular image
CN105894574A (en) * 2016-03-30 2016-08-24 清华大学深圳研究生院 Binocular three-dimensional reconstruction method
CN106709499A (en) * 2017-03-02 2017-05-24 西北工业大学 SIFT image feature point extraction method based on Canny operator and Hilbert-Huang transform

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2003098922A1 (en) * 2002-05-15 2003-11-27 The Board Of Governors For Higher Education, State Of Rhode Island And Providence Plantations An imaging system and method for tracking the motion of an object
CN104778721A (en) * 2015-05-08 2015-07-15 哈尔滨工业大学 Distance measuring method of significant target in binocular image
CN105894574A (en) * 2016-03-30 2016-08-24 清华大学深圳研究生院 Binocular three-dimensional reconstruction method
CN106709499A (en) * 2017-03-02 2017-05-24 西北工业大学 SIFT image feature point extraction method based on Canny operator and Hilbert-Huang transform

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
salient object detection:a benchmark;Borji A et al;《IEEE Transactions on Image Pricessing》;20151007;第24卷(第12期);第5706-5722页 *
salient region detection based on binocular vision;Zhong Liu et al;《2012 7th IEEE Conference on Industrial Electronics and Applications》;20120720;第1862-1866页 *
一种改进的Canny边缘检测算法;吉玲等;《微处理机》;20150228(第01期);第40-43页 *
基于双目视觉的测距方法研究与应用;王代东;《中国优秀硕士学位论文全文数据库 信息科技辑》;20170515(第05期);第I138-771页 *

Also Published As

Publication number Publication date
CN107392929A (en) 2017-11-24

Similar Documents

Publication Publication Date Title
CN107392929B (en) Intelligent target detection and size measurement method based on human eye vision model
CN110781827B (en) Road edge detection system and method based on laser radar and fan-shaped space division
CN106650640B (en) Negative obstacle detection method based on laser radar point cloud local structure characteristics
CN106651752B (en) Three-dimensional point cloud data registration method and splicing method
CN109211198B (en) Intelligent target detection and measurement system and method based on trinocular vision
CN107248159A (en) A kind of metal works defect inspection method based on binocular vision
CN111340797A (en) Laser radar and binocular camera data fusion detection method and system
CN110647850A (en) Automatic lane deviation measuring method based on inverse perspective principle
CN106969706A (en) Workpiece sensing and three-dimension measuring system and detection method based on binocular stereo vision
CN106251353A (en) Weak texture workpiece and the recognition detection method and system of three-dimensional pose thereof
CN106595500A (en) Transmission line ice coating thickness measurement method based on unmanned aerial vehicle binocular vision
CN105865344A (en) Workpiece dimension measuring method and device based on machine vision
CN110334678A (en) A kind of pedestrian detection method of view-based access control model fusion
CN113470090A (en) Multi-solid-state laser radar external reference calibration method based on SIFT-SHOT characteristics
CN111046843A (en) Monocular distance measurement method under intelligent driving environment
CN113205604A (en) Feasible region detection method based on camera and laser radar
CN107392953B (en) Depth image identification method based on contour line
CN113221648A (en) Fusion point cloud sequence image guideboard detection method based on mobile measurement system
CN116486287A (en) Target detection method and system based on environment self-adaptive robot vision system
CN106709432B (en) Human head detection counting method based on binocular stereo vision
CN116079749A (en) Robot vision obstacle avoidance method based on cluster separation conditional random field and robot
CN116703895B (en) Small sample 3D visual detection method and system based on generation countermeasure network
CN108388854A (en) A kind of localization method based on improvement FAST-SURF algorithms
CN117496401A (en) Full-automatic identification and tracking method for oval target points of video measurement image sequences
CN105718929B (en) The quick round object localization method of high-precision and system under round-the-clock circumstances not known

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant