CN109255818A - A kind of extracting method of novel target and its sub-pixel angle point - Google Patents
A kind of extracting method of novel target and its sub-pixel angle point Download PDFInfo
- Publication number
- CN109255818A CN109255818A CN201810918877.7A CN201810918877A CN109255818A CN 109255818 A CN109255818 A CN 109255818A CN 201810918877 A CN201810918877 A CN 201810918877A CN 109255818 A CN109255818 A CN 109255818A
- Authority
- CN
- China
- Prior art keywords
- camera
- angle point
- pixel
- target
- angle
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 46
- 238000001514 detection method Methods 0.000 claims abstract description 17
- 229910000831 Steel Inorganic materials 0.000 claims abstract description 8
- 239000010959 steel Substances 0.000 claims abstract description 8
- 238000000605 extraction Methods 0.000 abstract description 20
- 239000000284 extract Substances 0.000 abstract description 9
- 230000003287 optical effect Effects 0.000 abstract description 8
- 230000000694 effects Effects 0.000 abstract description 7
- 238000003384 imaging method Methods 0.000 description 24
- 238000012937 correction Methods 0.000 description 10
- 238000005259 measurement Methods 0.000 description 9
- 244000062793 Sorghum vulgare Species 0.000 description 8
- 235000019713 millet Nutrition 0.000 description 8
- 238000011160 research Methods 0.000 description 5
- 238000004364 calculation method Methods 0.000 description 4
- 238000012886 linear function Methods 0.000 description 4
- 238000010606 normalization Methods 0.000 description 4
- 238000010586 diagram Methods 0.000 description 3
- 238000001914 filtration Methods 0.000 description 3
- 239000000463 material Substances 0.000 description 3
- 238000013178 mathematical model Methods 0.000 description 3
- 239000011159 matrix material Substances 0.000 description 3
- NAWXUBYGYWOOIX-SFHVURJKSA-N (2s)-2-[[4-[2-(2,4-diaminoquinazolin-6-yl)ethyl]benzoyl]amino]-4-methylidenepentanedioic acid Chemical compound C1=CC2=NC(N)=NC(N)=C2C=C1CCC1=CC=C(C(=O)N[C@@H](CC(=C)C(O)=O)C(O)=O)C=C1 NAWXUBYGYWOOIX-SFHVURJKSA-N 0.000 description 2
- 241000208340 Araliaceae Species 0.000 description 2
- 235000005035 Panax pseudoginseng ssp. pseudoginseng Nutrition 0.000 description 2
- 235000003140 Panax quinquefolius Nutrition 0.000 description 2
- 238000010220 Pearson correlation analysis Methods 0.000 description 2
- 238000006243 chemical reaction Methods 0.000 description 2
- 235000008434 ginseng Nutrition 0.000 description 2
- 239000003550 marker Substances 0.000 description 2
- 238000012360 testing method Methods 0.000 description 2
- 241000406668 Loxodonta cyclotis Species 0.000 description 1
- 230000002159 abnormal effect Effects 0.000 description 1
- 238000004458 analytical method Methods 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000002860 competitive effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000013461 design Methods 0.000 description 1
- 238000003708 edge detection Methods 0.000 description 1
- 235000013399 edible fruits Nutrition 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000002474 experimental method Methods 0.000 description 1
- 230000005484 gravity Effects 0.000 description 1
- 238000013383 initial experiment Methods 0.000 description 1
- 238000007689 inspection Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000004519 manufacturing process Methods 0.000 description 1
- 238000013507 mapping Methods 0.000 description 1
- 230000000877 morphologic effect Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
- 230000000007 visual effect Effects 0.000 description 1
- 230000003245 working effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/80—Analysis of captured images to determine intrinsic or extrinsic camera parameters, i.e. camera calibration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/80—Geometric correction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20164—Salient point detection; Corner detection
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Length Measuring Devices By Optical Means (AREA)
Abstract
The invention discloses the extracting methods of a kind of novel target and its sub-pixel angle point, the novel target includes rectangle black block and the gridiron pattern target that rectangular white block is staggeredly stacked, the width of each rectangle black block and rectangular white block is equal, target is horizontal positioned, and latter row's length is Δ d than previous row's length incrementiAre as follows: d*f ' (x) mm.The extracting method, two different angle point templates are defined using the checkerboard angle point detection process based on growth, sub-pixel Corner character is carried out with the cornerSubPix () function in OpenCV, it is grown according to energy function and rebuilds gridiron pattern, angle point is marked, sub-pixel angular coordinate is exported.The present invention accurate can extract optical axis direction angle point, and reasonably evade influence caused by perspective transform angle steel joint extraction accuracy;The extracting method of the sub-pixel angle point, it is preferable to the biggish picture extraction effect of distortion degree.
Description
Technical field
The present invention relates to ground close-range photogrammetry field, especially a kind of single camera vision system lower pinhole camera it is novel
The extracting method of target and its sub-pixel angle point.
Background technique
Object ranging based on image is current research hotspot, and many fields industrially have obtained widely
Using, such as forestry survey, automatic Pilot.Object ranging based on image, is broadly divided into initiative range measurement and passive ranging
Two methods[1].Initiative range measurement is to carry out ranging using laser radar[2-4].Passive ranging is by machine vision to two-dimensional digital
Object depth information is calculated in image, then according to image pixel information and camera imaging principle calculate object away from
From[5-6].It not only can replace many manual workings using machine vision, improve production automation level, improve detection accuracy, more
It is effective solution route that many general measuring methods cannot achieve.Machine vision ranging is broadly divided into monocular vision ranging, double
Two class of mesh visual token [1-9].The depth information acquisition method of early stage is mainly binocular stereo vision and camera motion information, is needed
Multiple image is wanted to complete the acquisition of image depth information[10-16].Compared with binocular distance measurement, the acquisition of monocular range images is not required to
Want stringent hardware condition, more competitive superiority.The geometric position of the picture point on image obtained due to camera and true generation
Corresponded in boundary the geometric position of object point be it is closely related, these positions and its correlation are then the geometry moulds by camera imaging
What type determined, therefore once can determine the parameter of this geometrical model, we can completely express two-dimentional picture point and three
The corresponding relationship between object point is tieed up, to calculate image object object distance.In the prior art, the object of single camera vision system
There are many methods for Depth Information Acquistion.The depth information of target to be measured is such as obtained using corresponding points standardization[17-19].Document
[17] a kind of robot target positioning distance measuring method based on monocular vision is had studied, this method is usually to obtain by camera calibration
The inside and outside parameter of camera is taken, the transformational relation between image coordinate system and world coordinate system is solved in conjunction with projection model, to count
Calculate object depth information.Unfortunately, the method needs to acquire the target image of different direction, and accurately records each point and exist
Respective coordinates in world coordinate system and image coordinate system, stated accuracy are affected for measurement accuracy.Further, it is also possible to logical
The relationship building depth extraction model in research world coordinate system between object actual imaging and image pixel is crossed, image is calculated
Middle object is at a distance from actual scene between camera.Document [20] puts object of reference on road surface and measures its distance, choosing
Suitable mathematical model is selected, the corresponding relationship being fitted between object of reference distance and pixel recycles this relationship extract real-time depth
Information.Unfortunately, the method precision of document [20] will receive the influence of telemeasurement error and error of fitting.Document [21]
A kind of vertical target image is devised, the angle point data by detecting the image establish image ordinate pixel value and actual measurement
Mapping relations between angle combine known vehicle-mounted monocular camera height to obtain vehicle-mounted depth in image and believe using this relationship
Breath.Since different cameral equipment inner parameter has differences, for the camera apparatus of different model, this method needs are resurveyed
Target image information, establishes camera depth information extraction model, and different in-vehicle cameras camera lens make with assemble etc. due to,
So that camera pitch angle can also have differences, therefore the method versatility of document [21] is poor.In addition, the method for document [21] is adopted
It is applied to horizontal plane with the relationship between vertical target research perpendicular picture point imaging angle and ordinate pixel value, and by this
The measurement of upper object distance, so that range accuracy is relatively low, because of camera level and the incomplete phase of vertical direction Distortion Law
Together.The selection of target is most important to the building of depth extraction model, however, real using traditional gridiron pattern target research object point
It is existing by perspective transform due to its as broad as long characteristic when relationship between border imaging angle and corresponding picture point ordinate pixel
So that models fitting effect is poor, measurement accuracy is not high for the influence of elephant.
It can make model that there is equipment interoperability in addition, constructing depth extraction model by the method for camera calibration, it should
Model needs to obtain camera internal parameter, the stated accuracy of camera, the object ranging to image by the method for camera calibration
Precision has vital effect.Zhang Zhengyou calibration method is most common camera calibration method, using tradition
Black and white square interlock stacked gridiron pattern target.This target can only be used to realize the calibration of camera in same plane, if will
The horizontal positioned then stated accuracy of target can be reduced accordingly.Application No. is 201710849961.3 patent applications, disclose one kind
Improved camera calibration model and distortion correction model suitable for intelligent sliding moved end camera is (hereinafter referred to as: improved with non-
The peg model of linear distortion item), this method can help to correct scaling board picture, obtain the inside and outside ginseng of camera of higher precision
Number, but the distortion correction model is only used for the nonlinear distortion of correction camera, the perspective phenomenon not being suitable in imaging.
Bibliography:
[1] He Ruofei, Tian Xuetao, Liu Hongjuan wait unmanned plane target localization method of the based on Monte Carlo Kalman filtering
[J] Northwestern Polytechnical University journal, 2017,35 (3): 435-441.
[2] Lin F, Dong X, Chen B M, et al.A Robust Real-Time Embedded Vision
System on an Unmanned Rotorcraft for Ground Target Following[J].IEEE Trans on
Industrial Electronics, 2012,59 (2): 1038-1049.
[3] Zhang Wanlin, Hu Zhengliang, Zhu Jianjun wait one of the comprehensive view instrument of individual soldier target position calculation method [J]
Electronic measurement technique, 2014,37 (11): 1-3.
[4] Sun Junling, Sun Guangmin, Ma Pengge wait laser eyepiece of the based on symmetrical wavelet noise reduction and asymmetric Gauss curve fitting
Position [J] Chinese laser, 2017,44 (6): 178-185.
[5] Shi Jie, Li Yin child, Qi Guoqing, passive tracking algorithm [J] China under waiting not exclusively to measure based on machine vision
Middle University of Science and Technology's journal, 2017,45 (6): 33-37.
[6] Xu Cheng, yellow grand celebration, a kind of passive target positioning of the numerous small drone of clanging or clanking sound in hole and precision analysis [J] instrument instrument
Table journal, 2015,36 (5): 1115-1122.
[7] Li Kehong, Jiang Lingmin, Gong Yong justice .2 are tieed up to 3 d image/Video Quality Metric image depth extracting method and are summarized
[J] Journal of Image and Graphics, 2014,19 (10): 1393-1406.
[8] Wang Hao, Xu Zhiwen, Xie Kun, wait binocular range-measurement system [J] Jilin University journal of the based on OpenCV, and 2014,
32 (2): 188-194.
[9] Sun W, Chen L, Hu B, et al.Binocular vision-based position
determination algorithm and system[C]//Proceedings of the 2012International
Conference on Computer Distributed Control and Intelligent Environmental
Monitoring.Piscataway:IEEE Computer Society, 2012:170-173.
[10]Ikeuchi K.Determining a depth map using a dual photometric stereo
[J] .The International Journal of Robotics Research, 1987,6 (1): 15-31.
[11] Shao M, Simchony T, Chellappa R.New algorithms from reconstruction
of a 3-d depth map from one or more images[C]//Proceedings of CVPR’88.Ann
Arbor:IEEE, 1988:530-535.
[12] Matthies L, Kanade T, Szeliski R.Kalman filter-based algorithms for
estimating depth from image sequences[J].International Journal of Computer
Vision, 1989,3 (3): 209-238.
[13] Mathies L, Szeliski R, Kanade T.Incremental estimation of dense
depth maps from image sequence[C]//Proceedings of CVPR'88.Ann Arbor;IEEE,
1988:366-374.
[14] Mori T, Yamamoto M.A dynamic depth extraction method [C] //
Proceedings of Third International Conference on Computer Vision.Osaka:IEEE,
1990:672-676.
[15] Inoue H, Tachikawa T, Inaba M.Robot vision system with a
Correlation chip for real-time tracking, optical flow and depth map generation
[C] //Proceeding of Robotics and Automation.Nice:IEEE, 1992:1621-1626.
[16] Tree image distance measuring method [J] the agricultural mechanics of Hu Tianxiang, Zheng Jiaqiang, Zhou Hongping based on binocular vision
Report, 2010,41 (11): 158-162.
[17] Yu Naigong, Huang Can, Lin Jia are calculated based on robot target positioning distance measuring technique study [J] of monocular vision
Machine measurement and control, 2012,20 (10): 2654-2660.
[18] ranging research [J] the robot in Wu Gang, Tang Zhen people's monocular formula autonomous robot vision guided navigation, 2010,
32 (6): 828-832.
[19] Lu Weiwei, Xiao Zhitao, Lei Meilin study [J] with distance measuring method based on the front vehicles detection of monocular vision
Video Applications and engineering, 2011,35 (1): 125-128.
[20] Wu C F, Lin C J, Lee C Y, et al.Applying a functional neurofuzzy
network to real-time lane detection and front-vehicle distance measurement
[J] .IEEE Transactions on Systems, Man and Cybernetics-Part C:Applications and
Reviews, 2012,42 (4): 577-589.
[21] yellow cloudling, peak, Xu Guoyan wait monocular depth information extraction [J] of based on single width vertical target image
BJ University of Aeronautics & Astronautics's journal, 2015,41 (4): 649-655.
Summary of the invention
The object of the present invention is to provide the extracting method of a kind of novel target and its sub-pixel angle point, the novel target
Optical axis direction angle point accurate can be not only extracted, but also has reasonably evaded perspective transform angle steel joint extraction accuracy and has caused
Influence;The extracting method of the sub-pixel angle point does not need to specify gridiron pattern number in advance and algorithm robustness is high, to abnormal
It is preferable to become more picture extraction effect.
To achieve the above object, the present invention adopts the following technical scheme:
A kind of novel target, including the gridiron pattern target that rectangle black block and rectangular white block are staggeredly stacked, each rectangle
The width of black block and rectangular white block is fixed length and equal, it is characterised in that: the target is horizontal positioned, setting target away from
A row color lump size nearest from camera is d mm*d mm, and subsequent latter row's length is than previous row's length incrementIf
xiFor the actual range of i-th of angle point to camera, yiFor the length of each color lump, then the difference DELTA d of adjacent color lump lengthiAre as follows:
If the relationship between the computational length and actual range of each color lump is f (x), can be obtained according to formula (8):
Therefore, when a target row grid size nearest apart from camera is dmm*d mm, subsequent every row's width fixes,
Length incrementFor d*f (x) mm.
A kind of extracting method of the sub-pixel angle point of novel target described in claim 1, it is characterised in that including such as
Lower step:
It is defined using the checkerboard angle point detection process based on growth of the propositions such as Andreas Geiger two different
Angle point template, a kind of for the angle point parallel with reference axis, another kind is for rotating 45 ° of angle point, according to pixel each in image
Point and the similarity parameter of template find angle point on the image, carry out initial Corner Detection;The position and direction of angle steel joint carry out
Sub-pixel finely extracts, and sub-pixel Corner character is carried out with the cornerSubPix () function in OpenCV, by most
The standard deviation rate of smallization gradient image refines edge direction vector;It is finally label angle point and exports its subpixel coordinate,
Gridiron pattern is grown and rebuild according to energy function, marks angle point, exports sub-pixel angular coordinate.
Compared with prior art, the beneficial effects of the present invention are: due to the adoption of the above technical scheme,
(1) on the basis of traditional gridiron pattern target, existing perspective transform is existing when according to camera shooting level ground
As using a kind of incremental novel chess case marker target of the width equal length of specific standard as experimental material, this target is not
Optical axis direction angle point accurate can be only extracted, and has reasonably been evaded caused by perspective transform angle steel joint extraction accuracy
It influences;
(2) checkerboard angle point detection process by propositions such as Andreas Geiger based on growth and OpenCV are provided
CornerSubPix () function combines, and algorithm does not need to specify gridiron pattern number in advance and algorithm robustness is high, to distortion journey
It is preferable to spend biggish picture extraction effect.
Detailed description of the invention
Fig. 1 is novel target figure;
Fig. 2 is Corner Detection Algorithm implementation flow chart;
Fig. 3 is that camera upward angle of visibility is higher than horizontal line shooting geometrical model figure;
Fig. 4 is camera upward angle of visibility lower than horizontal line shooting geometrical model figure;
Fig. 5 is camera shooting perspective geometry model;
Fig. 6 is each coordinate system schematic diagram in pin-hole model;
Fig. 7 is camera stereo imaging system principle;
Fig. 8 is the relational graph between object ordinate pixel value and imaging angle;
Specific embodiment
In order to be more clear technical solution of the present invention, below in conjunction with attached drawing 1 to 8, the present invention is described in detail.
It should be understood that specific embodiment described in this specification is not intended to limit just for the sake of explaining the present invention
Protection scope of the present invention.
The present invention is a kind of novel target, and interlock stacked gridiron pattern target including rectangle black block and rectangular white block,
The width of each rectangle black block and rectangular white block is fixed length and equal, and the target is horizontal positioned, sets target distance
The nearest row's color lump size of camera is d mm*d mm, and subsequent latter row's length is than previous row's length incrementIf xi
For the actual range of i-th of angle point to camera, yiFor the length of each color lump, then the difference DELTA d of adjacent color lump lengthiAre as follows:
If the relationship between the computational length and actual range of each color lump is f (x), can be obtained according to formula (8):
Therefore, when a target row grid size nearest apart from camera is dmm*d mm, subsequent every row's width fixes,
Length incrementFor d*f (x) mm.
It is verified using novel target of the invention, it is in extremely significant line between the computational length and actual range of each color lump
Property correlativity, therefore can preferably f (x)) be constant.
A kind of extracting method of the sub-pixel angle point of above-mentioned novel target includes the following steps: to utilize Andreas
The checkerboard angle point detection process based on growth of the propositions such as Geiger defines two different angle point templates, and one kind is used for and sits
The parallel angle point of parameter, another kind exist for rotating 45 ° of angle point according to the similarity parameter of pixel each in image and template
Angle point is found on image, carries out initial Corner Detection;The position and direction of angle steel joint carry out sub-pixel and finely extract, and use
CornerSubPix () function in OpenCV carries out sub-pixel Corner character, by minimize the standard of gradient image from
Rate refines edge direction vector;It is finally label angle point and exports its subpixel coordinate, is laid equal stress on according to energy function growth
Gridiron pattern is built, angle point is marked, exports sub-pixel angular coordinate.
Embodiment 1
A kind of depth extraction and passive ranging method of the optimization based on monocular vision, include the following steps:
One, mobile phone camera is demarcated, obtains camera internal parameter and image resolution ratio.The calibration uses Zhang Zhengyou
Standardization, and introduce the improved peg model with nonlinear distortion variable and camera internal parameter is corrected.
The physical size size as pixel each in plane is set first as dx*dy (unit: mm), image coordinate system (x,
Y) coordinate of the origin in pixel coordinate system (u, v) is (u0, v0), (x, y) is the normalized coordinate of picture point in real image, figure
Any pixel meets following relationship in two coordinate systems as in:
fx、fyFor the normalization focal length in x-axis and y-axis, any point P in camera coordinates systemc(Xc, Yc, Zc) project to image
It is (x on coordinate systemc, yc, f), it is f with initial point distance that image coordinate system plane is vertical with optical axis z-axis, according to similar triangles original
Reason it follows that
Introduce the improved peg model with nonlinear distortion variable, including the diameter as caused by lens shape defect
To distortion and since there are tangential distortion caused by different degrees of bias, radial distortion mathematical models for optical system are as follows:
Wherein r2=x2+y2, (x ', y ') it is sat for the normalization of the ideal linearity camera coordinates system without distortion term after correction
Scale value, radial distortion value is related with the position of picture point in the picture, and the radial distortion value at image border is larger,
Tangential distortion model mathematical model are as follows:
It wherein include k1、k2、k3、p1、p2Totally 5 kilrrfactors obtain distortion correction Function Modules by formula (3), (4)
Type is as follows:
There are following relationships for conversion from world coordinate transformation to camera coordinates:
Pc=R (PW- C)=RPW+T (6)
Convolution (1)~(6), may be expressed as: with homogeneous coordinates and matrix form
Mint、MextIt is the inside and outside parameter matrix of camera calibration respectively, wherein camera internal parameter includes image center pixel
Value u0、v0, fx、fyFor in x-axis and y-axis normalization focal length, by Java combination OpenCV realize mobile phone camera calibration, acquisition
Inner parameter described in mobile phone camera, camera lens distortion parameter and image resolution ratio vmax、umax;Calibration gained camera internal ginseng
Number are as follows: fx=3486.5637, u0=1569.0383, fy=3497.4652, v0=2107.9899, image resolution ratio 3120
× 4208, camera lens distortion parameter are as follows: [0.0981, -0.1678,0.0003, -0.0025,0.0975],
Two, it by the acquisition to novel target image, establishes camera depth and extracts model.Existing target is that length and width are equal
Black and white chessboard case marker target.The difference of modulation of the invention and existing target is, setting target is apart from camera nearest
One row's grid size is d*d mm, and the width of subsequent every row's grid is to fix, and the previous row's value added of the latter parallelism of length is
X in following formulaiFor the actual range of i-th of angle point to camera, yiFor the length of each grid, then adjacent square length
Difference DELTA diAre as follows:
If the relationship between the computational length and actual range of each grid is f (x), can be obtained according to formula (8):
Through Pearson correlation analysis, it is between length and actual range extremely significant linear relationship (p < 0.01),
Correlation coefficient r be equal to 0.975, by least square method can in the hope of calculate f (x) derivative f ' (x),
Therefore, when a target row grid size nearest apart from camera is d*d mm (the survey when range of d takes 30~60mm
Accuracy of measurement highest) when, subsequent every row's width fixes, length incrementFor d*f ' (x) mm, novel target as shown in Figure 1,
There are the angles that perspective transform phenomenon makes Harris and Shi-Tomasi etc. common when object on shooting level ground
Point detection algorithm robustness is poor, and can also detect mistake when camera is larger along camera coordinates system ox axis rotated counterclockwise by angle
It loses, therefore the checkerboard angle point detection process based on growth of the propositions such as present invention combination Andreas Geiger and OpenCV are mentioned
The comerSubPix () function of confession carries out the detection of sub-pixel corner location, and the algorithm robustness is high, larger to distortion degree
Picture extraction effect it is preferable,
The implementation process of Corner Detection Algorithm as shown in Fig. 2, the above-mentioned modulation of the present invention sub-pixel angle point grid
Step are as follows:
1) angle point is found according to the similarity parameter of pixel each in image and template on the image, positions target angle point position
It sets;
Two different angle point templates are defined first, and a kind of for the angle point parallel with reference axis, another kind is for rotating
45 ° of angle point, each template are made of 4 filtering cores { A, B, C, E }, with carrying out convolution operation with image later;Then sharp
The similarity of each inflection point and angle point is calculated with the two angle point templates:
WhereinIndicate that convolution kernel X (X=A, B, C, E) and template i (i=1,2) are responded in the convolution of some pixel,
WithIt indicates the similarity of two kinds of possible inflection points of template i, calculates the available angle of similarity of each pixel in image
Point similar diagram;It is handled using non-maxima suppression algorithm angle steel joint pixel map to obtain candidate point;Then it is counted with gradient
Method verify these candidate points in the nxn neighborhood of a local, first local area grayscale image carries out sobel filtering, then counts
Weighting direction histogram (32bins) is calculated, finds two therein main mode γ with meanshift algorithm1And γ2;According to
The direction at edge, for desired gradient intensityConstruct a template T.(* indicates cross-correlation operation symbol)
Then product with angle point similarity is judged just to obtain initial angle point with threshold value as angle point score value.
2) the position and direction progress sub-pixel of angle steel joint finely extracts;
Sub-pixel Corner character is carried out with the cornerSubPix () function in OpenCV, by Corner character to sub- picture
Element, to obtain the other Corner Detection effect of sub-pixel;To refine edge direction vector, it is minimized according to image gradient value
Standard deviation rate:
WhereinIt is adjacent pixel collection, the gradient value m with module ii
=[cos (γi)sin(γi)]TMatch.(ask calculation scheme according to document Geiger A, Moosmann F, Caret
a1.Automatic camera and range sensor calibration using a single shot[C]//
Robotics and Automation (ICRA), 2012 IEEE International Conference on.IEEE,
2012:3936-3943.)
3) it is finally label angle point and exports its subpixel coordinate, gridiron pattern is grown and rebuild according to energy function, marks
Remember angle point, exports sub-pixel angular coordinate;
According to document " Geiger A, MoosmannF, Caret al.Automatic camera and range
Sensor calibration using a single shot [C] //Robotics and Automation (ICRA),
2012IEEE International Conference on.IEEE, the method that 2012:3936-3943. " is provided optimize energy
Function rebuilds gridiron pattern and marks angle point, energy growth function formula are as follows:
E (x, y)=Ecorners(y)+Estruct(x, y) (16)
Wherein, EcornersIt is the negative value of current chessboard angle point sum, EstructIt is of two adjacent corner points and prediction angle point
With degree;Angle point pixel value is exported by OpenCV.
Linear correlative analysis is carried out to image objects angle, ordinate pixel value using SPSS 22, exports Pearson phase
Relationship number, output Pearson correlation coefficient r are as shown in table 2.It is verified, under equipment and camera the rotation angle of different model, object
Body ordinate pixel value and actual imaging angle are in extremely significant negative correlativing relation (p < 0.01), in addition, the present invention is also to different
The slope difference of linear function carries out between object ordinate pixel value and imaging angle under device model and camera rotation angle
Significance test, the results showed that, under distinct device model and camera rotation angle object ordinate pixel value and imaging angle it
Between linear function slope differences it is heteropolar significant (p < 0.01), illustrate that the equipment of different model and camera rotation angle, depth mention
Modulus type is different,
2 object ordinate pixel value of table and imaging angle related coefficient
Table 2 Pearson correlation coefficient of image ordinate pixel
values and actual imaging angles
Note: * * indicates extremely significant (p < 0.01).
Note:**represents very significant correlation (p < 0.01)
Verified, equipment and camera with model rotate under angle, and object ordinate pixel value is in actual imaging angle
Extremely significant negative correlativing relation (p < 0.01), correlation coefficient r are greater than 0.99.In addition, the present invention is also to different device model and phase
The slope difference that machine rotates linear function between object ordinate pixel value and imaging angle under angle carries out significance test.Knot
Fruit show distinct device model and camera rotation angle under between object ordinate pixel value and imaging angle linear function it is oblique
Rate difference is extremely significant (p < 0.01), illustrates the equipment and camera rotation angle of different model, depth extraction model is not
Together.
Abstract function is set according to the linear relationship between object imaging angle α and ordinate pixel value v, establishes and contains mesh
Mark tri- object imaging angle α, ordinate pixel value v and camera rotation angle β parameter space relational models, i.e. α=F (v, β),
Under equipment and camera the rotation angle of different model, subject ordinate pixel value and imaging angle are in pole
Significant negative linear correlation, and the slope of the linear relationship and intercept are different, therefore set:
α=F (v, β)=av+b (17)
Wherein parameter a, b is related with camera model and camera rotation angle,
When α is minimized α=αminWhen=90- θ-β, θ is the half at camera vertical field of view angle, i.e. subject projects
When to picture lowermost end, v=vmax(vmaxFor camera CMOS or ccd image sensor column coordinate valid pixel number), substitute into formula
(17) it can obtain:
90- β-θ=avmax+b (18)
Work as αminWhen+2 90 ° of θ >, i.e. θ > β, camera upward angle of visibility is higher than horizontal line at this time, and camera shoots perspective geometry model
Such as Fig. 3, ground level unlimited distance, α is infinitely close to 90 °, and v is substantially equal to v at this time0-tanβ*fy, fyFor phase under pixel unit
The focal length of machine, when β rotates counterclockwise for negative value, that is, camera also similarly, therefore, substituting into formula (17) can obtain:
90=a (v0-tanβ·fy)+b (19)
Work as αminWhen+2 90 ° of θ <, i.e. θ < β, camera upward angle of visibility is lower than horizontal line at this time, and camera shoots perspective geometry model
If Fig. 4, ground level unlimited distance object imaging angle α are maximized, αmax=αminWhen+2 θ=90- β+θ, i.e. subject
When body projects to picture highest point, v=0, substituting into formula (17) can be obtained:
90- β+θ=b (20)
According to pinhole camera aufbauprinciple, the tangent value of the camera vertical field of view angle θ of half is schemed equal to camera CMOS or CCD
As the half of sensor side length is divided by camera focus, therefore θ can be calculated:
L in formula (21)CMOSFor the side length of camera CMOS or ccd image sensor, convolution (18)~(21), F (α, β)
Are as follows:
δ is camera nonlinear distortion variable error in formula (10), in conjunction with mobile phone camera shooting height h, according to trigonometric function
Principle establishes mobile phone camera depth extraction model:
3 mobile phone camera inner parameter of millet is substituted into formula (10) to obtain:
The specific depth extraction model of the equipment is obtained according to trigonometric function principle are as follows:
Three, by the Image Acquisition to target to be measured, target point pixel value u, v are obtained.Figure is carried out by mobile phone camera
As acquisition, perspective geometry model such as Fig. 5 is established, wherein f is camera focus, and θ is the half at camera vertical field of view angle, and h is camera
It takes pictures highly, β is rotation angle of the camera along camera coordinates system ox axis, and camera rotates clockwise β value and is positive, is negative counterclockwise, β value
It is obtained by camera internal gravity sensor, α is object imaging angle;The camera lens obtained in conjunction with first step camera calibration
Distortion parameter carries out nonlinear distortion correction to radial distortion existing for image and tangential distortion error;By the ideal after correction
Linear normalization coordinate value (x, y) substitutes into formula (1), image each point pixel coordinate value after asking calculating to correct, by bilinearity
Image after slotting method corrects pixel value progress interpolation processing after correction;Using computer vision and image procossing
Technology pre-processes the image after correction, including image binaryzation, morphological image operation and the inspection of object contour edge
It surveys, obtains the edge of object, and then calculate the geometric center point pixel value (u, v) at the edge of object and ground face contact.
Use (MI 3) camera of millet mobile phone 3 as picture collection equipment, carries out picture collection by camera trivets, and
The height h for measuring camera to ground is equal to 305mm, and camera rotation angle β is equal to 0 °,
Nonlinear distortion correction is carried out to radial distortion existing for image and tangential distortion error;
The camera lens distortion parameter obtained according to first step camera calibration: [0.0981, -0.1678,0.0003, -
0.0025,0.0975], ideal linearity normalized coordinate value after correcting is calculated according to formula (5):
Image each point pixel coordinate value after correcting is calculated in conjunction with formula (1) and (2), is handled and is rectified by bilinear interpolation
Image after just;
The present invention measures its depth and distance by taking the cuboid box being placed on level ground as an example, first to acquisition
Image carries out binary conversion treatment, then carries out edge detection to cuboid box using Canny operator, extracts object profile.
Extracting cuboid box bottom margin central point pixel value is (1851.23,3490).
Four, the camera internal parameter and target point pixel value and combining camera depth extraction mould of above-mentioned steps acquisition are utilized
Type calculates object to be measured object image and takes up an official post meaning point to the distance between mobile phone camera L.Angle beta and half are rotated according to camera
Camera vertical field of view angle θ between size relation, select corresponding depth model, ask the camera internal of calculation to join above-mentioned steps
Number image center pixel value v0, normalized focal length f in y-axisyAnd image resolution ratio vmaxAnd above-mentioned steps ask the to be measured of calculation
Object ordinate pixel value v, camera rotation angle beta and mobile phone camera shooting height h substitute into the depth extraction model, calculate
Target point depth value D,
Fig. 6 is camera stereo imaging system schematic diagram, and midpoint P is camera position, and the straight line and image where point A, B are flat
Face is parallel, and coordinate of the A under camera coordinates system is (X, Y, Z), and the coordinate of point B is (X+Tx, Y, Z), project to plane of delineation A '
(xl, yl)、B’(xr, yr) on, it can be obtained according to formula (2):
In conjunction with formula (1) and formula (22), the two o'clock A ' that Y value is identical and depth Z is equal, the horizontal parallax of B ' can be derived
D:
It is thus known that camera focus f, image center coordinate (u0, v0) and as pixel each in plane in the direction of the x axis
Physical size size dx, in conjunction with depth extraction model, calculate target point to optical axis direction vertical range Tx:
In pin-hole model, the transformational relation between each coordinate system of camera is as shown in fig. 7, calculating target point depth value D
And its vertical range T to optical axis directionxOn the basis of, according to formula (11)~(12), arbitrary point on image can be calculated
To shooting the distance between camera L:
Camera internal parameter, camera are taken pictures height h, rotation angle beta and cuboid box bottom margin central point is vertical sits
Mark pixel value v, which substitutes into formula (24), can calculate the object actual imaging angle equal to 69.58 °.According to trigonometric function original
Reason calculates target point depth value D (unit: mm):
D=305*tan 69.58 °=819.21 (27)
By parameter fx, u0, D and cuboid box bottom margin central point abscissa pixel value u substitute into formula (12) and can count
Vertical range T of the calculation object geometric center point to optical axis directionx:
Therefore, which reaches shooting camera in the distance L of floor projection point are as follows:
By tape measuring, distance of the cuboid box apart from camera floor projection point is 827mm, therefore uses the present invention
Carry out ranging, relative error 0.62%.
Embodiment 2
Below by taking millet 3 (MI 3) mobile phone as an example, novel target and its sub-pixel angle point of the invention are illustrated
Extracting method.
One, mobile phone camera is demarcated, obtains camera internal parameter and image resolution ratio
Use ranks number be 8*9 size be 20*20 gridiron pattern scaling board as the experimental material of camera calibration, lead to
The scaling board picture that 3 mobile phone camera of millet acquires 20 different angles is crossed, using OpenCV according to above-mentioned improved with non-thread
The camera calibration model of sex distortion item demarcates millet 3 (MI 3) mobile phone camera,
Scaling board picture is read using fin () function first, and obtains the image of the first picture by .cols .rows
Resolution ratio;Then sub-pixel angle point in scaling board picture is extracted by find4QuadCornerSubpix () function, be used in combination
DrawChessboardCorners () function marks angle point;CalibrateCamera () function is called to demarcate camera,
It is used for obtained camera interior and exterior parameter to carry out projection again to the three-dimensional point in space calculating, obtains new subpoint, calculate
Error between new subpoint and old subpoint;Camera internal reference matrix and distortion parameter are exported and save,
Calibration gained camera internal parameter are as follows: fx=3486.5637, u0=1569.0383, fy=3497.4652, v0=
2107.9899, image resolution ratio is 3120 × 4208, camera lens distortion parameter are as follows: [0.0981, -0.1678,0.0003, -
0.0025,0.0975],
Two, it by the acquisition to novel target image, establishes camera depth and extracts model
The initial experiment material that the present invention uses traditional gridiron pattern scaling board of 45*45mm to design as target, to calculate
The difference of adjacent square length, the present invention devise 6 groups of experiments, extract traditional X-comers that grid size is 45*45mm
Value, and ask calculate adjacent corner points between the actual physics distance that is represented under world coordinate system of unit pixel, to guarantee to indulge between angle point
Coordinate pixel value difference is roughly equal, the length y of each gridiValue it is as shown in table 1,
The calculating width of each grid of table 1
Table 1 Computing width of each grid
Through Pearson correlation analysis, it is between length and actual range extremely significant linear relationship (p < 0.01),
Correlation coefficient r is equal to 0.975, can be in the hope of derivative f ' (x)=0.262 of calculating f (x), therefore, when the mark by least square method
When a range row grid size nearest from camera is 45*45mm, then every row's width is fixed, width value added Δ d is
11.79mm,
The angle point of the novel target is extracted by the Robust Algorithm of Image Corner Extraction in specific implementation step,
The present invention use millet 3 choose respectively millet, Huawei, tri- kinds of different models of iPhone smart phone as image
Equipment is acquired, camera rotates angle beta=0 °.Data are acquired using the Corner Detection Algorithm, and it is quasi- to carry out function to its relationship
It closes, Fig. 8 is the relationship between the ordinate pixel value fitted and image objects angle, can according to angle point grid data and Fig. 8
, 3 mobile phone of millet depth extraction model when camera rotates angle beta=0 ° are as follows:
α=- 0.015v+112.6 (30)
Extracting cuboid box bottom margin central point pixel value is (1762.05,2360), by cuboid box bottom sides
Edge central point ordinate pixel value v, which substitutes into formula (29), can calculate the object actual imaging angle equal to 74.2 °.According to
Trigonometric function principle calculates target point depth value D (unit: mm):
D=305*tan 74.2 °=1077.85 (31).
Claims (2)
1. a kind of novel target, including the gridiron pattern target that rectangle black block and rectangular white block are staggeredly stacked, each rectangle is black
The width of color lump and rectangular white block is fixed length and equal, it is characterised in that: the target is horizontal positioned, sets target distance
The nearest row's color lump size of camera is d mm*d mm, and subsequent latter row's length is than previous row's length incrementIf xi
For the actual range of i-th of angle point to camera, yiFor the length of each color lump, then the difference DELTA d of adjacent color lump lengthiAre as follows:
If the relationship between the computational length and actual range of each color lump is f (x), can be obtained according to formula (8):
Therefore, when a target row grid size nearest apart from camera is dmm*d mm, subsequent every row's width fixed, length
Value addedFor d*f ' (x) mm.
2. a kind of extracting method of the sub-pixel angle point of novel target described in claim 1, it is characterised in that including as follows
Step:
Two different angle points are defined using the checkerboard angle point detection process based on growth of the propositions such as Andreas Geiger
Template, it is a kind of for the angle point parallel with reference axis, another kind for rotate 45 ° of angle point, according to pixel each in image and
The similarity parameter of template finds angle point on the image, carries out initial Corner Detection;The position and direction of angle steel joint carry out sub- picture
Plain grade is finely extracted, and is carried out sub-pixel Corner character with the comerSubPix () function in OpenCV, is passed through minimum
The standard deviation rate of gradient image refines edge direction vector;It is finally label angle point and exports its subpixel coordinate, according to
Energy function grows and rebuilds gridiron pattern, marks angle point, exports sub-pixel angular coordinate.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810918877.7A CN109255818B (en) | 2018-08-12 | 2018-08-12 | Novel target and extraction method of sub-pixel level angular points thereof |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810918877.7A CN109255818B (en) | 2018-08-12 | 2018-08-12 | Novel target and extraction method of sub-pixel level angular points thereof |
Publications (2)
Publication Number | Publication Date |
---|---|
CN109255818A true CN109255818A (en) | 2019-01-22 |
CN109255818B CN109255818B (en) | 2021-05-28 |
Family
ID=65049244
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201810918877.7A Active CN109255818B (en) | 2018-08-12 | 2018-08-12 | Novel target and extraction method of sub-pixel level angular points thereof |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109255818B (en) |
Cited By (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109859137A (en) * | 2019-02-14 | 2019-06-07 | 重庆邮电大学 | A kind of irregular distortion universe bearing calibration of wide angle camera |
CN110443856A (en) * | 2019-08-12 | 2019-11-12 | 广州图语信息科技有限公司 | A kind of 3D structure optical mode group scaling method, storage medium, electronic equipment |
CN111798422A (en) * | 2020-06-29 | 2020-10-20 | 福建汇川物联网技术科技股份有限公司 | Checkerboard angular point identification method, device, equipment and storage medium |
CN113256735A (en) * | 2021-06-02 | 2021-08-13 | 杭州灵西机器人智能科技有限公司 | Camera calibration method and system based on binocular calibration |
CN113658272A (en) * | 2021-08-19 | 2021-11-16 | 湖北亿咖通科技有限公司 | Vehicle-mounted camera calibration method, device, equipment and storage medium |
CN114187363A (en) * | 2021-11-24 | 2022-03-15 | 北京极豪科技有限公司 | Method and device for obtaining radial distortion parameter value and mobile terminal |
CN114255272A (en) * | 2020-09-25 | 2022-03-29 | 广东博智林机器人有限公司 | Positioning method and device based on target image |
Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202066514U (en) * | 2011-03-21 | 2011-12-07 | 李为民 | Compound target for measuring large-scale coordinate |
CN202622812U (en) * | 2012-05-25 | 2012-12-26 | 山东天泽软控技术股份有限公司 | Calibrating plate for visual system of robot |
CN103019643A (en) * | 2012-12-30 | 2013-04-03 | 中国海洋大学 | Method for automatic correction and tiled display of plug-and-play large screen projections |
CN203149664U (en) * | 2013-03-27 | 2013-08-21 | 黑龙江科技学院 | Calibration plate for binocular vision camera |
CN103292710A (en) * | 2013-05-27 | 2013-09-11 | 华南理工大学 | Distance measuring method applying binocular visual parallax error distance-measuring principle |
CN203217624U (en) * | 2013-05-04 | 2013-09-25 | 长春工业大学 | Novel checkerboarded calibration target |
CN103927750A (en) * | 2014-04-18 | 2014-07-16 | 上海理工大学 | Detection method of checkboard grid image angular point sub pixel |
CN204388802U (en) * | 2015-01-19 | 2015-06-10 | 长春师范大学 | Line-structured light vision system calibration plate |
CN105105779A (en) * | 2014-04-18 | 2015-12-02 | Fei公司 | High aspect ratio X-ray targets and uses of same |
CN106803273A (en) * | 2017-01-17 | 2017-06-06 | 湖南优象科技有限公司 | A kind of panoramic camera scaling method |
CN206574134U (en) * | 2017-03-09 | 2017-10-20 | 昆山鹰之眼软件技术有限公司 | Automate scaling board |
-
2018
- 2018-08-12 CN CN201810918877.7A patent/CN109255818B/en active Active
Patent Citations (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN202066514U (en) * | 2011-03-21 | 2011-12-07 | 李为民 | Compound target for measuring large-scale coordinate |
CN202622812U (en) * | 2012-05-25 | 2012-12-26 | 山东天泽软控技术股份有限公司 | Calibrating plate for visual system of robot |
CN103019643A (en) * | 2012-12-30 | 2013-04-03 | 中国海洋大学 | Method for automatic correction and tiled display of plug-and-play large screen projections |
CN203149664U (en) * | 2013-03-27 | 2013-08-21 | 黑龙江科技学院 | Calibration plate for binocular vision camera |
CN203217624U (en) * | 2013-05-04 | 2013-09-25 | 长春工业大学 | Novel checkerboarded calibration target |
CN103292710A (en) * | 2013-05-27 | 2013-09-11 | 华南理工大学 | Distance measuring method applying binocular visual parallax error distance-measuring principle |
CN103927750A (en) * | 2014-04-18 | 2014-07-16 | 上海理工大学 | Detection method of checkboard grid image angular point sub pixel |
CN105105779A (en) * | 2014-04-18 | 2015-12-02 | Fei公司 | High aspect ratio X-ray targets and uses of same |
CN204388802U (en) * | 2015-01-19 | 2015-06-10 | 长春师范大学 | Line-structured light vision system calibration plate |
CN106803273A (en) * | 2017-01-17 | 2017-06-06 | 湖南优象科技有限公司 | A kind of panoramic camera scaling method |
CN206574134U (en) * | 2017-03-09 | 2017-10-20 | 昆山鹰之眼软件技术有限公司 | Automate scaling board |
Non-Patent Citations (3)
Title |
---|
周传德: "基于平面标靶的物体空间姿态精确测量", 《重庆大学学报》 * |
张胜男: "圆阵列平面靶标特征点的自动标记", 《计算机工程与应用》 * |
白瑞林: "一种实用的X型靶标亚像素角点提取方法", 《光学技术》 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109859137A (en) * | 2019-02-14 | 2019-06-07 | 重庆邮电大学 | A kind of irregular distortion universe bearing calibration of wide angle camera |
CN109859137B (en) * | 2019-02-14 | 2023-02-17 | 重庆邮电大学 | Wide-angle camera irregular distortion global correction method |
CN110443856A (en) * | 2019-08-12 | 2019-11-12 | 广州图语信息科技有限公司 | A kind of 3D structure optical mode group scaling method, storage medium, electronic equipment |
CN111798422A (en) * | 2020-06-29 | 2020-10-20 | 福建汇川物联网技术科技股份有限公司 | Checkerboard angular point identification method, device, equipment and storage medium |
CN111798422B (en) * | 2020-06-29 | 2023-10-20 | 福建汇川物联网技术科技股份有限公司 | Checkerboard corner recognition method, device, equipment and storage medium |
CN114255272A (en) * | 2020-09-25 | 2022-03-29 | 广东博智林机器人有限公司 | Positioning method and device based on target image |
CN113256735A (en) * | 2021-06-02 | 2021-08-13 | 杭州灵西机器人智能科技有限公司 | Camera calibration method and system based on binocular calibration |
CN113658272A (en) * | 2021-08-19 | 2021-11-16 | 湖北亿咖通科技有限公司 | Vehicle-mounted camera calibration method, device, equipment and storage medium |
CN113658272B (en) * | 2021-08-19 | 2023-11-17 | 亿咖通(湖北)技术有限公司 | Vehicle-mounted camera calibration method, device, equipment and storage medium |
CN114187363A (en) * | 2021-11-24 | 2022-03-15 | 北京极豪科技有限公司 | Method and device for obtaining radial distortion parameter value and mobile terminal |
Also Published As
Publication number | Publication date |
---|---|
CN109255818B (en) | 2021-05-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109146980A (en) | The depth extraction and passive ranging method of optimization based on monocular vision | |
CN109035320A (en) | Depth extraction method based on monocular vision | |
CN109255818A (en) | A kind of extracting method of novel target and its sub-pixel angle point | |
CN109269430B (en) | Multi-standing-tree breast height diameter passive measurement method based on deep extraction model | |
CN102376089B (en) | Target correction method and system | |
Hu et al. | Extrinsic calibration of 2-D laser rangefinder and camera from single shot based on minimal solution | |
CN111563921B (en) | Underwater point cloud acquisition method based on binocular camera | |
CN103411553B (en) | The quick calibrating method of multi-linear structured light vision sensors | |
CN103971378A (en) | Three-dimensional reconstruction method of panoramic image in mixed vision system | |
CN107977996B (en) | Space target positioning method based on target calibration positioning model | |
CN106595528A (en) | Digital speckle-based telecentric microscopic binocular stereoscopic vision measurement method | |
CN109579695B (en) | Part measuring method based on heterogeneous stereoscopic vision | |
CN103278138A (en) | Method for measuring three-dimensional position and posture of thin component with complex structure | |
CN109163657A (en) | A kind of circular target position and posture detection method rebuild based on binocular vision 3 D | |
CN103473771A (en) | Method for calibrating camera | |
CN107084680A (en) | Target depth measuring method based on machine monocular vision | |
CN108010086A (en) | Camera marking method, device and medium based on tennis court markings intersection point | |
CN107481288A (en) | The inside and outside ginseng of binocular camera determines method and apparatus | |
CN110763204A (en) | Planar coding target and pose measurement method thereof | |
CN102914295A (en) | Computer vision cube calibration based three-dimensional measurement method | |
CN104167001B (en) | Large-visual-field camera calibration method based on orthogonal compensation | |
CN107680035B (en) | Parameter calibration method and device, server and readable storage medium | |
CN116205993A (en) | Double-telecentric lens high-precision calibration method for 3D AOI | |
CN104123726B (en) | Heavy forging measuring system scaling method based on vanishing point | |
CN104596486B (en) | Pose measuring method based on the rotationally symmetrical feature of target |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |