CN108759666A - A kind of dimension measurement method based on flight time three-dimensional camera - Google Patents

A kind of dimension measurement method based on flight time three-dimensional camera Download PDF

Info

Publication number
CN108759666A
CN108759666A CN201810524023.0A CN201810524023A CN108759666A CN 108759666 A CN108759666 A CN 108759666A CN 201810524023 A CN201810524023 A CN 201810524023A CN 108759666 A CN108759666 A CN 108759666A
Authority
CN
China
Prior art keywords
point
babinet
dimensional
flight time
method based
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810524023.0A
Other languages
Chinese (zh)
Other versions
CN108759666B (en
Inventor
黄毛毛
余晨阳
郭云锋
付敏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan Dot Eye 3d Technology Co Ltd
Original Assignee
Wuhan Dot Eye 3d Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan Dot Eye 3d Technology Co Ltd filed Critical Wuhan Dot Eye 3d Technology Co Ltd
Priority to CN201810524023.0A priority Critical patent/CN108759666B/en
Publication of CN108759666A publication Critical patent/CN108759666A/en
Application granted granted Critical
Publication of CN108759666B publication Critical patent/CN108759666B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/02Measuring arrangements characterised by the use of optical techniques for measuring length, width or thickness
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T19/00Manipulating 3D models or images for computer graphics
    • G06T19/20Editing of 3D images, e.g. changing shapes or colours, aligning objects or positioning parts
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/60Analysis of geometric attributes
    • G06T7/62Analysis of geometric attributes of area, perimeter, diameter or volume
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10004Still image; Photographic image
    • G06T2207/10012Stereo images

Landscapes

  • Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Geometry (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Architecture (AREA)
  • Computer Graphics (AREA)
  • Computer Hardware Design (AREA)
  • General Engineering & Computer Science (AREA)
  • Software Systems (AREA)
  • Length Measuring Devices By Optical Means (AREA)

Abstract

The present invention relates to a kind of dimension measurement methods based on flight time three-dimensional camera, light source is subjected to radio-frequency modulations, reflected light is collected by three-dimensional ToF cameras and makees related operation with demodulated signal, actively change the initial phase delay of demodulated signal, four groups of correlations are obtained, solve depth value accordingly and eliminate ambient light;For the present invention by the way that the algorithm in data conversion module and module is arranged, the camera coordinates system three-dimensional coordinate that control system is obtained to each point is converted to the length, width and height of babinet, can quick and precisely obtain the size and volume of babinet.

Description

A kind of dimension measurement method based on flight time three-dimensional camera
Technical field
The present invention relates to flow volume field of measuring technique, specifically a kind of ruler based on flight time three-dimensional camera Very little measurement method.
Background technology
The existing 3-D measuring apparatus for logistics field is the scheme based on structure light, to ambient light anti-interference ability Difference, the infrared discrete spot especially projected in sun light direct beam can be flooded by sunlight, and Measurement Algorithm is caused to fail.
Based on this, for limitation present in above-mentioned present situation, the present invention, which proposes one kind, can solve existing size survey By the dimension measurement method based on flight time three-dimensional camera of ambient light interference problem in amount scheme.
Invention content
In order to solve above-mentioned problems of the prior art, the present invention, which provides one kind, can solve existing dimensional measurement side By the dimension measurement method based on flight time three-dimensional camera of ambient light interference problem in case.
The technical solution adopted by the present invention to solve the technical problems is:
A kind of dimension measurement method based on flight time three-dimensional camera, includes the following steps:
Step 1: radio-frequency modulations processing is carried out to light source by the magazine control systems of three-dimensional ToF, and by modulation treatment Light source afterwards is emitted on babinet;
Step 2: observing reflected corresponding light by the magazine imaging systems of three-dimensional ToF, pixel seat is acquired Each pixel point coordinates (u, v) in mark system middle case;
Step 3: by the magazine data processing modules of three-dimensional ToF, each pixel in pixel coordinate system middle case is obtained Point coordinates (u, v) simultaneously obtains the camera coordinates system three-dimensional coordinate (X of each point by calculating conversionc, Yc, Zc);
Step 4: control system obtains the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) after be transmitted to terminal, terminal Interior data conversion module is by the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) length, width and height (L, W, H) of babinet are converted to, And display is exported by the data outputting module in terminal.
Further, the light source after a kind of modulation treatment of the step is square wave.
Further, for conversion process specifically, setting k × l as the size of pixel, f is that imaging is in the step 3 The focal length of system, θ are the gradient of camera coordinates system, and d is measurement distance of the light source to babinet,To emit the phase of light and reflected light Potential difference, c are light velocity constant, then the transformational relation between pixel coordinate system and camera coordinates system is,
Formula (1):
Formula (2):
Formula (3):
Association type (1), formula (2), formula (3), obtain (Xc, Yc, Zc)。
Further, in the step 3, emit the phase difference of light and reflected lightSpecific calculating process be, set τ Actively to change retardation in measuring every time,
It is measured by four times, and changes retardation τ in each measure, ambient light is constant K, is obtained:
To above-mentioned relation formula by way of differential mode, constant term K is eliminated, is obtained:
Further, in the step 4, the conversion process of the length, width and height (L, W, H) of babinet is the data conversion mould Block is provided with judgment criterion model, using the rejecting of judgment criterion iteration and the inconsistent input data of estimated parameter, by defeated Enter correct data estimation model parameter, specially:
S1, the initial point cloud data of the camera coordinates system three-dimensional coordinate of control system acquisition each point is randomly choosed Point forms subset L1, and parameter A, B, C of initialization areal model, model AX+BY+CZ+D=0 are obtained with least square method;
It is the data except subset L1 that S2, the model AX+BY+CZ+D=0 obtained with step S1, which remove test subset L2, L2, It is distance of any point to areal model that threshold value d, d, which is arranged, and the point set L ' and L1 for being less than d in subset L2 with model error constitute collection Close L*, set L*For interior point set, remaining is point not in the know;
Point set L in S3, statistics*The number of interior point judges whether to be more than the 1/3 of the number put in initial point cloud, if so, Then with interior point set L*Interior data estimate new model parameter A ', B ', C ', as the output of new model, if it is not, then with current Maximum interior point set L*The number of interior point compares, if more than the number of current most imperial palace point set is then replaced, and store current mould The parameter of type;
S4, step S1-S3 is repeated, is iterated processing, the model generated to each iteration screens, when in interior point set Point is then given up when very little, is then replaced when more preferable than "current" model, gets rid of the point cloud of non-babinet, obtains the point of babinet Cloud;
S5, the point cloud of babinet is filtered and is clustered, obtain the point cloud P of more accurate babinet;
S6, marginality detection is carried out to the point cloud P of acquisition, obtains the point cloud p at babinet edge;
S7, the linear edge profile of babinet is obtained to the point cloud p progress straight-line detection of acquisition;
S8, find babinet linear edge profile extreme coordinates, the length, width and height (L, W, H) of babinet are obtained by calculation;
S9, box volume is calculated by the product of length, width and height.
Further, in the step S2, threshold value d is set as 0.01m.
Further, in the step S4, iterations are set as 50 times.
Further, in the step S5,
Filtering includes straight-through filtering and Statis Outlier Removal filter filterings, and the straight-through filtering was used for Left and right noise is filtered, the Statis Outlier Removal filters filter outliers calculate each point and arrive and this institute There are the average distance of point of proximity, point of the average distance except a standard deviation range, as outlier;
Cluster is distance cluster, the noise to get up with other cohesions for isolating babinet.
Further, in the step S6, marginality detection process is:
For in cloud P a point q and the nearest k point of distance q be fitted, obtain in a least square meaning Part plan, be denoted as ax+by+cz+d=0, the normal (a, b, c) in this face is the normal of the point, finds out the normal of all the points Behind direction, the normal direction of k neighbor point is compared, takes the point that normal direction mutates as boundary point.
Further, in the step S7, detailed process is:
By sampling consistency, the point cloud of a plurality of line segment is obtained, the endpoint of line segment is respectively x coordinate minimax point, is led to The direction vector of every line segment can be calculated by crossing the coordinate of two endpoints, by the direction vector angle of every line segment, to case The length, width and height of body are classified.
Compared with prior art, the beneficial effects of the invention are as follows:
1, light source is carried out radio-frequency modulations by the present invention, is collected reflected light by three-dimensional ToF cameras and is made phase with demodulated signal Operation is closed, actively changes the initial phase delay of demodulated signal, obtains four groups of correlations, solve depth value accordingly and eliminate environment Light;
2, control system is obtained the camera of each point by the present invention by the way that the algorithm in data conversion module and module is arranged Coordinate system three-dimensional coordinate is converted to the length, width and height of babinet, can quick and precisely obtain the size and volume of babinet.
Description of the drawings
Present invention will be further explained below with reference to the attached drawings and examples.
Fig. 1 is the schematic diagram that the present invention is implemented;
Fig. 2 is the work flow diagram of the present invention;
Fig. 3 is in one embodiment of the present of invention, for the figure for getting rid of before the point cloud of non-babinet 2;
Fig. 4 is in one embodiment of the present of invention, for the figure for getting rid of after the point cloud of non-babinet 2;
Fig. 5 is the figure obtained after Fig. 4 is filtered and is clustered;
Fig. 6 is the figure for Fig. 5 obtain after marginality detection;
Fig. 7 is the figure for Fig. 6 obtain after straight-line detection.
Specific implementation mode
In order to make the purpose , technical scheme and advantage of the present invention be clearer, with reference to the accompanying drawings and embodiments, right The present invention is further elaborated.It should be appreciated that the specific embodiments described herein are merely illustrative of the present invention, and It is not used in the restriction present invention.
As depicted in figs. 1 and 2, a kind of dimension measurement method based on flight time three-dimensional camera of the present invention, packet Include following steps:
Step 1: radio-frequency modulations processing is carried out to light source 104 by the control system 103 in three-dimensional ToF cameras 1, and will Light source 101 after modulation treatment is emitted on babinet 2;
Step 2: observing reflected corresponding light 102 by the imaging system 105 in three-dimensional ToF cameras 1, obtain Each pixel point coordinates (u, v) on to pixel coordinate system middle case 2;
Step 3: by the data processing module 106 in three-dimensional ToF cameras 1, obtain each in pixel coordinate system middle case 2 Pixel point coordinates (u, v) simultaneously obtains the camera coordinates system three-dimensional coordinate (X of each point by calculating conversionc, Yc, Zc);
Step 4: control system 103 obtains the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) after be transmitted to terminal 3, Data conversion module 301 in terminal 3 is by the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) be converted to the length and width of babinet 2 High (L, W, H), and display is exported by the data outputting module 302 in terminal 3.
When it is implemented, the light source 101 after a kind of modulation treatment of step is square wave.
When it is implemented, conversion process is specifically, set k × l as the size of pixel, f is imaging in the step 3 The focal length of system 105, θ are the gradient of camera coordinates system, and d is measurement distance of the light source 104 to babinet 2,For transmitting light with The phase difference of reflected light, c are light velocity constant, then the transformational relation between pixel coordinate system and camera coordinates system is,
Formula (1):
Formula (2):
Formula (3):
Association type (1), formula (2), formula (3), obtain (Xc, Yc, Zc)。
When it is implemented, in the step 3, emit the phase difference of light and reflected lightSpecific calculating process be, set τ Actively to change retardation in measuring every time,
It is measured by four times, and changes retardation τ in each measure, ambient light is constant K, is obtained:
To above-mentioned relation formula by way of differential mode, constant term K is eliminated, is obtained:
When it is implemented, in the step 4, the conversion process of the length, width and height (L, W, H) of babinet 2 is the data conversion Module 301 is provided with judgment criterion model, using the rejecting of judgment criterion iteration and the inconsistent input data of estimated parameter, leads to It crosses and inputs correct data estimation model parameter, specially:
S1, the initial point cloud data of the camera coordinates system three-dimensional coordinate of the acquisition each point of control system 103 is randomly choosed one It is a little, subset L1 is formed, parameter A, B, C of initialization areal model, model AX+BY+CZ+D=are obtained with least square method 0;
It is the data except subset L1 that S2, the model AX+BY+CZ+D=0 obtained with step S1, which remove test subset L2, L2, It is distance of any point to areal model that threshold value d, d, which is arranged, and the point set L ' and L1 for being less than d in subset L2 with model error constitute collection Close L*, set L*For interior point set, remaining is point not in the know;
Point set L in S3, statistics*The number of interior point judges whether to be more than the 1/3 of the number put in initial point cloud, if so, Then with interior point set L*Interior data estimate new model parameter A ', B ', C ', as the output of new model, if it is not, then with current Maximum interior point set L*The number of interior point compares, if more than the number of current most imperial palace point set is then replaced, and store current mould The parameter of type;
S4, step S1-S3 is repeated, is iterated processing, the model generated to each iteration screens, when in interior point set Point is then given up when very little, is then replaced when more preferable than "current" model, gets rid of the point cloud of non-babinet 2, obtains the point of babinet 2 Cloud, as shown in Figure 3 and Figure 4, Fig. 3 are the figure before getting rid of the point cloud of non-babinet 2, and Fig. 4 is the point cloud for getting rid of non-babinet 2 Figure afterwards;
S5, the point cloud of babinet 2 is filtered and is clustered, obtain the point cloud P of more accurate babinet 2, as shown in Figure 5;
S6, marginality detection is carried out to the point cloud P of acquisition, obtains the point cloud p at 2 edge of babinet, as shown in Figure 6;
S7, the linear edge profile of babinet 2 is obtained to the point cloud p progress straight-line detection of acquisition, as shown in Figure 7;
S8, find babinet 2 linear edge profile extreme coordinates, be obtained by calculation babinet 2 length, width and height (L, W, H);
S9,2 volume of babinet is calculated by the product of length, width and height.
When it is implemented, in the step S2, threshold value d is set as 0.01m.
When it is implemented, in the step S4, iterations are set as 50 times, are meeting the unsuitable long feelings of iteration time The maximum times of iteration are capable of in selection under condition.
When it is implemented, in the step S5,
Filtering includes straight-through filtering and Statis Outlier Removal filter filterings, and the straight-through filtering was used for Left and right noise is filtered, the Statis Outlier Removal filters filter outliers calculate each point and arrive and this institute There are the average distance of point of proximity, point of the average distance except a standard deviation range, as outlier;
Cluster is distance cluster, the noise to get up with other cohesions for isolating babinet.
When it is implemented, in the step S6, marginality detection process is:
For in cloud P a point q and the nearest k point of distance q be fitted, obtain in a least square meaning Part plan, be denoted as ax+by+cz+d=0, the normal (a, b, c) in this face is the normal of the point, finds out the normal of all the points Behind direction, the normal direction of k neighbor point is compared, takes the point that normal direction mutates as boundary point.
When it is implemented, in the step S7, detailed process is:
By sampling consistency, the point cloud of a plurality of line segment is obtained, the endpoint of line segment is respectively x coordinate minimax point, is led to The direction vector of every line segment can be calculated by crossing the coordinate of two endpoints, by the direction vector angle of every line segment, to case The length, width and height of body 2 are classified.
The basic principles, main features and advantages of the present invention have been shown and described above.The technology of the industry Personnel are it should be appreciated that the present invention is not limited to the above embodiments, and the above embodiments and description only describe this The principle of invention, without departing from the spirit and scope of the present invention, various changes and improvements may be made to the invention, these changes Change and improvement all fall within the protetion scope of the claimed invention.The claimed scope of the invention by appended claims and its Equivalent thereof.

Claims (10)

1. a kind of dimension measurement method based on flight time three-dimensional camera, which is characterized in that include the following steps:
Step 1: radio-frequency modulations processing is carried out to light source by the magazine control systems of three-dimensional ToF, and will be after modulation treatment Light source is emitted on babinet;
Step 2: collecting reflected corresponding light by the magazine imaging systems of three-dimensional ToF, pixel coordinate system is acquired Each pixel point coordinates (u, v) in middle case;
Step 3: by the magazine data processing modules of three-dimensional ToF, obtains each pixel point in pixel coordinate system middle case and sit Mark (u, v) simultaneously obtains the camera coordinates system three-dimensional coordinate (X of each point by calculating conversionc, Yc, Zc);
Step 4: control system obtains the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) after be transmitted to terminal, in terminal Data conversion module is by the camera coordinates system three-dimensional coordinate (X of each pointc, Yc, Zc) length, width and height (L, W, H) of babinet are converted to, and lead to Cross the data outputting module output display in terminal.
2. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 1, which is characterized in that institute It is square wave to state the light source after a kind of modulation treatment of step.
3. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 1, which is characterized in that institute Conversion process in step 3 is stated specifically, setting k × l as the area of pixel, f is the focal length of imaging system, and θ is camera coordinates The gradient of system, d are measurement distance of the light source to babinet,To emit the phase difference of light and reflected light, c is light velocity constant, then Transformational relation between pixel coordinate system and camera coordinates system is,
Formula (1):
Formula (2):
Formula (3):
Association type (1), formula (2), formula (3), obtain (Xc, Yc, Zc)。
4. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 3, which is characterized in that institute It states in step 3, emits the phase difference of light and reflected lightSpecific calculating process be set τ as every time measure in actively change Retardation,
It is measured by four times, and changes retardation τ in each measure, ambient light is constant K, is obtained:
To above-mentioned relation formula by way of differential mode, constant term K is eliminated, is obtained:
5. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 1, which is characterized in that institute It states in step 4, the conversion process of the length, width and height (L, W, H) of babinet is that the data conversion module is provided with judgment criterion mould Type passes through using the rejecting of judgment criterion iteration and the inconsistent input data of estimated parameter and inputs correct data estimation mould Shape parameter, specially:
S1, the initial point cloud data of the camera coordinates system three-dimensional coordinate of control system acquisition each point is randomly choosed into some points, group At subset L1, parameter A, B, C of initialization areal model, model AX+BY+CZ+D=0 are obtained with least square method;
It is the data except subset L1, setting that S2, the model AX+BY+CZ+D=0 obtained with step S1, which remove test subset L2, L2, Threshold value d, d are distance of any point to areal model, and the point set L ' and L1 for being less than d in subset L2 with model error constitute set L*, set L*For interior point set, remaining is point not in the know;
Point set L in S3, statistics*The number of interior point judges whether to be more than the 1/3 of the number put in initial point cloud, if it is, with Interior point set L*Interior data estimate new model parameter A ', B ', C ', as the output of new model, if it is not, then with current maximum Interior point set L*The number of interior point compares, if more than the number of current most imperial palace point set is then replaced, and store "current" model Parameter;
S4, step S1-S3 is repeated, is iterated processing, the model of each iteration generation screened, when being put too in interior point set Then give up when few, is then replaced when more preferable than "current" model, gets rid of the point cloud of non-babinet, obtain the point cloud of babinet;
S5, the point cloud of babinet is filtered and is clustered, obtain the point cloud P of more accurate babinet;
S6, marginality detection is carried out to the point cloud P of acquisition, obtains the point cloud p at babinet edge;
S7, the linear edge profile of babinet is obtained to the point cloud p progress straight-line detection of acquisition;
S8, find babinet linear edge profile extreme coordinates, the length, width and height (L, W, H) of babinet are obtained by calculation;
S9, box volume is calculated by the product of length, width and height.
6. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 5, which is characterized in that institute It states in step S2, threshold value d is set as 0.01m.
7. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 5, which is characterized in that institute It states in step S4, iterations are set as 50 times.
8. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 5, which is characterized in that institute It states in step S5,
Filtering includes straight-through filtering and Statis Outlier Removal filter filterings, and the straight-through filtering is for filtering a left side Right noise, the Statis Outlier Removal filters filter outliers calculate each point to point is all faces with this The average distance of near point, point of the average distance except a standard deviation range, as outlier;
Cluster is distance cluster, the noise to get up with other cohesions for isolating babinet.
9. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 5, which is characterized in that institute It states in step S6, marginality detection process is:
For in cloud P a point q and the nearest k point of distance q be fitted, obtain a least squares sense office Facial planes is denoted as ax+by+cz+d=0, and the normal (a, b, c) in this face is the normal of the point, finds out the normal direction of all the points Afterwards, the normal direction of k neighbor point is compared, takes the point that normal direction mutates as boundary point.
10. a kind of dimension measurement method based on flight time three-dimensional camera according to claim 5, which is characterized in that In the step S7, detailed process is:
By sampling consistency, the point cloud of a plurality of line segment is obtained, the endpoint of line segment is respectively x coordinate minimax point, passes through two The coordinate of a endpoint can calculate the direction vector of every line segment, by the direction vector angle of every line segment, to babinet Length, width and height are classified.
CN201810524023.0A 2018-05-28 2018-05-28 Size measurement method based on time-of-flight three-dimensional camera Expired - Fee Related CN108759666B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810524023.0A CN108759666B (en) 2018-05-28 2018-05-28 Size measurement method based on time-of-flight three-dimensional camera

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810524023.0A CN108759666B (en) 2018-05-28 2018-05-28 Size measurement method based on time-of-flight three-dimensional camera

Publications (2)

Publication Number Publication Date
CN108759666A true CN108759666A (en) 2018-11-06
CN108759666B CN108759666B (en) 2020-06-23

Family

ID=64003086

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810524023.0A Expired - Fee Related CN108759666B (en) 2018-05-28 2018-05-28 Size measurement method based on time-of-flight three-dimensional camera

Country Status (1)

Country Link
CN (1) CN108759666B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109991583A (en) * 2019-03-14 2019-07-09 深圳奥比中光科技有限公司 A kind of jamproof distance measurement method and depth camera
CN109991584A (en) * 2019-03-14 2019-07-09 深圳奥比中光科技有限公司 A kind of jamproof distance measurement method and depth camera
CN110686600A (en) * 2019-11-05 2020-01-14 北京深测科技有限公司 Measuring method and system based on flight time measurement
CN112504125A (en) * 2020-12-15 2021-03-16 广东工业大学 Non-contact cuboid volume measurement method based on TOF
CN113418467A (en) * 2021-06-16 2021-09-21 厦门硅谷动能信息技术有限公司 Method for detecting general and black luggage size based on ToF point cloud data
CN113671519A (en) * 2021-10-22 2021-11-19 武汉市聚芯微电子有限责任公司 Time-of-flight ranging method and device, time-of-flight camera and electronic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110169922A1 (en) * 2010-01-09 2011-07-14 Hon Hai Precision Industry Co., Ltd. Three dimension model building system and method
CN102252653A (en) * 2011-06-27 2011-11-23 合肥工业大学 Position and attitude measurement method based on time of flight (TOF) scanning-free three-dimensional imaging
CN104316083A (en) * 2014-11-15 2015-01-28 中国科学院光电研究院 Three-dimensional coordinate calibration device and method of TOF (Time-of-Flight) depth camera based on sphere center positioning of virtual multiple spheres
CN109798838A (en) * 2018-12-19 2019-05-24 西安交通大学 A kind of ToF depth transducer and its distance measuring method based on laser speckle projection

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110169922A1 (en) * 2010-01-09 2011-07-14 Hon Hai Precision Industry Co., Ltd. Three dimension model building system and method
CN102252653A (en) * 2011-06-27 2011-11-23 合肥工业大学 Position and attitude measurement method based on time of flight (TOF) scanning-free three-dimensional imaging
CN104316083A (en) * 2014-11-15 2015-01-28 中国科学院光电研究院 Three-dimensional coordinate calibration device and method of TOF (Time-of-Flight) depth camera based on sphere center positioning of virtual multiple spheres
CN109798838A (en) * 2018-12-19 2019-05-24 西安交通大学 A kind of ToF depth transducer and its distance measuring method based on laser speckle projection

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
叶子瑞: "基于TOF相机的相对位姿测量方法", 《万方数据库》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109991583A (en) * 2019-03-14 2019-07-09 深圳奥比中光科技有限公司 A kind of jamproof distance measurement method and depth camera
CN109991584A (en) * 2019-03-14 2019-07-09 深圳奥比中光科技有限公司 A kind of jamproof distance measurement method and depth camera
CN110686600A (en) * 2019-11-05 2020-01-14 北京深测科技有限公司 Measuring method and system based on flight time measurement
CN110686600B (en) * 2019-11-05 2021-08-06 北京深测科技有限公司 Measuring method and system based on flight time measurement
CN112504125A (en) * 2020-12-15 2021-03-16 广东工业大学 Non-contact cuboid volume measurement method based on TOF
CN112504125B (en) * 2020-12-15 2022-03-25 广东工业大学 Non-contact cuboid volume measurement method based on TOF
CN113418467A (en) * 2021-06-16 2021-09-21 厦门硅谷动能信息技术有限公司 Method for detecting general and black luggage size based on ToF point cloud data
CN113671519A (en) * 2021-10-22 2021-11-19 武汉市聚芯微电子有限责任公司 Time-of-flight ranging method and device, time-of-flight camera and electronic equipment

Also Published As

Publication number Publication date
CN108759666B (en) 2020-06-23

Similar Documents

Publication Publication Date Title
CN108759666A (en) A kind of dimension measurement method based on flight time three-dimensional camera
CN110942449B (en) Vehicle detection method based on laser and vision fusion
US7660436B2 (en) Stereo-vision based imminent collision detection
Pantilie et al. SORT-SGM: Subpixel optimized real-time semiglobal matching for intelligent vehicles
Barrois et al. 3D pose estimation of vehicles using a stereo camera
CN109263637B (en) Collision prediction method and device
CN107796373B (en) Distance measurement method based on monocular vision of front vehicle driven by lane plane geometric model
WO2016193716A1 (en) A computer implemented method of detecting the distance of an object from an image sensor
CN103996202A (en) Stereo matching method based on hybrid matching cost and adaptive window
CN105890542B (en) A kind of motor vehicle contour dimension measuring system and method
Pinggera et al. High-performance long range obstacle detection using stereo vision
CN114758504B (en) Online vehicle overspeed early warning method and system based on filtering correction
CN108305277A (en) A kind of heterologous image matching method based on straightway
Shivakumar et al. Real time dense depth estimation by fusing stereo with sparse depth measurements
CN113989758A (en) Anchor guide 3D target detection method and device for automatic driving
CN116109601A (en) Real-time target detection method based on three-dimensional laser radar point cloud
CN115187955A (en) Target positioning method based on post-fusion of visual image and sparse point cloud
CN114137562A (en) Multi-target tracking method based on improved global nearest neighbor
CN113838069A (en) Point cloud segmentation method and system based on flatness constraint
CN110702015B (en) Method and device for measuring icing thickness of power transmission line
CN117058236A (en) Target identification positioning method based on multi-vision system self-switching
Cigla et al. Image-based visual perception and representation for collision avoidance
RU2650857C1 (en) System for determining geometric parameters of three-dimensional objects
CN109785261A (en) A kind of airborne LIDAR three-dimensional filtering method based on gray scale volume element model
Lim et al. Integration of Vehicle Detection and Distance Estimation using Stereo Vision for Real-Time AEB System.

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20200623

CF01 Termination of patent right due to non-payment of annual fee