CN108592822A - A kind of measuring system and method based on binocular camera and structure light encoding and decoding - Google Patents

A kind of measuring system and method based on binocular camera and structure light encoding and decoding Download PDF

Info

Publication number
CN108592822A
CN108592822A CN201810424252.5A CN201810424252A CN108592822A CN 108592822 A CN108592822 A CN 108592822A CN 201810424252 A CN201810424252 A CN 201810424252A CN 108592822 A CN108592822 A CN 108592822A
Authority
CN
China
Prior art keywords
striped
gray scale
color
decoding
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201810424252.5A
Other languages
Chinese (zh)
Other versions
CN108592822B (en
Inventor
余洪山
夏长得
冯云
关佳
胡华虎
武彦飞
朱江
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan University
Original Assignee
Hunan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan University filed Critical Hunan University
Publication of CN108592822A publication Critical patent/CN108592822A/en
Application granted granted Critical
Publication of CN108592822B publication Critical patent/CN108592822B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01BMEASURING LENGTH, THICKNESS OR SIMILAR LINEAR DIMENSIONS; MEASURING ANGLES; MEASURING AREAS; MEASURING IRREGULARITIES OF SURFACES OR CONTOURS
    • G01B11/00Measuring arrangements characterised by the use of optical techniques
    • G01B11/24Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures
    • G01B11/25Measuring arrangements characterised by the use of optical techniques for measuring contours or curvatures by projecting a pattern, e.g. one or more lines, moiré fringes on the object

Abstract

The invention discloses a kind of measuring system and method based on binocular camera and structure light encoding and decoding, measuring system include control host, projecting apparatus and 2 cameras;Projecting apparatus and 2 cameras are controlled by control host;Projecting apparatus is located between 2 gray scale cameras;Control host forms structure fringe light and the image transmitting projected will be needed to projecting apparatus;It will be in structure striped light projection to the measured object as projection objects by projecting apparatus;2 gray scale cameras are arranged side by side to form binocular gray scale phase unit, and left and right projected image is obtained respectively by 2 gray scale cameras;And the left and right projected image of acquisition is sent to control host;Control host is additionally operable to be decoded left and right projected image, grating matching and disparity computation are to realize that the 3D vision to measured object is rebuild.The measuring system and method based on binocular camera and structure light encoding and decoding are easy to implement, and imaging and measuring speed are fast.

Description

A kind of measuring system and method based on binocular camera and structure light encoding and decoding
Technical field
The present invention relates to a kind of measuring system and method based on binocular camera and structure light encoding and decoding.
Background technology
Since nearly twenty or thirty year, with the rapid development of computer technology, the application of computer vision is from fixed Two dimensional surface has been transitioned into intuitive flexible stereoscopic three-dimensional.Produce the traditional two-dimensional measurement of supermatic industrial age and Three-dimensional touch measurement method cannot meet the needs of modern production, and therefore, the industrial production of modernization is non-to three-dimensional to be connect Touch measurement proposes urgent demand.
Optical three-dimensional measurement technology is a kind of non-contact measurement, has high certainty of measurement, easy to implement and can measure The advantages that object of surface complexity, so optical three-dimensional measurement technology has been widely used in vision guided navigation, industrial detection etc. Field.Structural light measurement technology is a kind of typical measuring method, be widely used in reverse-engineering, industrial detection, Artistic statue appearance is restored, the fields such as medical cosmetic beauty, with the continuous development of three-dimensional measurement technology, the requirement of measurement accuracy Also higher and higher.
In field of industry detection, for example, the vibrations of the blade, body of a motor car of turbine, aircraft wing wind tunnel test Caused deformation and vibrations etc. are required for the three-dimensional information of the vibrations to their critical component, deformation, movement to measure, point Its performance is analysed, to ensure its quality and safety.And we can also be timely anti-to the design and process of product progress Feedback, further increases its technical performance.In product design process, people also increasingly focus on customer-oriented personalized designs Such as shoes, clothes, glasses industry are more conducive to product individuality by the three-dimensional information locally or globally of acquisition human body The design and rapid processing of change.
Currently, in three-dimensional measurement field, most of measurement method is all based on testee surface characteristics, effect It is relatively good.But the detection result of the object less to testee surface texture-free or texture is undesirable.As industry is raw The high speed development of automation is produced, industrially increasingly for the detection demand of the less object of body surface texture-free or texture Greatly, but current three dimensional detection technology can not fully meet its demand.
Therefore, it is necessary to design a kind of measuring system and method based on binocular camera and structure light encoding and decoding.
Invention content
Technical problem to be solved by the invention is to provide a kind of measurement system based on binocular camera and structure light encoding and decoding System and method, measuring system and method that should be based on binocular camera and structure light encoding and decoding be easy to implement, imaging and measuring speed Soon.
The technical solution of invention is as follows:
A kind of measuring system based on binocular camera and structure light encoding and decoding, including control host, projecting apparatus and 2 phases Machine;Projecting apparatus and 2 cameras are controlled by control host;Projecting apparatus is located between 2 gray scale cameras;
Control host forms structure fringe light and the image transmitting projected will be needed to projecting apparatus;By projecting apparatus by structure bar In line light projection to the measured object as projection objects;
2 gray scale cameras are arranged side by side to form binocular gray scale phase unit, obtain left and right projection respectively by 2 gray scale cameras Image;And the left and right projected image of acquisition is sent to control host;
Control host is additionally operable to be decoded left and right projected image, grating matching and disparity computation are to realize to quilt The 3D vision for surveying object is rebuild.
The structure fringe light is the non-repetitive De Bruijn sequences of 5 yuan of 3 rank;The sequence is:
{2,5,4,1,3,5,4,5,2,3,5,1,2,1,2,3,2,5,3,1,4,1,2,5,1,4,3,1,3,4,1,5,4,3, 2,1,4,2,1,3,2,4,5,4,2,5,2,4,2,4,1,4,5,1,5,1,3,1,2,4,3,5,2,1,5,3,2,3,4,3,4,5, 3,5,3,4,2,3,1,5}。
It is equipped with blackstreak between adjacent color fringe as interval, so that color fringe edge is more clear, enhances The anti-interference ability of fringe counting method;
2 color codeword look-up table of table
The correspondence of color and code word such as table 2.
Projecting apparatus is DLP projector, and gray scale camera is CCD camera or CMOS cameras
The camera is gray scale camera;Colored structures striations is decomposed, the ash in tri- channels corresponding R, G, B is obtained Spend striated structure light pattern;The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to respectively by projecting apparatus Testee surface.
Control host is equipped with the digital picture that 2 image pick-up cards are used to acquire the acquisition of 2 gray scale cameras.
A kind of measurement method based on binocular camera and structure light encoding and decoding is based on binocular camera and structure using above-mentioned The measuring system of light encoding and decoding;Its step is:
Step 1:Striped pumped FIR laser and generating structure striped light image are formed by control host;
Step 2:Projected fringe light image and acquisition left and right projected image;
Colored structures striations is decomposed, the gray scale striped structured light patterns in tri- channels corresponding R, G, B are obtained;
The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to measured object body surface respectively by projecting apparatus Face is used in combination 2 gray scale cameras to obtain projected image;
Control the gray scale striated structure in tri- channels R, G, B for being projected to body surface that host synthesis gray scale camera obtains Light pattern obtains the Three Channel Color structural light stripes image that 2 width contain measured object texture position information, is denoted as left and right perspective view Picture;
Step 3:Control host is respectively decoded left and right projected image;
Step 4:Grating matching and disparity computation are executed in control host.
In step 3, first left and right projected image is pre-processed;The pretreatment refers to using at medium filtering Reason.
Include the following steps in step 3:
Step 31:With canny operator extraction stripe edges, edge coordinate is preserved;
Step 32:Geometrical center method extraction algorithm is used to each striped of every row, obtains the geometric center lines of striped, i.e., Rough center line:Coordinate point set centered on geometric center lines;The coordinate point set that gray value in bianry image is 255;
Step 33:In rough center (certain point on i.e. rough center line) P of stripedcThe right and left at place respectively takes M pixel, so that width (the corresponding width of 2m+1 pixel) between them is more than the width of striped, (such as 1.1~1.5 times of stripeds are wide Degree);Gray threshold is determined using Adaptive Thresholding, background pixel is detached with stripe pixel;
Step 34:The Grad G (x, y) of stripe pixel is calculated according to formula 1;
For the gray threshold of striped boundary;
Step 35:The sub-pixel location P (x, y) of stripe pixel is calculated according to formula 2 mentioned above;
Step 36:The sub-pixel location of stripe centerline is calculated using gradient weight method extraction centerline algorithms formula 3, wherein G (x, y) is weight, and P (x, y) is the gradient locations of respective weights;
Because striped used is nicking, and it is to extract center line by row, so row coordinate is known, sub- picture here Plain location formula is to calculate row coordinate.
The ranks coordinate of sub-pixel location is deposited in Mat class images, the coordinate point set of intensity value non-zero points in image;
In step 3, the clustering algorithm based on HSI color spaces clusters the color of the striped in image, each color The range of corresponding tone value H is as shown in table 1:
1 HSI color spaces color of table table corresponding with tone angular region
Based on table 2, the corresponding code word of striped is obtained according to specific striped color.
Advantageous effect:
The present invention provides a kind of measuring system and method based on binocular camera and structure light encoding and decoding.
The coding method of the present invention, first passes through the color structured light candy strip that computer code goes out De-Bruijn sequences, And it isolates its BGR and is projected to testee surface respectively per channel gray scale fringe structure light;It is obtained again by gray scale camera One step is projected to the structure light on measured object surface, three width gray scales of the BGR triple channels for the body surface that gray scale camera is obtained Figure one width color structured light bar graph of synthesis, completes cataloged procedure.This coding method can significantly improve in later stage decoding process The accuracy rate and efficiency of the identification of the extraction striped of code word, increase the anti-interference ability and robustness of system.Generate color fringe Coding pattern meets following several conditions:
(1) color of two stripes of arbitrary neighborhood differs;
(2) any one subsequence being made of adjacent n stripes does not repeat in entire sequence, i.e., subsequence is whole It is unique in a sequence;
(3) color is not excessive when encoding scheme, and the discrimination of each color is very big, and guarantee projects testee surface Color fringe is not readily susceptible to the interference of testee surface color and ambient light and generates error code, influences the precision rebuild.
The above feature provides advantage for the extraction of the edge line or center line of subsequent stripe, especially in measured object Surface is discontinuously or in the case of depth information change dramatically.
In an encoding process, ensure that subsequence striped is encoded in whole picture using the uniqueness of De Bruijn window sequences Uniqueness in pattern, while ensuring that adjacent two color fringes differ by improving De Bruijn sequences.It is also logical simultaneously Crossing introducing blackstreak interval makes color fringe edge be more clear, and enhances the anti-interference ability of system.Although the encoding scheme Precision there is no the precision of space-time code method high, but improve the precision of three-dimensional reconstruction by the way that colouring information is added, increase The interference performance and robustness of system, realizes accurate fringe counting method and matches.
A width coloud coding pattern is generated according to De Bruijn sequences, each color stripe represents one in sequence Element, according to the window uniqueness of De Bruijn sequences it is found that the color fringe of some specific width combines in coding pattern All it is unique, so only needing correctly to extract the colouring information of striped and arrangement in corresponding child window in left images Sequence can complete the decoding of corresponding fringe position information.When being tested, due to by ambient light, the property of hardware device The influence of energy, measured object surface color and its series of factors such as the reflection on surface and absorption, camera get striped color very It is that the identification of subsequent color increases difficulty to be easy distortion.Therefore the present invention is also further devised is decomposed into ash by color fringe Spend the strategy of striped.
Present invention is alternatively directed to colored De Bruijn coding binocular vision systems to propose a set of efficiently accurate decoding side Method can find out the sub-pixel pixel coordinate of center line and the pixel matching of left images, to obtain the parallax of image Matrix.Canny operators detect edge, and there are one the process of inhibitions of non-maximum, carry out edge connection using two threshold values, have Good anti-interference ability.Since there are edge effects, i.e. high intensity striped to be deviated to low intensity background in actual photographed, So calculating fringe center line coordinates is used for disparity computation.The Adaptive Thresholding that the present invention uses can be effectively by striate region Domain is separated with background area, and the subpixel coordinate of stripe centerline can be calculated using gradient gravity model appoach.But it is sub- Pixel coordinate is extremely sensitive to noise, needs that constraints is arranged in conjunction with rough Pixel-level coordinate before, calculates in this way The sub-pixel center line come will be very accurate, also improve the anti-interference ability of system.The coloured image of acquisition is transformed into HIS Color space carries out color cluster, can overcome a system such as color, reflection and the absorption on illumination measured object surface in shooting process The interference of row factor is had modified to a certain extent between the color value of color fringe and standard value when projection of camera acquisition Deviation.Then rgb space is rotated back into again writes code word label to color fringe.Binocular camera can not almost be made in actual photographed Strictly parallel, this is obvious, and carrying out three-dimensional correction to image with epipolar-line constraint principle can make matching be converted into one Dimension space.Striped is traversed in same a line of left images and decodes code word label, can be realized using De Bruijn windows uniqueness Accurate matching between pixel and pixel, then the poor parallax that can be obtained the point is asked to corresponding left images pixel point coordinates.When Parallax matrix can be obtained by traversing all rows.
The binocular colored structures light measurement system of the present invention is by the way that binocular vision technology and structured light technique are tied It closing, the object less to surface texture-free or texture carries out three-dimensional reconstruction, and the system structure is fairly simple, and precision is higher, at As speed is fast.
Description of the drawings
Fig. 1 is binocular structured-light system overall construction drawing;
Fig. 2 is binocular Constructed Lighting Vision System schematic diagram;
Fig. 3 is the overview flow chart of the measurement method based on binocular camera and structure light encoding and decoding;
Fig. 4 is that (Fig. 4 (a) is the DeBruijn digraphs of 2 yuan of 2 rank to De Bruijn sequences digraph, is (b) 3 yuan of 2 rank DeBruijn digraphs)
Fig. 5 is 5 yuan of 3 grades of non-duplicate De Brujin code patterns;
Fig. 6 is structure light coding figure, and wherein Fig. 6 (a)-(d) is coloud coding bar graph, the channels R gray-scale map, G logical respectively Road gray-scale map and channel B gray-scale map.
Fig. 7 is workpiece projection schematic diagram;Wherein Fig. 7 (a) is the channels R gray scale striped capture figure, and Fig. 7 (b) is channel B gray scale Striped capture figure, Fig. 7 (c) are channel B gray scale striped capture figure, and Fig. 7 (d) is color fringe composite diagram;
Fig. 8 is structured light decoding flow chart;
Fig. 9 is geometrical center method flow chart;
Figure 10 is Adaptive Thresholding schematic diagram;
Figure 11 is fixed board sub-pix center line chart;
Figure 12 is workpiece sub-pix central line pick-up figure;
Figure 13 is color identification process figure;
Figure 14 is scaling board color center line chart;
Figure 15 is that color fringe matches schematic diagram.
Specific implementation mode
The present invention is described in further details below with reference to the drawings and specific embodiments:
Embodiment 1:Such as Fig. 1-3, a kind of measuring system based on binocular camera and structure light encoding and decoding, including control master Machine, projecting apparatus and 2 cameras;Projecting apparatus and 2 cameras are controlled by control host;Projecting apparatus is located between 2 gray scale cameras;
Control host forms structure fringe light and the image transmitting projected will be needed to projecting apparatus;By projecting apparatus by structure bar In line light projection to the measured object as projection objects;
2 gray scale cameras are arranged side by side to form binocular gray scale phase unit, obtain left and right projection respectively by 2 gray scale cameras Image;And the left and right projected image of acquisition is sent to control host;
Control host is additionally operable to be decoded left and right projected image, grating matching and disparity computation are to realize to quilt The 3D vision for surveying object is rebuild.
The structure fringe light is the non-repetitive De Bruijn sequences of 5 yuan of 3 rank;The sequence is:
{2,5,4,1,3,5,4,5,2,3,5,1,2,1,2,3,2,5,3,1,4,1,2,5,1,4,3,1,3,4,1,5,4,3, 2,1,4,2,1,3,2,4,5,4,2,5,2,4,2,4,1,4,5,1,5,1,3,1,2,4,3,5,2,1,5,3,2,3,4,3,4,5, 3,5,3,4,2,3,1,5}。
It is equipped with blackstreak between adjacent color fringe as interval, so that color fringe edge is more clear, enhances The anti-interference ability of fringe counting method;
2 color codeword look-up table of table
The correspondence of color and code word such as table 2.
Projecting apparatus is DLP projector, and gray scale camera is CCD camera or CMOS cameras
The camera is gray scale camera;Colored structures striations is decomposed, the ash in tri- channels corresponding R, G, B is obtained Spend striated structure light pattern;The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to respectively by projecting apparatus Testee surface.
Control host is equipped with the digital picture that 2 image pick-up cards are used to acquire the acquisition of 2 gray scale cameras.
A kind of measurement method based on binocular camera and structure light encoding and decoding is based on binocular camera and structure using above-mentioned The measuring system of light encoding and decoding;Its step is:
Step 1:Striped pumped FIR laser and generating structure striped light image are formed by control host;
Step 2:Projected fringe light image and acquisition left and right projected image;
Colored structures striations is decomposed, the gray scale striped structured light patterns in tri- channels corresponding R, G, B are obtained;
The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to measured object body surface respectively by projecting apparatus Face is used in combination 2 gray scale cameras to obtain projected image;
Control the gray scale striated structure in tri- channels R, G, B for being projected to body surface that host synthesis gray scale camera obtains Light pattern obtains the Three Channel Color structural light stripes image that 2 width contain measured object texture position information, is denoted as left and right perspective view Picture;
Step 3:Control host is respectively decoded left and right projected image;
Step 4:Grating matching and disparity computation are executed in control host.
To coding method, detailed description are as follows below:
A kind of structural light stripes coding method, if K=(1,2,3 ... k) be a k element set, n be one it is just whole Number, then the De Bruijn sequence definitions of k member n ranks are as follows:(di,di+1,di+2,…di+n-1) it is that n member is chosen from set K Element carries out the k of permutation and combination generationnA different subsequence carries out striped pumped FIR laser using non-repetitive De Bruijn sequences; Non-repetitive De Bruijn sequence maximum lengths are k* (k-1)n-1;A kind of corresponding color of each coding, different codings pair Answer different colors.Striped pumped FIR laser is used to form the strip encoding light projected on testee, then is obtained by two-sided camera Image realizes the three-dimensional measurement for measured object.
Alternative color have red (255,0,0), green (0,255,0), blue (0,0,255) yellow (255,255,0), Pinkish red (255,0,255), cyan (0,255,255);Numerical value wherein in color unquote is the corresponding rgb value of the color.It compiles Each color corresponds to a kind of code word in table 2 when code, as taken a computer capacity to know to the structural light stripes of each color Other name.
Hamiltonian circuit or Euler's circuit search are carried out in De Bruijns, the label sequence on side is just in each circuit Constitute the De Bruijn sequences of a k member n rank.Grid in figure constitutes node, and the representative of grid number is selected from set K Take the permutation and combination (number is the corresponding code word of various different colours fringe lights) that n element carries out, total node number kn.It breathes out Close circuit refers to from a node in figure, along the side in figure, just past each node in figure, and only By primary circuit.
Table 2a color codeword look-up tables
Based on table 2a, the corresponding code word of striped is obtained according to specific striped color.
Non-repetitive De Bruijn sequences are the sequence of the sequence of 2 yuan of 2 rank, the sequence of 3 yuan of 2 rank or 5 yuan of 3 rank.
Such as:2 yuan 2 grades of non-duplicate De Bruijn sequences (12), the non-duplicate De Bruijn sequences of 3 yuan of 2 rank (123132), two code words of arbitrary neighborhood do not repeat wherein in 2 rank sequence of ternary
Corresponding 5 yuan of 3 rank sequence be
{2,5,4,1,3,5,4,5,2,3,5,1,2,1,2,3,2,5,3,1,4,1,2,5,1,4,3,1,3,4,1,5,4,3, 2,1,4,2,1,3,2,4,5,4,2,5,2,4,2,4,1,4,5,1,5,1,3,1,2,4,3,5,2,1,5,3,2,3,4,3,4,5, 3,5,3,4,2,3,1,5}。
It is equipped with blackstreak between adjacent color fringe as interval, so that color fringe edge is more clear, enhances The anti-interference ability of fringe counting method;
2 color codeword look-up table of table
The correspondence of color and code word such as table 2.
A kind of projected image acquisition system, including control host, projecting apparatus and 2 gray scale cameras;Projecting apparatus and 2 gray scales Camera is controlled by control host;
The image that control host forms structure fringe light based on the structural light stripes coding method and projects needs It is transferred to projecting apparatus;It will be in structure striped light projection to projection objects by projecting apparatus;
2 gray scale cameras are arranged side by side to form binocular gray scale phase unit, obtain left and right projection respectively by 2 gray scale cameras Image;And the left and right projected image of acquisition is sent to control host and carries out subsequent processing.
Projecting apparatus is DLP projector, and gray scale camera is CCD camera or CMOS cameras.
A kind of projected image acquisition methods, which is characterized in that system is obtained using the projected image, feature exists In including the following steps:
Step 1:The colored structures striation of De-Bruijn sequences is encoded out by controlling host (such as PC machine or industrial personal computer) Line pattern;
Step 2:Colored structures striations is decomposed, the gray scale fringe structure light figure in tri- channels corresponding R, G, B is obtained Case;
Step 3:The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to measured object body surface respectively Face is used in combination 2 gray scale cameras to obtain projected image;
Step 4:Synthesize the gray scale fringe structure light in tri- channels R, G, B for being projected to body surface that gray scale camera obtains Pattern obtains the Three Channel Color structural light stripes image (i.e. left and right projected image) that 2 width contain measured object texture position information.
Such as Fig. 4, De-bruijn sequences are improved non-duplicate De-bruijn sequences, i.e., in decoding process if adjacent The identical striped of two colors has a missing, may will produce error hiding, and influence three-dimensional reconstruction as a result, in addition in order to meet The requirement of color fringe structure light coding needs to be improved traditional De Bruijn sequences.By to De Bruijn sequences The mathematical models of row and De Bruijns the study found that identical in order to eliminate two neighboring element in De Bruijn sequences Situation, it is only necessary to by De Bruijns G (n, K) containing the identical subsequence of adjacent element vertex and the vertex correspondence Side removes, and a modified De Bruijn G ' (n, k) is obtained, if there are hamiltonian circuit, circuit subscripts in G ' (n, K) Label sequence be exactly improved non-duplicate De Bruijn sequences, according to permutation and combination it is found that non-repetitive De Bruijn sequences most Long length is k* (k-1)n-1.Following Fig. 4 (a) is the De of 2 yuan 2 grades of non-duplicate De Bruijn sequences (12) after improving The circuit of Bruijn, solid line side composition is its corresponding hamiltonian circuit, and following Fig. 4 (b) is 3 yuan of 2 rank after improving The circuit of the De Bruijns of non-duplicate De Bruijn sequences (123132), solid line side composition is returned for its corresponding Hamilton Road.
The colored structures striations is eliminated the interference of ambient light and noise, is preferably distinguished when in order to make decoding Identify striped color, alternative color have red (255,0,0), green (0,255,0), blue (0,0,255) yellow (255, 255,0), pinkish red (255,0,255), cyan (0,255,255).
Coding/decoding method is described in detail below:
A kind of coding/decoding method based on binocular vision color fringe coding
A kind of coding/decoding method based on binocular vision color fringe coding, includes the following steps:
Step 1:Projected image is shot using binocular camera;
Projected image is 2 width, i.e., left image and has image;Step 2:The projected image of acquisition is pre-processed;
The pretreatment refers to being handled using medium filtering;
Step 3:Center line of the extraction per stripe;
Step 4:Identify the corresponding code word of striped color;
Step 5:Grating matching and disparity computation;
Wherein, step 2~5 are respectively handled 2 width images.
The step 3 includes the following steps:
Step 31:With canny operator extraction stripe edges, edge coordinate is preserved;
Step 32:Geometrical center method extraction algorithm is used to each striped of every row, obtains the geometric center lines of striped, i.e., Rough center line:Coordinate point set centered on geometric center lines;The coordinate point set that gray value in bianry image is 255;Step 33: In rough center (certain point on i.e. rough center line) P of stripedcThe right and left at place respectively takes m pixel, makes them Between width (the corresponding width of m+1 pixel) be more than striped width (such as 1.1~1.5 times of widths of fringe);Using adaptive Threshold method determines gray threshold, and background pixel is detached with stripe pixel;
Step 34:The Grad G (x, y) of stripe pixel is calculated according to formula 1;
For the gray threshold of striped boundary;
Step 35:The sub-pixel location P (x, y) of stripe pixel is calculated according to formula 2 mentioned above;
Step 36:The sub-pixel location of stripe centerline is calculated using gradient weight method extraction centerline algorithms formula 3, wherein G (x, y) is weight, and P (x, y) is the gradient locations of respective weights;
Because striped used is nicking, and it is to extract center line by row, so row coordinate is known, sub- picture here Plain location formula is to calculate row coordinate.
The ranks coordinate of sub-pixel location is deposited in Mat class images, the coordinate point set of intensity value non-zero points in image.
In step 4, the clustering algorithm based on HSI color spaces clusters the color of the striped in image, each color The range of corresponding tone value H is as shown in table 1:
1 HSI color spaces color of table table corresponding with tone angular region
2 color codeword look-up table of table
Based on table 2, the corresponding code word of striped is obtained according to specific striped color.
Step 5 includes the following steps:
Step 51:Referring to Figure 14,15, first by the center line L1 of left image (be when matching since Far Left, this In only take centre position illustratively matching process) (corresponding code word be 3) as matching starting striped, institute's bid section in figure Coded sequence is 35121 ...), need to find the adjacent n stripes on the right of the striped, composition target window sequence LP1 is (corresponding Series of windows be 351), since n stripes form a target window sequence, when matching be a window and a window into Row matching;Step 52:
It is carried out on the mutually colleague of the fringe center line chart of right image using the target window sequence of left image as template time It goes through, if finding match window sequence RP2 identical with target window sequence, that is, completes matching;
Step 53:
The difference for the center line corresponding points subpixel coordinates that left images have matched is the parallax of the point;
Traversal entire image has obtained the parallax matrix of corresponding left images.
Adaptive Thresholding in step 33 includes the following steps:
Step a:First calculate the P on xth row in image ci - m is arranged to P ci Gray value I's (x, y) within the scope of+m row Average value Ia
Step b:If IthFor the gray threshold of striped boundary, IthFor the P on striped ci - m is arranged to P ci Within the scope of+m row Gray value I (x, y) be less than average value IaPixel gray scale average value;Using IthCan be adaptive by fringe area It is separated with background area.
When seeking of central point is carried out, only the coordinate of (put to) is just it is to be understood that two, any one striped section or so endpoint It can obtain the coordinate of fringe center.The white wire that the edge line after edge detection is all single pixel is carried out by Canny operators Item, T1And T2For the desired width of color fringe and blackstreak, a complete point to the modulation by testee surface it Difference, that is, X of the ordinate of rear left right endpointR-XLRange be greater than 4/5T1Less than 6/5T1
Extraction stripe centerline is as follows:
Extraction stripe centerline is as follows:
1) P is setL(xL,yL) be striped section left end point, PR(xR,yR) be striped section right end point coordinates, extracting Using first marginal point of reading as left end point P when being traversed to certain level direction on the design sketch of edge lineL(xL, yL), second marginal point is as right endpoint PR(xR,yR);
If 2) 4/5T1≤yR-yL, then PL, PRIt is a complete point pair, the coordinate of central point is Pc (xc,yc).If yR-yL>=6/5T1, then right hand edge point missing, enables PL=PR, PRFor next edge point coordinates;
If 3) yR-yL≤ 4/5 T1, works as 4/5T2<yR-yL<6/5 T2 is true, then takes point P to the right1(x1,y1)、P2(x2, y2);
If 4) 4/5T1<yR-yL<6/5 T1 and 4/5T2<yR-yL<6/5T2 is true, then PL, PRIt is two of blackstreak Endpoint enables PL=PR, PRFor next marginal point, work as 4/5T2<yR-yL<6/5 T2 is false, continues to take next marginal point PR1 (xR1,yR1),
If 5) yR1-yL≤ 4/5T1, then enable PR=PR1, continue to take next marginal point PR1;If yR1-yL>=6/5T1, Then PL, PRFor complete edge pair, then it is further continued for being judged.
Central point Pc(xc, yc) coordinate calculation formula it is as follows:
Step 14:In the rough center P of stripedcThe right and left at place respectively takes m pixel, and width omits between making them More than the width of striped.After the range of light belt is determined, it is also necessary to which the accurate boundary for determining striped utilizes ladder to improve Spend the accuracy of the sub-pix center line of gravity model appoach extraction.The gray scale on striped boundary is determined there is employed herein Adaptive Thresholding Threshold value, the algorithm are as follows:
1) in image on xth row is first calculatedRow are extremelyThe average value of gray value I (x, y) within the scope of row Ia
2) it is set as the gray threshold of striped boundary, is the on stripedRow are extremelyGray value I within the scope of row (x, y) is less than average value IaPixel gray scale average value.The background gray scale of fringe area can be slightly larger than, it can be adaptive That answers is separated by fringe area and background area.
3) gradient herein refers to the exhausted of the difference result of the gray value of adjacent pixel on the line direction of image To value, if G (x, y) is the Grad of y row and y+1 row centre position on image xth row.
4) position corresponding to the Grad of striped boundary is calculated, it can be logical according to gray value I (x, y) and I (x, y+1) Linear interpolation is crossed to be calculated.The expression formula of position P (x, y) corresponding to Grad is as follows;
5) G (x, y) is weight in the method for extracting center line using gradient weight method, and P (x, y) is the gradient of respective weights Position, wherein PcFor rough position of center line, w is distance of the rough position of center line to striped boundary.To each traveling of image After the above-mentioned processing of row, so that it may each to can be obtained by obtain the sub-pixel location of the fringe center point on per a line The sub-pix center line of stripe.The calculation formula of the sub-pixel location of fringe center is as follows
Step 21:The standard value when color value for the color fringe that camera obtains and projection has certain deviation, so We need that these colors are identified and are handled, so as to will be used in the coding pattern of the color of every stripe and projection Five kinds of colors correspond.Present invention employs the clustering algorithms based on HSI color spaces to gather the color in image Class.The corresponding tone H of color of RGB color, saturation degree S, intensity I calculation formula as shown in formula:
It, only need to be to chrominance component H processing, without to brightness I processing, especially when being clustered to color To have shade or uneven illumination it is even coloured image it is especially effective.The striped color selected herein is all that RGB color model is vertical The color on the vertex in cube, including black and red.By calculation formula above it is found that black and red tone value All it is 0, so we can not only be judged by tone value H.When judging red and black, introduce bright Component I is spent, as H=0, brightness I is more than threshold value th, then is red, is otherwise black, the th=25 chosen herein.It is other Color is judged according to its corresponding tone value, and is set to corresponding projected color, and following table is that each color is corresponding The range of tone value H.
1 HSI color spaces color of table table corresponding with tone angular region
Step 22:Red (255,0,0) that the present invention selects, green (0,255,0), yellow (255,255,0), pinkish red Five kinds of (255,0,255), cyan (0,255,255) colors are all the colors on five vertex of square in RGB color model, often Value on three channels of kind color is all 255 or 0.The corresponding code word of each color is as shown in the table.
2 color codeword look-up table of table
According to the colouring information of reading, color is identified, and then completes the identification of code word.In actual processing, If the value in channel is less than 50, by it as 0 processing, it is more than 200, then presses 255 processing.
Step 23:Grating matching in the present invention is based on two criterion:Epipolar-line constraint criterion and De Bruijn window sequences Uniqueness criterion.Epipolar-line constraint criterion can ensure that the row of two images is alignment, then the matching based on window uniqueness It can be carried out on the correspondence row of left images.
1) it regard the center line L1 of left image (corresponding code word is 1) as target striped first, we select when due to coding What is taken is 5 yuan 3 grades of De Bruijn sequences, so needing to find the two adjacent stripes on the right of it, forms target window sequence Arrange LP1 (corresponding series of windows is 121);
2) it is carried out on the mutually colleague of the fringe center line chart of right image using the target window sequence of left image as template Traversal, can find match window sequence RP2 identical with target window sequence soon;
3) difference of left images corresponding points subpixel coordinates is the parallax of the point;
Due to obtain color fringe image, striped have it is spatial ordering, will not become because of the pattern on testee surface Change and striped sequence is made to cause confusion, so after LP1 successful match, the matching of next target window sequence LP2 is from item The next column of line R2 columns is traversed.The traversal range for reducing target sequence in this way, improves the matching efficiency of image With matched accuracy, traversal entire image has just obtained parallax matrix.
Input is the projected image of binocular camera shooting, exports the left and right camera for image to parallax matrix:
A kind of coding/decoding method being encoded binocular vision based on colored De Bruijn, is included the following steps:
Step 1:Binocular camera shoots projected image;
Two cameras need suitably to rotate inward an angle (10 ° or so), keep pair in two camera horizontal directions Claim, testee needs to be placed in the public view field of binocular camera;
Step 2:The original image of acquisition is pre-processed, noise is removed with medium filtering;
Step 3:Extract center line;
It is as follows:
Step 31:With canny operator extraction stripe edges, edge coordinate is preserved;
Step 32:Since known edge coordinate, using improved geometrical center method extraction algorithm proposed above, such as scheme Shown in 3, so that it may to obtain the geometric center lines of striped, i.e. coarse filtration center line;
Step 33:In the rough center P of stripedcThe right and left at place respectively takes m pixel, and width omits between making them More than the width of striped.After the range of light belt is determined, it is also necessary to which the accurate boundary for determining striped utilizes ladder to improve Spend the accuracy of the sub-pix center line of gravity model appoach extraction.Gray threshold is determined using Adaptive Thresholding, by background pixel with Stripe pixel is discernable,;
Step 34:The Grad G (x, y) of stripe pixel is calculated according to formula mentioned above (1);
Step 35:The sub-pixel location P (x, y) of stripe pixel is calculated according to formula mentioned above (2);
Step 36:The sub-pixel location of stripe centerline is calculated using gradient weight method extraction centerline algorithms formula (3), Middle G (x, y) is weight, and P (x, y) is the gradient locations of respective weights;
Step 4:Identify the code word of striped color:
RGB image is transformed into HSI color spaces using formula (4), the color of striped, then profit are judged using Hue values With threshold value to the value of the endowed word of striped color;
Step 5:Grating matching and disparity computation:
Grating matching is based on two criterion:Epipolar-line constraint criterion and De Bruijn window sequence uniqueness criterions.Polar curve is about Beam criterion can ensure that the row of two images is alignment, then the matching based on window uniqueness can be in left images It is carried out on corresponding row.
Step 51:
It regard the center line L1 of left image (corresponding code word is 1) as target striped first, we choose when due to coding Be 5 yuan 3 grades of De Bruijn sequences, so need to find the two adjacent stripes on the right of it, form target window sequence LP1 (corresponding series of windows is 121);
Step 52:
It is carried out on the mutually colleague of the fringe center line chart of right image using the target window sequence of left image as template time It goes through, match window sequence RP2 identical with target window sequence can be found soon;
Step 53:
The difference of left images corresponding points subpixel coordinates is the parallax of the point.

Claims (10)

1. a kind of measuring system based on binocular camera and structure light encoding and decoding, which is characterized in that including control host, projecting apparatus With 2 cameras;Projecting apparatus and 2 cameras are controlled by control host;Projecting apparatus is located between 2 gray scale cameras;
Control host forms structure fringe light and the image transmitting projected will be needed to projecting apparatus;By projecting apparatus by structure fringe light It projects on the measured object as projection objects;
2 gray scale cameras are arranged side by side to form binocular gray scale phase unit, and left and right perspective view is obtained respectively by 2 gray scale cameras Picture;And the left and right projected image of acquisition is sent to control host;
Control host is additionally operable to be decoded left and right projected image, grating matching and disparity computation are to realize to measured object 3D vision rebuild.
2. the measuring system according to claim 1 based on binocular camera and structure light encoding and decoding, the structure striped Light is the non-repetitive De Bruijn sequences of 5 yuan of 3 rank;The sequence is:
{2,5,4,1,3,5,4,5,2,3,5,1,2,1,2,3,2,5,3,1,4,1,2,5,1,4,3,1,3,4,1,5,4,3,2,1, 4,2,1,3,2,4,5,4,2,5,2,4,2,4,1,4,5,1,5,1,3,1,2,4,3,5,2,1,5,3,2,3,4,3,4,5,3,5, 3,4,2,3,1,5}。
3. the measuring system according to claim 2 based on binocular camera and structure light encoding and decoding, in adjacent color-bar It is equipped with blackstreak between line as interval, so that color fringe edge is more clear, enhances the anti-interference ability of fringe counting method;
2 color codeword look-up table of table
The correspondence of color and code word such as table 2.
4. the measuring system according to claim 1 based on binocular camera and structure light encoding and decoding, projecting apparatus projects for DLP Instrument, gray scale camera are CCD camera or CMOS cameras.
5. the measuring system according to claim 1 based on binocular camera and structure light encoding and decoding, which is characterized in that described Camera be gray scale camera;Colored structures striations is decomposed, the gray scale fringe structure light in tri- channels corresponding R, G, B is obtained Pattern;The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to testee surface respectively by projecting apparatus.
6. the measuring system according to claim 1 based on binocular camera and structure light encoding and decoding, which is characterized in that control Host is equipped with the digital picture that 2 image pick-up cards are used to acquire the acquisition of 2 gray scale cameras.
7. a kind of measurement method based on binocular camera and structure light encoding and decoding, which is characterized in that any using claim 1-7 The measuring system based on binocular camera and structure light encoding and decoding described in;Its step is:
Step 1:Striped pumped FIR laser and generating structure striped light image are formed by control host;
Step 2:Projected fringe light image and acquisition left and right projected image;
Colored structures striations is decomposed, the gray scale striped structured light patterns in tri- channels corresponding R, G, B are obtained;
The gray scale striped structured light patterns in obtained tri- channels R, G, B are projected to testee surface respectively by projecting apparatus, It is used in combination 2 gray scale cameras to obtain projected image;
Control the gray scale fringe structure light figure in tri- channels R, G, B for being projected to body surface that host synthesis gray scale camera obtains Case obtains the Three Channel Color structural light stripes image that 2 width contain measured object texture position information, is denoted as left and right projected image;
Step 3:Control host is respectively decoded left and right projected image;
Step 4:Grating matching and disparity computation are executed in control host.
8. the measurement method according to claim 7 based on binocular camera and structure light encoding and decoding, which is characterized in that step In 3, first left and right projected image is pre-processed;The pretreatment refers to being handled using medium filtering.
9. the measurement method according to claim 8 based on binocular camera and structure light encoding and decoding, which is characterized in that step Include the following steps in 3:
Step 31:With canny operator extraction stripe edges, edge coordinate is preserved;
Step 32:Geometrical center method extraction algorithm is used to each striped of every row, obtains the geometric center lines of striped, i.e., roughly Center line:Coordinate point set centered on geometric center lines;
Step 33:In rough center (certain point on i.e. rough center line) P of stripedcThe right and left at place respectively takes m picture Element makes width between them be more than the width of striped;Gray threshold is determined using Adaptive Thresholding, by background pixel and striped Pixel separation;
Step 34:The Grad G (x, y) of stripe pixel is calculated according to formula 1;
For the gray threshold of striped boundary;
Step 35:The sub-pixel location P (x, y) of stripe pixel is calculated according to formula 2 mentioned above;
Step 36:Calculate the sub-pixel location of stripe centerline using gradient weight method extraction centerline algorithms formula 3, wherein G (x, Y) it is weight, P (x, y) is the gradient locations of respective weights;
10. the measurement method according to claim 9 based on binocular camera and structure light encoding and decoding, which is characterized in that step In rapid 3, the clustering algorithm based on HSI color spaces clusters the color of the striped in image, the corresponding tone of each color The range of value H is as shown in table 1:
1 HSI color spaces color of table table corresponding with tone angular region
Based on table 2, the corresponding code word of striped is obtained according to specific striped color.
CN201810424252.5A 2017-12-04 2018-05-07 Measuring system and method based on binocular camera and structured light coding and decoding Active CN108592822B (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN2017112612182 2017-12-04
CN201711261218 2017-12-04

Publications (2)

Publication Number Publication Date
CN108592822A true CN108592822A (en) 2018-09-28
CN108592822B CN108592822B (en) 2020-01-07

Family

ID=63621061

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810424252.5A Active CN108592822B (en) 2017-12-04 2018-05-07 Measuring system and method based on binocular camera and structured light coding and decoding

Country Status (1)

Country Link
CN (1) CN108592822B (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109993826A (en) * 2019-03-26 2019-07-09 中国科学院深圳先进技术研究院 A kind of structural light three-dimensional image rebuilding method, equipment and system
CN110440712A (en) * 2019-08-26 2019-11-12 英特维科技(苏州)有限公司 Adaptive big depth of field 3-D scanning method and system
CN110686599A (en) * 2019-10-31 2020-01-14 中国科学院自动化研究所 Three-dimensional measurement method, system and device based on colored Gray code structured light
CN111023999A (en) * 2019-12-26 2020-04-17 北京交通大学 Dense point cloud generation method based on spatial coding structured light
CN111508012A (en) * 2019-01-31 2020-08-07 先临三维科技股份有限公司 Method and device for line stripe mismatching detection and three-dimensional reconstruction
WO2020186825A1 (en) * 2019-03-15 2020-09-24 上海图漾信息科技有限公司 Depth data measuring head, measurement device and measurement method
TWI722886B (en) * 2020-04-30 2021-03-21 國立陽明交通大學 Rotary code disk and method for designing the same
WO2022021678A1 (en) * 2020-07-31 2022-02-03 中国科学院深圳先进技术研究院 Three-dimensional object reconstruction method and terminal device
CN114688994A (en) * 2020-12-31 2022-07-01 先临三维科技股份有限公司 Color projection module, stripe decoding method, apparatus, medium, device, and system
CN111508012B (en) * 2019-01-31 2024-04-19 先临三维科技股份有限公司 Method and device for line stripe mismatching detection and three-dimensional reconstruction

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101082481A (en) * 2007-07-16 2007-12-05 北京航空航天大学 Colorful encode grating visible sensation measurement method based on phase displacement
CN102445165A (en) * 2011-08-05 2012-05-09 南京航空航天大学 Stereo vision measurement method based on single-frame color coding grating
CN103292741A (en) * 2013-05-29 2013-09-11 哈尔滨工程大学 Structured light vision measurement method for 3D surface profiles of objects on the basis of K-means color clustering
US20150145959A1 (en) * 2013-11-26 2015-05-28 Rowan University Use of Spatially Structured Light for Dynamic Three Dimensional Reconstruction and Reality Augmentation
CN106251376A (en) * 2016-08-12 2016-12-21 南京航空航天大学 A kind of towards colored structures pumped FIR laser and edge extracting method

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101082481A (en) * 2007-07-16 2007-12-05 北京航空航天大学 Colorful encode grating visible sensation measurement method based on phase displacement
CN102445165A (en) * 2011-08-05 2012-05-09 南京航空航天大学 Stereo vision measurement method based on single-frame color coding grating
CN103292741A (en) * 2013-05-29 2013-09-11 哈尔滨工程大学 Structured light vision measurement method for 3D surface profiles of objects on the basis of K-means color clustering
US20150145959A1 (en) * 2013-11-26 2015-05-28 Rowan University Use of Spatially Structured Light for Dynamic Three Dimensional Reconstruction and Reality Augmentation
CN106251376A (en) * 2016-08-12 2016-12-21 南京航空航天大学 A kind of towards colored structures pumped FIR laser and edge extracting method

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
LI ZHANG, ET AL.: ""Rapid Shape Acquisition Using Color Structured Light and Multi-pass Dynamic Programming"", 《PROCEEDINGS OF THE FIRST INTERNATIONAL SYMPOSIUM ON 3D DATA PROCESSING VISUALIZATION AND TRANSMISSION》 *
SANDRO BARONE, ET AL.: ""A Coded Structured Light System Based on Primary Color Stripe Projection and Monochrome Imaging"", 《SENSORS》 *
郭富强: ""基于结构光的颅骨部位三维重建方法研究"", 《中国优秀硕士学位论文全文数据库 信息科技辑》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111508012A (en) * 2019-01-31 2020-08-07 先临三维科技股份有限公司 Method and device for line stripe mismatching detection and three-dimensional reconstruction
CN111508012B (en) * 2019-01-31 2024-04-19 先临三维科技股份有限公司 Method and device for line stripe mismatching detection and three-dimensional reconstruction
WO2020186825A1 (en) * 2019-03-15 2020-09-24 上海图漾信息科技有限公司 Depth data measuring head, measurement device and measurement method
CN109993826B (en) * 2019-03-26 2023-06-13 中国科学院深圳先进技术研究院 Structured light three-dimensional image reconstruction method, device and system
CN109993826A (en) * 2019-03-26 2019-07-09 中国科学院深圳先进技术研究院 A kind of structural light three-dimensional image rebuilding method, equipment and system
CN110440712A (en) * 2019-08-26 2019-11-12 英特维科技(苏州)有限公司 Adaptive big depth of field 3-D scanning method and system
CN110686599A (en) * 2019-10-31 2020-01-14 中国科学院自动化研究所 Three-dimensional measurement method, system and device based on colored Gray code structured light
CN110686599B (en) * 2019-10-31 2020-07-03 中国科学院自动化研究所 Three-dimensional measurement method, system and device based on colored Gray code structured light
CN111023999A (en) * 2019-12-26 2020-04-17 北京交通大学 Dense point cloud generation method based on spatial coding structured light
CN111023999B (en) * 2019-12-26 2020-12-01 北京交通大学 Dense point cloud generation method based on spatial coding structured light
TWI722886B (en) * 2020-04-30 2021-03-21 國立陽明交通大學 Rotary code disk and method for designing the same
WO2022021678A1 (en) * 2020-07-31 2022-02-03 中国科学院深圳先进技术研究院 Three-dimensional object reconstruction method and terminal device
CN114688994A (en) * 2020-12-31 2022-07-01 先临三维科技股份有限公司 Color projection module, stripe decoding method, apparatus, medium, device, and system
CN114688994B (en) * 2020-12-31 2023-03-14 先临三维科技股份有限公司 Color projection module, stripe decoding method, device, medium, equipment and system

Also Published As

Publication number Publication date
CN108592822B (en) 2020-01-07

Similar Documents

Publication Publication Date Title
CN108592822A (en) A kind of measuring system and method based on binocular camera and structure light encoding and decoding
CN108592823A (en) A kind of coding/decoding method based on binocular vision color fringe coding
CN107945268B (en) A kind of high-precision three-dimensional method for reconstructing and system based on binary area-structure light
CN110686599B (en) Three-dimensional measurement method, system and device based on colored Gray code structured light
Kawasaki et al. Dynamic scene shape reconstruction using a single structured light pattern
CN104077804B (en) A kind of method based on multi-frame video picture construction three-dimensional face model
US7430312B2 (en) Creating 3D images of objects by illuminating with infrared patterns
CN101667303B (en) Three-dimensional reconstruction method based on coding structured light
CN103292741B (en) A kind of structured light vision method for measuring of the object surface tri-dimensional profile based on k average color cluster
CN105844633B (en) Single frames structure optical depth acquisition methods based on De sequence and phase code
CN106155299B (en) A kind of pair of smart machine carries out the method and device of gesture control
Starck et al. The multiple-camera 3-d production studio
Anderson et al. Color photometric stereo for multicolored surfaces
CN103562676B (en) Method and 3D scanner of using structured lighting
CN106780726A (en) The dynamic non-rigid three-dimensional digital method of fusion RGB D cameras and colored stereo photometry
CN108122254A (en) Three-dimensional image reconstruction method, device and storage medium based on structure light
CN107850782A (en) Represent that strengthening depth map represents with reflectance map
CN108225216A (en) Structured-light system scaling method and device, structured-light system and mobile equipment
CN107734267A (en) Image processing method and device
CN101482398B (en) Fast three-dimensional appearance measuring method and device
Petković et al. Single-shot dense 3D reconstruction using self-equalizing De Bruijn sequence
CN106251376B (en) One kind is towards colored structures pumped FIR laser and edge extracting method
Klaudiny et al. High-detail 3D capture and non-sequential alignment of facial performance
CN109191562A (en) Three-dimensional rebuilding method based on colored false random coding structure light
CN109903377A (en) A kind of three-dimensional face modeling method and system without phase unwrapping

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant