CN101794459A - Seamless integration method of stereoscopic vision image and three-dimensional virtual object - Google Patents

Seamless integration method of stereoscopic vision image and three-dimensional virtual object Download PDF

Info

Publication number
CN101794459A
CN101794459A CN201010110440A CN201010110440A CN101794459A CN 101794459 A CN101794459 A CN 101794459A CN 201010110440 A CN201010110440 A CN 201010110440A CN 201010110440 A CN201010110440 A CN 201010110440A CN 101794459 A CN101794459 A CN 101794459A
Authority
CN
China
Prior art keywords
virtual
model
image
stereoscopic vision
virtual image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201010110440A
Other languages
Chinese (zh)
Inventor
王晨升
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Posts and Telecommunications
Original Assignee
Beijing University of Posts and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Posts and Telecommunications filed Critical Beijing University of Posts and Telecommunications
Priority to CN201010110440A priority Critical patent/CN101794459A/en
Publication of CN101794459A publication Critical patent/CN101794459A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Processing Or Creating Images (AREA)

Abstract

The invention discloses a seamless integration method of a stereoscopic vision image and a three-dimensional virtual object, which comprises the following steps: 1) acquiring a practical stereoscopic vision image; 2) carrying out feature extraction and match on the acquired stereoscopic vision image and a virtual image model; 3) determining pose of the matched virtual image model by matching score calculation; 4) adjusting scale of the matched virtual image model; and 5) replacing a real object in the operation site with the completed matched virtual image model. The seamless integration method of the stereoscopic vision image and the three-dimensional virtual object realizes the seamless integration of the stereoscopic image in real environment and the three-dimensional virtual object and is required to integrate the gestures of stereoscopic image in teleoperation environment with those of the three-dimensional virtual object and a virtual robot, and is favor of breaking through the bottleneck that the teleoperation depends on two dimension images for a long time and has poor operation positioning accuracy and provides virtual simulation environment with deep immersion for teleoperation and control.

Description

The seamless integration method of a kind of stereoscopic vision image and three-dimensional virtual object
Technical field
The present invention relates to the teleoperation of robot field, refer to be used for the stereoscopic vision image of robot manipulation location and the seamless integration method of three-dimensional virtual object especially.
Background technology
The distant operation control of existing robots depends on bidimensional image for a long time, and its 3-dimensional image gap of comparing in the actual environment is quite big, therefore adopts bidimensional image that teleoperation of robot is positioned all the time, and ubiquity the inaccurate problem in operation location.
Realization is based on the stereopsis of true environment and the seamless fusion of three-dimensional target object, need be with the stereopsis of distant operation scenario and virtual three-dimensional object, virtual robot attitude dynamic fusion together.The purpose of doing like this is to help to break through the bottleneck that distant operation control depends on bidimensional image, operation bearing accuracy difference for a long time, and the virtual emulation environment of degree of depth feeling of immersion is provided for distant operation control.
Summary of the invention
At the problem that prior art exists, the invention provides and a kind ofly robot is shaken operation control the stereoscopic vision image of virtual emulation environment and the seamless integration method of three-dimensional virtual object are provided.
For achieving the above object, the seamless integration method of stereoscopic vision image of the present invention and three-dimensional virtual object is specially: 1) gather actual stereoscopic vision image; 2) according to the characteristic information extraction algorithm to collecting to such an extent that stereoscopic vision image and virtual image model carry out feature extraction, and then carry out the coupling of stereoscopic vision image and virtual image model; 3), mate score value and calculate, to determine the pose of matching virtual iconic model by the virtual image model after the rotation coupling; 4) the yardstick adjustment of matching virtual iconic model; 5) will mate completely virtual image model generation puts the real object object of distant operation site.
Further, described virtual image model is stored in the model bank, and can call by different instruction, the access model information.
Further, the concrete steps that the pose of virtual image model is determined in the step 3) are: the experience threshold values of setting up the pose coupling, virtual image model of every rotation, by calculating actual threshold values, when actual threshold values is lower than the experience threshold values, the pose of assert virtual image model and stereoscopic vision image is inconsistent, continue rotation virtual image model, when actual threshold values is higher than the experience threshold values, this pose is noted, and it is consistent with the target object pose in the stereoscopic vision image that statistics obtains the virtual image model pose of actual threshold values when the highest at last.
Further, the yardstick adjustment of virtual image model need be carried out convergent-divergent to the virtual image model in the step 4) on dimension, the ratio of convergent-divergent, is adjusted as coefficient by the scale ratio of the containing box of virtual image model under the containing box of target object in the real scene image and the identical pose.
Further, in the step 5) in the real scene virtual image model of yardstick such as target object be taken back in the real scene, need carry out being tied to the conversion of camera coordinates system from world coordinates.
The seamless integration method of stereoscopic vision image of the present invention and three-dimensional virtual object, realization is based on the stereopsis of true environment and the seamless fusion of three-dimensional target object, need be with the stereopsis of distant operation scenario and virtual three-dimensional object, virtual robot attitude dynamic fusion together.The purpose of doing like this is to help to break through the bottleneck that distant operation control depends on bidimensional image, operation bearing accuracy difference for a long time, and the virtual emulation environment of degree of depth feeling of immersion is provided for distant operation control.
Description of drawings
Fig. 1 is the process flow diagram of fusion method of the present invention;
Fig. 2 is a virtual image model pose calculation flow chart;
Fig. 3 is the binocular solid imaging schematic diagram;
Fig. 4 a and Fig. 4 b are the actual situation fusion exemplary plot based on the two-way video.
Embodiment
As shown in Figure 1, the seamless integration method of stereoscopic vision image of the present invention and three-dimensional virtual object is specially:
1. choosing alternately of target object:
Choosing of target object can be carried out in the video image of the left and right sides, and concrete operations are chosen similar to screen area.
2. the aspect of model extracts and the coupling retrieval:
Utilize the characteristic information extraction algorithm to carry out the model that feature extraction reaches and the model knowledge base is interior and carry out feature extraction and coupling.
Feature Extraction is relevant with choosing target object residing angle in video image, and reasonable angle should be the angle that can represent object singularity to greatest extent.When carrying out characteristic matching, system adopts the method for incremental rotation model to guarantee the accuracy of mating.
Feature extraction SIFT algorithm detailed step:
(1) generation of metric space
The theoretical purpose of metric space is the multiple dimensioned feature of simulated image data.
Gaussian convolution nuclear is the unique linear kernel that realizes change of scale, so the metric space of a secondary two dimensional image is defined as:
L(x,y,σ)=G(x,yσ)*I(x,y) (1)
Wherein G (x, y σ) are the changeable scale Gaussian function,
Figure GSA00000030358900031
(x y) is volume coordinate, and σ is the yardstick coordinate.
In order effectively to detect stable key point, difference of Gaussian metric space (DOG scale-space) has been proposed at metric space.Utilize the Gaussian difference pyrene and the image convolution of different scale to generate.
D(x,y,σ)=(G(x,y,kσ)-G(x,y,σ))*I(x,y)=L(x,y,kσ)-L(x,y,σ) (3)
The DOG operator calculates simple, is the approximate of the normalized LoG operator of yardstick.
(2) make up the parameter that metric space need be determined
σ-metric space coordinate
The O-octave coordinate
The S-sub-level coordinate
σ and O, S concern σ (o, s)=σ 02 O+s/S, o ∈ o Min+ [0 ..., O-1], s ∈ [0 ..., S-1]
σ wherein 0It is the key horizon yardstick.The o-octave coordinate, the s-sub-level coordinate.Annotate: the index of octaves may be born.First group index usually is made as 0 or-1, and when being made as-1, image expansion earlier before calculating Gauss's metric space is twice.
Volume coordinate x is the function of group octave, establishes x 0Be 0 group volume coordinate, then
x=2 ox 0,o∈Z,x 0∈[0,...,N 0-1]×[0,...,M 0-1]
If (M 0, N 0) be the resolution of base set o=0, then the resolution of other groups is obtained by following formula:
Figure GSA00000030358900033
(3) accurately determine the extreme point position
By fitting three-dimensional quadratic function accurately to determine the position and the yardstick (reaching sub-pixel precision) of key point, remove the key point and the unsettled skirt response point (because the DoG operator can produce stronger skirt response) of low contrast simultaneously, to strengthen coupling stability, to improve noise resisting ability.
The removal of skirt response
An extreme value that defines bad difference of Gaussian operator has bigger principal curvatures in the place across the edge, and in the direction of vertical edge less principal curvatures is arranged.Principal curvatures is obtained by the Hessian matrix H of a 2x2:
H = D xx D xy D xy D yy - - - ( 4 )
Derivative is estimated to obtain by the adjacent difference of sampled point.
The principal curvatures of D and the eigenwert of H are directly proportional, and make that α is an eigenvalue of maximum, and β is minimum eigenwert, then
Tr(H)=D xx+D yy=α+β,
Det(H)=D xxD yy-(D xy) 2=αβ.
Make α=γ β, then:
Tr ( H ) 2 Det ( H ) = ( α + β ) 2 αβ = ( rβ + β ) 2 r β 2 = ( r + 1 ) 2 r ,
(r+1) 2The value of/r is minimum when two eigenwerts equate, increases along with the increase of r, therefore, in order to detect principal curvatures whether under certain thresholding r, only needs to detect
Tr ( H ) 2 Det ( H ) < ( r + 1 ) 2 r .
(4) the key point direction is distributed
Utilize the gradient direction distribution character of key point neighborhood territory pixel to be each key point assigned direction parameter, make operator possess rotational invariance.
m ( x , y ) = ( L ( x + 1 , y ) - L ( x - 1 , y ) ) 2 + ( L ( x , y + 1 ) - L ( x , y - 1 ) ) 2
θ(x,y)=αtan2((L(x,y+1)-L(x,y-1))/(L(x+1,y)-L(x-1,y))) (5)
Formula (5) is that (x y) locates the mould value and the direction formula of gradient.Wherein the used yardstick of L is each key point yardstick at place separately.
When actual computation, we sample in the neighborhood window that with the key point is the center, and with the gradient direction of statistics with histogram neighborhood territory pixel.The scope of histogram of gradients is 0~360 degree, wherein per a 10 degree post, 36 posts altogether.Histogrammic peak value has then been represented the principal direction of this key point place neighborhood gradient, promptly as the direction of this key point.
In gradient orientation histogram, when existing another to be equivalent to the peak value of main peak value 80% energy, then this direction is thought the auxilliary direction of this key point.A key point may designatedly have a plurality of directions (principal direction, auxilliary direction more than), and this can strengthen the robustness of coupling.
So far, the key point of image has detected and has finished, and each key point has three information: position, yardstick of living in, direction.Can determine a SIFT characteristic area (representing with ellipse or arrow) thus at the experiment chapters and sections.
(5) the unique point descriptor generates
At first coordinate axis is rotated to be the direction of key point, to guarantee rotational invariance.Next be that 8 * 8 window is got at the center with the key point, on per 4 * 4 fritter, calculate the gradient orientation histogram of 8 directions then, draw the accumulated value of each gradient direction, can form a seed points.The thought of this neighborhood directivity information associating has strengthened the antimierophonic ability of algorithm, also provides fault-tolerance preferably for the characteristic matching that contains positioning error simultaneously.
In the actual computation process, in order to strengthen the robustness of coupling, suggestion to each key point use 4 * 4 totally 16 seed points describe, just can produce 128 data for a key point like this, promptly finally form the 128 SIFT proper vectors of tieing up.The influence that this moment, the SIFT proper vector was removed geometry deformation factors such as dimensional variation, rotation continues the length normalization method with proper vector again, then can further remove the influence of illumination variation.
After the SIFT proper vector of two width of cloth images generated, we adopted the Euclidean distance of key point proper vector to be used as the similarity determination tolerance of key point in two width of cloth images next step.Get certain key point in the image 1, and find out European nearest preceding two key points in itself and the image 2, in these two key points, near distance is less than certain proportion threshold value if nearest distance is removed in proper order, then accepts this a pair of match point.Reduce this proportion threshold value, SIFT match point number can reduce, but more stable.
3. calculate the pose conversion of Matching Model and locus:
Fundamental purpose is to determine orientation and the exact position in scene and the pose of dummy model.Determining of pose is by continuous rotation dummy model and to matching result marking, and be the highest and when reaching the experience threshold values of setting, system thinks that this bit model appearance is consistent with the pose of target object in the real scene when the coupling score value.This computation process as shown in Figure 2.
4. model dimension adjustment:
To the model on the coupling, sometimes might be inconsistent on yardstick with the real object object, this just need carry out convergent-divergent to dummy model on dimension.Shown in Fig. 4 a and Fig. 4 b, the ratio of convergent-divergent by the scale ratio of the containing box of dummy model under the containing box of target object in the real scene image and the identical pose as coefficient.When putting back in the real scene, at first carry out scale transformation, to dummy model generation to guarantee consistent on space scale of dummy model and real object object.
5. actual situation merges
The core that actual situation merges is Matching Model generation of extracting with model bank to put the real object object of distant operation site.In this process, be taken back in the real scene with the dummy model of yardstick such as target object in the real scene, need carry out being tied to the conversion of camera coordinates system from world coordinates.
Usually, two video camera centers keep on the same surface level, i.e. camera horizontal positioned, and then the image coordinate Y coordinate of unique point P is identical, i.e. Y Left=Y Right=Y, according to shown in Figure 3,
Obtain following equation by the triangle geometric relationship:
X left = f x c z c X right = f ( x c - B ) z c Y = f y c z c
Parallax is in the following formula: Disparity=X Left-X RightCan calculate the three-dimensional coordinate of unique point P under camera coordinates system thus is:
x c = t c &CenterDot; X left Disparity y c = t c &CenterDot; Y Disparity z c = t c &CenterDot; f Disparity (4-1-3)
Wherein, (X Left, Y), (X Right, Y) being the picture point of spatial point P left and right sides image centering, picture point is to obtaining by the correlativity of calculating pixel:
Corr ( X , Y ) = &Sigma; i = 1 m ( x i - x &OverBar; ) ( y i - y &OverBar; ) [ &Sigma; i = 1 m ( x i - x &OverBar; ) 2 &Sigma; i = 1 m ( y i - y &OverBar; ) 2 ] 1 / 2
Wherein (x y) represents given pixel; (x i, y i) represent image to the pixel in another width of cloth image; The size of m template.To given P (x c, y c, z c) (x y), when following formula is obtained minimum value under given template size m, then thinks (x to the projection image's vegetarian refreshments in the width of cloth image at stereopsis i, y i) be (x, the y) projection of correspondence in another width of cloth image of stereopsis centering.
Formula 4-1-3 is used to set up the actual coordinate of model under world coordinate system and its dummy model projection coordinate related under camera coordinates system.
Fig. 4 a and Fig. 4 b have provided actual situation and have merged the sectional drawing of example in the video image of the left and right sides.In this example, target object is replaced by virtual three-dimensional model, and its model can be selected.
It is to be noted and any distortion of having done according to the specific embodiment of the present invention all do not break away from the scope that spirit of the present invention and claim are put down in writing.

Claims (5)

1. the seamless integration method of stereoscopic vision image and three-dimensional virtual object is specially: 1) gather actual stereoscopic vision image; 2) according to the characteristic information extraction algorithm to collecting to such an extent that stereoscopic vision image and virtual image model carry out feature extraction, and then carry out the coupling of stereoscopic vision image and virtual image model; 3), mate score value and calculate, to determine the pose of matching virtual iconic model by the virtual image model after the rotation coupling; 4) the yardstick adjustment of matching virtual iconic model; 5) will mate completely virtual image model generation puts the real object object of distant operation site.
2. the seamless integration method of stereoscopic vision image as claimed in claim 1 and three-dimensional virtual object is characterized in that, described virtual image model is stored in the model bank, and can call by different instruction, the access model information.
3. the seamless integration method of stereoscopic vision image as claimed in claim 1 and three-dimensional virtual object, it is characterized in that, the concrete steps that the pose of virtual image model is determined in the step 3) are: the experience threshold values of setting up the pose coupling, virtual image model of every rotation, by calculating actual threshold values, when actual threshold values is lower than the experience threshold values, the pose of assert virtual image model and stereoscopic vision image is inconsistent, continue rotation virtual image model, when actual threshold values is higher than the experience threshold values, this pose is noted, and it is consistent with the target object pose in the stereoscopic vision image that statistics obtains the virtual image model pose of actual threshold values when the highest at last.
4. the seamless integration method of stereoscopic vision image as claimed in claim 1 and three-dimensional virtual object, it is characterized in that, the yardstick adjustment of virtual image model need be carried out convergent-divergent to the virtual image model in the step 4) on dimension, the ratio of convergent-divergent, is adjusted as coefficient by the scale ratio of the containing box of virtual image model under the containing box of target object in the real scene image and the identical pose.
5. the seamless integration method of stereoscopic vision image as claimed in claim 1 and three-dimensional virtual object, it is characterized in that, in the step 5) in the real scene virtual image model of yardstick such as target object be taken back in the real scene, need carry out being tied to the conversion of camera coordinates system from world coordinates.
CN201010110440A 2010-02-09 2010-02-09 Seamless integration method of stereoscopic vision image and three-dimensional virtual object Pending CN101794459A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201010110440A CN101794459A (en) 2010-02-09 2010-02-09 Seamless integration method of stereoscopic vision image and three-dimensional virtual object

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201010110440A CN101794459A (en) 2010-02-09 2010-02-09 Seamless integration method of stereoscopic vision image and three-dimensional virtual object

Publications (1)

Publication Number Publication Date
CN101794459A true CN101794459A (en) 2010-08-04

Family

ID=42587130

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201010110440A Pending CN101794459A (en) 2010-02-09 2010-02-09 Seamless integration method of stereoscopic vision image and three-dimensional virtual object

Country Status (1)

Country Link
CN (1) CN101794459A (en)

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102096941A (en) * 2011-01-30 2011-06-15 北京航空航天大学 Consistent lighting method under falsehood-reality fused environment
CN102157011A (en) * 2010-12-10 2011-08-17 北京大学 Method for carrying out dynamic texture acquisition and virtuality-reality fusion by using mobile shooting equipment
CN103258350A (en) * 2013-03-28 2013-08-21 广东欧珀移动通信有限公司 Method and device for displaying 3D images
CN105488801A (en) * 2015-12-01 2016-04-13 深圳华强数码电影有限公司 Method and system for combining real shooting of full dome film with three-dimensional virtual scene
CN106097454A (en) * 2016-06-06 2016-11-09 成都天福创造机器人有限公司 A kind of man-machine interactive system and exchange method
CN107341827A (en) * 2017-07-27 2017-11-10 腾讯科技(深圳)有限公司 A kind of method for processing video frequency, device and storage medium
CN109003305A (en) * 2018-07-18 2018-12-14 江苏实景信息科技有限公司 A kind of positioning and orientation method and device
CN110969682A (en) * 2019-11-27 2020-04-07 深圳追一科技有限公司 Virtual image switching method and device, electronic equipment and storage medium
CN112639888A (en) * 2018-08-09 2021-04-09 祖克斯有限公司 Programmed world generation
CN114359522A (en) * 2021-12-23 2022-04-15 阿依瓦(北京)技术有限公司 AR model placing method and device

Cited By (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102157011A (en) * 2010-12-10 2011-08-17 北京大学 Method for carrying out dynamic texture acquisition and virtuality-reality fusion by using mobile shooting equipment
CN102096941B (en) * 2011-01-30 2013-03-20 北京航空航天大学 Consistent lighting method under falsehood-reality fused environment
CN102096941A (en) * 2011-01-30 2011-06-15 北京航空航天大学 Consistent lighting method under falsehood-reality fused environment
CN103258350A (en) * 2013-03-28 2013-08-21 广东欧珀移动通信有限公司 Method and device for displaying 3D images
CN105488801B (en) * 2015-12-01 2019-02-15 深圳华强数码电影有限公司 The method and system that spherical screen stereoscopic film real scene shooting and three-dimensional virtual scene combine
CN105488801A (en) * 2015-12-01 2016-04-13 深圳华强数码电影有限公司 Method and system for combining real shooting of full dome film with three-dimensional virtual scene
CN106097454A (en) * 2016-06-06 2016-11-09 成都天福创造机器人有限公司 A kind of man-machine interactive system and exchange method
CN107341827A (en) * 2017-07-27 2017-11-10 腾讯科技(深圳)有限公司 A kind of method for processing video frequency, device and storage medium
CN109003305A (en) * 2018-07-18 2018-12-14 江苏实景信息科技有限公司 A kind of positioning and orientation method and device
CN109003305B (en) * 2018-07-18 2021-07-20 江苏实景信息科技有限公司 Positioning and attitude determining method and device
CN112639888A (en) * 2018-08-09 2021-04-09 祖克斯有限公司 Programmed world generation
US11861790B2 (en) 2018-08-09 2024-01-02 Zoox, Inc. Procedural world generation using tertiary data
CN110969682A (en) * 2019-11-27 2020-04-07 深圳追一科技有限公司 Virtual image switching method and device, electronic equipment and storage medium
CN110969682B (en) * 2019-11-27 2021-03-02 深圳追一科技有限公司 Virtual image switching method and device, electronic equipment and storage medium
CN114359522A (en) * 2021-12-23 2022-04-15 阿依瓦(北京)技术有限公司 AR model placing method and device

Similar Documents

Publication Publication Date Title
CN101794459A (en) Seamless integration method of stereoscopic vision image and three-dimensional virtual object
CN103530881B (en) Be applicable to the Outdoor Augmented Reality no marks point Tracing Registration method of mobile terminal
CN106651942A (en) Three-dimensional rotation and motion detecting and rotation axis positioning method based on feature points
CN103729885B (en) Various visual angles projection registers united Freehandhand-drawing scene three-dimensional modeling method with three-dimensional
CN101877143B (en) Three-dimensional scene reconstruction method of two-dimensional image group
CN103604417B (en) The multi-view images bi-directional matching strategy that object space is information constrained
CN106683173A (en) Method of improving density of three-dimensional reconstructed point cloud based on neighborhood block matching
CN106485690A (en) Cloud data based on a feature and the autoregistration fusion method of optical image
CN102074015A (en) Two-dimensional image sequence based three-dimensional reconstruction method of target
CN107862744A (en) Aviation image three-dimensional modeling method and Related product
CN104346608A (en) Sparse depth map densing method and device
CN102902355A (en) Space interaction method of mobile equipment
CN102697508A (en) Method for performing gait recognition by adopting three-dimensional reconstruction of monocular vision
CN107944459A (en) A kind of RGB D object identification methods
CN104637090A (en) Indoor scene modeling method based on single picture
CN107329962A (en) Image retrieval data library generating method, the method and device of augmented reality
CN107134008A (en) A kind of method and system of the dynamic object identification based under three-dimensional reconstruction
CN103456043A (en) Panorama-based inter-viewpoint roaming method and device
CN105719277A (en) Transformer station three-dimensional modeling method and system based on surveying and mapping and two-dimensional image
CN103700082B (en) Image split-joint method based on dual quaterion relative orientation
CN106485741A (en) A kind of method of the non-firm point set registration retaining partial structurtes
CN105488491A (en) Human body sleep posture detection method based on pyramid matching histogram intersection kernel
CN104166995B (en) Harris-SIFT binocular vision positioning method based on horse pace measurement
CN104318552A (en) Convex hull projection graph matching based model registration method
Zhu et al. Large-scale architectural asset extraction from panoramic imagery

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C02 Deemed withdrawal of patent application after publication (patent law 2001)
WD01 Invention patent application deemed withdrawn after publication

Open date: 20100804