TWM594322U - Camera configuration system with omnidirectional stereo vision - Google Patents

Camera configuration system with omnidirectional stereo vision Download PDF

Info

Publication number
TWM594322U
TWM594322U TW108215064U TW108215064U TWM594322U TW M594322 U TWM594322 U TW M594322U TW 108215064 U TW108215064 U TW 108215064U TW 108215064 U TW108215064 U TW 108215064U TW M594322 U TWM594322 U TW M594322U
Authority
TW
Taiwan
Prior art keywords
camera
lens
main
depth
lenses
Prior art date
Application number
TW108215064U
Other languages
Chinese (zh)
Inventor
薛樂山
Original Assignee
大陸商南京深視光點科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 大陸商南京深視光點科技有限公司 filed Critical 大陸商南京深視光點科技有限公司
Priority to TW108215064U priority Critical patent/TWM594322U/en
Publication of TWM594322U publication Critical patent/TWM594322U/en

Links

Images

Abstract

一種全向立體視覺的相機配置系統,相機配置系統主要包括:至少四個主攝像鏡頭,分別排列於可構成一矩形的四基準線;至少四個輔攝像鏡頭,各主攝像鏡頭與各副攝像鏡頭,可沿著四個基準線於同一平面間隔排列設置為矩形;本創作之關鍵在於,各副攝像鏡頭或各主攝像鏡頭的光軸,被組態為可分為兩部分旋轉,使排列於同一基準線之攝像鏡頭的光軸彼此平行,以及使排列於同一基準線之攝像鏡頭的光軸皆與基準線垂直,以完成相機校正;藉此,可讓對應於各基準線的至少二深度圖於執行影像融合演算後,消除遮擋區域,進而能透過影像拼接獲取無遮擋的全向深度圖。An omnidirectional stereo vision camera configuration system. The camera configuration system mainly includes: at least four main camera lenses, which are respectively arranged on four reference lines that can form a rectangle; at least four auxiliary camera lenses, each main camera lens and each sub camera The lenses can be arranged as rectangles at four intervals along the same plane; the key to this creation is that the optical axis of each secondary camera lens or each main camera lens is configured to rotate in two parts to make the arrangement The optical axes of the camera lenses at the same reference line are parallel to each other, and the optical axes of the camera lenses arranged at the same reference line are perpendicular to the reference line to complete the camera calibration; thereby, at least two corresponding to each reference line can be made After performing the image fusion calculation on the depth map, the occlusion area is eliminated, and then the omnidirectional depth map without occlusion can be obtained through image stitching.

Description

全向立體視覺的相機配置系統Omnidirectional stereo vision camera configuration system

本創作涉及立體視覺(Stereo Vision)技術,尤指一種可將多個攝像鏡頭間隔排列設置為矩形,並使待校正之至少三個攝像鏡頭的鏡心係配置於同一基準線,以可藉此獲得無遮擋深度圖,進而能於完成影像拼接(image stitching)後獲取全向深度圖(omnidirectional depth map)的「全向立體視覺的相機配置系統」。This creation involves Stereo Vision (Stereo Vision) technology, especially one that can arrange a plurality of camera lenses at intervals and arrange the lens centers of at least three camera lenses to be corrected on the same reference line, so that Obtaining an unobstructed depth map, and then an omnidirectional depth map (omnidirectional depth map) camera configuration system that can obtain an omnidirectional depth map after image stitching is completed.

在雙目視覺中,需要對相機進行標定和校正,而相機校正(camera rectification)的目的是達成理想的雙目相機系統,使得至少兩個相機的光軸完全平行(即鏡心彼此之間僅存在X分量),並與基線垂直(左右相機之鏡心的連線即為基線),才能續行深度計算與三維重建。In binocular vision, the camera needs to be calibrated and corrected, and the purpose of camera rectification is to achieve an ideal binocular camera system, so that the optical axes of at least two cameras are completely parallel (that is, the mirror centers are only There is an X component), and it is perpendicular to the baseline (the line connecting the lens centers of the left and right cameras is the baseline) in order to continue the depth calculation and 3D reconstruction.

一般而言,校正前左右相機的鏡心(光心)並不是平行的,而校正後理想雙目系統的左右相機的光軸彼此平行、光軸和影像平面垂直,且成像點在左右影像上的高度一致,依此,後續進行立體匹配(stereo matching)時,只需在同一列的搜索區間(search range),搜索左右影像的匹配點即可,能使相機校正效率大大提高。Generally speaking, the lens centers (optical centers) of the left and right cameras before correction are not parallel, and the optical axes of the left and right cameras of the ideal binocular system after correction are parallel to each other, the optical axis and the image plane are perpendicular, and the imaging point is on the left and right images The heights are the same. Therefore, when performing stereo matching in the future, you only need to search the matching points of the left and right images in the search range of the same column, which can greatly improve the camera calibration efficiency.

目前雖有人提出可利用3D感測相機(Stereoscopic Camera)取得全向深度資訊,例如中華民國發明專利第TW201832547號「產生全景深度影像的影像裝置、相關方法及相關影像裝置」所揭(下稱台灣案),台灣案主要使四顆超廣角魚眼鏡頭(>190度)兩兩背對與上下排列,並將擷取的廣角影像投影至經緯(equirectangular)座標系,而位於同側的左右兩顆魚眼鏡頭可用於計算180x180度的立體深度影像(stereo depth),最後再將兩側的深度影像進行影像拼接,即可獲取360x180度的全向深度影像,然而,此種將廣角影像投影至經緯座標的等距柱狀投影法(Equirectangular Projection),在接近180度的影像解析度相當差(因使用魚眼鏡頭會產生鏡頭扭曲現象),同時,也會產生因採用立體視覺技術所產生的遮擋(Occlusion)問題,而遮擋問題將直接影響到深度估計的準確度。At present, although it has been proposed to use 3D sensing camera (Stereoscopic Camera) to obtain omnidirectional depth information, for example, the Republic of China Invention Patent No. TW201832547 "Image Device for Generating Panoramic Depth Image, Related Methods and Related Image Devices" (hereinafter referred to as Taiwan) Case), the Taiwan case mainly arranges four ultra-wide-angle fisheye lenses (>190 degrees) back-to-back and up and down, and projects the captured wide-angle image to the equirectangular coordinate system, which is located on the same side A fisheye lens can be used to calculate a 180x180 degree stereo depth image, and finally the two side depth images are stitched together to obtain a 360x180 degree omnidirectional depth image. However, this type of wide angle image is projected to The equidistant cylindrical projection method of the latitude and longitude coordinates (Equirectangular Projection), the image resolution near 180 degrees is quite poor (due to the use of fisheye lenses will cause lens distortion), at the same time, it will also be caused by the use of stereo vision technology Occlusion problem, which will directly affect the accuracy of depth estimation.

雖另有人提出可將N顆相機排列成正N邊形,並透過兩兩相機生成深度影像,最後再完成全景拼接以獲取全向深度影像,即如美國發明專利第US10244226B2「Camera rig and stereoscopic image capture」所揭(下稱美國案),然而,此種作法雖可產生解析度較高的深度影像,但美國案擺放各相機的位置,將無法處理前述的遮擋問題,使得其產生的深度影像產生缺陷。Although others have proposed that N cameras can be arranged into a regular N polygon, and depth images are generated by two or two cameras, and finally panoramic stitching is completed to obtain omnidirectional depth images, such as US Patent No. US10244226B2 "Camera rig and stereoscopic image capture It is revealed (hereinafter referred to as the US case) that, although this method can produce a high-resolution depth image, the location of each camera in the US case will not be able to deal with the aforementioned occlusion problem, resulting in the depth image it produces Defective.

承上,為了在雙目視覺領域獲取無遮擋深度圖(de-occlusion depth map),多鏡頭(至少三顆)的相機校正是必要的,並請搭配參閱「第1圖」,假設圖中所示的基線b12、b13分別為美國案校正前第一與第二相機、以及美國案校正前第一與第三相機之鏡心O的連線,而基線b12’、b13’即為校正後第一相機與第二相機、以及校正後第一相機與第三相機之鏡心O的連線,校正時,得以基線b13為基準線,故第一與第三相機只需透過旋轉即可讓X軸與基線b13的向量平行,至於第二相機雖亦可透過偏移矩陣T(

Figure 02_image001
,亦可稱平移矩陣),達到讓X軸與基線b12的向量平行,但由於第二相機偏離X軸太遠,導致偏移矩陣T的y、z分量較大,而當y、z分量達到一定程度(例如若高於1mm),將造成後續計算深度圖的誤差過大,進而影響到後續將相機座標轉換至世界座標的準確性,且即便有其它方式(例如調整第二相機的焦距)可讓第二相機的鏡心位置逼近於X軸(理想位置),但效果仍為有限,且屆時在計算深度圖時也會因此增加可觀的運算量,換言之,由於美國案的第一至第三相機的鏡心並非排列於同一基線(尤其像第二相機偏離X軸太遠),故在無法取得第二相機所拍攝場景物體的XYZ資訊(相機座標)前,將無法計算出無遮擋視差圖。 According to the above, in order to obtain a de-occlusion depth map in the field of binocular vision, multi-lens (at least three) camera correction is necessary, and please refer to "Picture 1". The baselines b12 and b13 shown are the connection of the lens centers O of the first and second cameras before correction in the US case and the first and third cameras before correction in the US case, and the baselines b12' and b13' are the first The connection between the first camera and the second camera, and the center O of the first camera and the third camera after calibration, the baseline b13 is used as the reference line during the calibration, so the first and third cameras only need to rotate to make X The axis is parallel to the vector of the baseline b13, although the second camera can also pass through the offset matrix T(
Figure 02_image001
, Can also be called a translation matrix), so that the X axis is parallel to the vector of the baseline b12, but because the second camera deviates too far from the X axis, the y and z components of the offset matrix T are large, and when the y and z components reach To a certain extent (for example, if it is higher than 1mm), it will cause excessive errors in the subsequent calculation of the depth map, which will affect the accuracy of the subsequent conversion of the camera coordinates to the world coordinates, and even if there are other ways (such as adjusting the focal length of the second camera) The position of the lens center of the second camera is approximated to the X axis (ideal position), but the effect is still limited, and the amount of calculation will also be increased when calculating the depth map. In other words, due to the first to third in the US case The lens centers of the cameras are not arranged on the same baseline (especially if the second camera is too far away from the X axis), so the unobstructed parallax map cannot be calculated before the XYZ information (camera coordinates) of the scene objects shot by the second camera cannot be obtained .

依此,如何提出一種可獲取無遮擋視差圖、毋須使用魚眼鏡頭的全向立體視覺的相機配置系統,乃有待解決之問題。Accordingly, how to propose an omnidirectional stereo vision camera configuration system that can obtain an unobstructed parallax map without using a fisheye lens is a problem to be solved.

為達上述目的,本創作提出一種全向立體視覺的相機配置系統,相機配置系統主要包括: 一主相機組、一輔相機組及一運算單元;其中,主相機組可至少包括四個主攝像鏡頭,各主攝像鏡頭可分別排列於可構成一矩形的四個基準線;輔相機組可至少包括四個副攝像鏡頭,且各主攝像鏡頭與各副攝像鏡頭,可沿著四個基準線於同一平面間隔排列設置為矩形;運算單元可分別與各主攝像鏡頭及各副攝像鏡頭呈資訊連結;各副攝像鏡頭或各主攝像鏡頭的光軸被組態為可旋轉,使排列於基準線之副攝像鏡頭與主攝像鏡頭的光軸彼此平行,以及使排列於基準線之副攝像鏡頭與主攝像鏡頭的光軸皆與對應的基準線垂直,以完成相機校正;位於同一基準線且完成相機校正的該等攝像鏡頭,其所擷取的多個影像可供運算單元演算出至少兩深度圖,且運算單元對各深度圖執行影像融合演算後,可消除各深度圖彼此之間的遮擋區域,而生成一去遮擋深度圖;最後,運算單元可對於各基準線所演算出的各去遮擋深度圖,進行一影像拼接而獲取一全向深度圖。To achieve the above purpose, this author proposes an omnidirectional stereo vision camera configuration system. The camera configuration system mainly includes: a main camera group, an auxiliary camera group, and an arithmetic unit; wherein, the main camera group may include at least four main cameras Lenses, each main camera lens can be arranged in four reference lines that can form a rectangle; the auxiliary camera group can include at least four auxiliary camera lenses, and each main camera lens and each auxiliary camera lens can be along four reference lines Arranged in the same plane at a rectangular interval; the arithmetic unit can be connected to each main camera lens and each sub camera lens in information; the optical axis of each sub camera lens or each main camera lens is configured to be rotatable, so that the arrangement is based on the reference The optical axes of the auxiliary camera lens and the main camera lens of the line are parallel to each other, and the optical axes of the auxiliary camera lens and the main camera lens arranged on the reference line are perpendicular to the corresponding reference line to complete the camera calibration; they are on the same reference line and For the camera lenses that have completed camera calibration, the multiple images captured by the computing unit can be used by the computing unit to calculate at least two depth maps, and after the computing unit performs image fusion calculation on each depth map, it can eliminate the depth maps between each other. The occlusion area generates a de-occlusion depth map; finally, the arithmetic unit may perform image stitching on each de-occlusion depth map calculated by each reference line to obtain an omnidirectional depth map.

藉此,本創作據以實施後,相較於既有獲取全向深度圖的作法,本創作至少可達成獲取無遮擋的全向深度度之有利功效,並且,即便採用的相機數量為四顆以上,仍可使該等相機維持矩形排列,而毋須排列為多邊形而增加相機系統的配置成本。In this way, after the implementation of this creation, compared with the existing method of obtaining omnidirectional depth maps, this creation can at least achieve the beneficial effect of obtaining unobstructed omnidirectional depth, and even if the number of cameras used is four The above can still maintain the rectangular arrangement of these cameras without arranging them as polygons, which increases the configuration cost of the camera system.

為使 貴審查委員得以清楚了解本創作之目的、技術特徵及其實施後之功效,茲以下列說明搭配圖示進行說明,敬請參閱。In order to enable your reviewing committee to clearly understand the purpose, technical features and effects of this creation, the following description is accompanied by illustrations, please refer to it.

請參閱「第2圖」,其為本創作的相機配置系統架構圖,本創作提出一種全向立體視覺的相機配置系統10,包括一主相機組101、一輔相機組102及一運算單元103,其中: (1)主相機組101至少包括四個主攝像鏡頭(1011~1014),各主攝像鏡頭(1011~1014)可分別排列於可構成一矩形的四個基準線(base line); (2)輔相機組102至少包括四個副攝像鏡頭(1021~1024),且各主攝像鏡頭(1011~1014)與各副攝像鏡頭(1021~1024),被配置為可沿著四個基準線於同一平面間隔排列設置為矩形; (3)運算單元103分別與各主攝像鏡頭(1011~1014)及各副攝像鏡頭(1021~1024)呈資訊連結,其中,運算單元103可至少具有一處理器(圖中未繪示,例如CPU、MCU),其供以運行運算單元103,並具備邏輯運算、暫存運算結果、保存執行指令位置、執行影像處理等功能; (4)各副攝像鏡頭(1021~1024)或各主攝像鏡頭(1011~1014)的鏡頭光軸(Optical Axis)均可被組態為可旋轉,使排列於基準線之副攝像鏡頭(1021~1024)與主攝像鏡頭(1011~1014)的鏡頭光軸彼此平行,以及使排列於基準線之副攝像鏡頭(1021~1024)與主攝像鏡頭(1011~1014)的鏡頭光軸皆與對應的基準線垂直,以對主相機組101與輔相機組102皆完成相機校正; (5)位於同一基準線且完成相機校正的該等攝像鏡頭(1011~1014、1021~1024),其所擷取的影像可供運算單元103演算出至少兩深度圖(depth map,亦可稱視差圖),且運算單元103對各深度圖執行影像融合演算(fusion)後,可消除各深度圖彼此之間的遮擋區域,而生成一去遮擋深度圖; (6)運算單元103可對於各基準線所演算出的各去遮擋深度圖,進行一影像拼接而獲取一全向深度圖; (7)承上,本創作在一較佳實施例中,輔相機組102的各副攝像鏡頭(1021~1024),或是主相機組101的主攝像鏡頭(1011~1014),係可被組態為以鏡頭光軸為基準向左或向右旋轉一特定角度,且位於同一基準線的主相機組101與輔相機組102,其鏡頭光軸於完成相機校正前,係可分別朝向同一方向或不同方向; (8)承上,本創作在一較佳實施例中,更可包括耦接於運算單元103的一慣性傳感器104(IMU),供以回傳多個自由度(DOF)的運動資訊與姿態資訊,以更精確追蹤被攝物體與拍攝場景如何在現實世界中運動,其中,自由度可例如區分為兩種類型:平移和旋轉,平移可包括X軸(前/後)、Y軸(左/右)、Z軸(上/下),旋轉可包括縱搖(Pitch)、橫搖(Roll)和垂搖(Yaw),但並不以此些自由度為限。 Please refer to "Figure 2", which is the architecture diagram of the camera configuration system created by the author. This creation proposes an omnidirectional stereo vision camera configuration system 10, which includes a main camera group 101, an auxiliary camera group 102, and an arithmetic unit 103 , Where: (1) The main camera group 101 includes at least four main camera lenses (1011~1014), and each main camera lens (1011~1014) can be arranged on four base lines that can form a rectangle; (2) The auxiliary camera group 102 includes at least four auxiliary camera lenses (1021 to 1024), and each main camera lens (1011 to 1014) and each auxiliary camera lens (1021 to 1024) are configured to be along four benchmarks The lines are arranged on the same plane at intervals, and are arranged as rectangles; (3) The computing unit 103 is respectively connected to the main camera lenses (1011~1014) and the sub camera lenses (1021~1024), wherein the computing unit 103 can have at least one A processor (not shown in the figure, such as a CPU and an MCU), which is used to run the arithmetic unit 103, and has functions such as logical operation, temporary storage of operation results, storage of execution instruction positions, and execution of image processing; (4) each sub-camera The lens (1021~1024) or the optical axis of each main camera lens (1011~1014) can be configured to be rotatable, so that the auxiliary camera lens (1021~1024) arranged on the reference line and the main camera The optical axes of the lenses (1011~1014) are parallel to each other, and the optical axes of the secondary camera lenses (1021~1024) and the main camera lenses (1011~1014) arranged on the reference line are perpendicular to the corresponding reference lines, Camera calibration is completed for both the main camera group 101 and the auxiliary camera group 102; (5) The camera lenses (1011~1014, 1021~1024) which are located on the same reference line and complete the camera correction, the captured images are available for calculation The unit 103 calculates at least two depth maps (also called disparity maps), and after the arithmetic unit 103 performs image fusion calculation on each depth map, it can eliminate the occlusion area between the depth maps and generate 1. Deblocking depth map; (6) The computing unit 103 can perform an image splicing on each deblocking depth map calculated by each reference line to obtain an omnidirectional depth map; (7) Continued In a preferred embodiment, each auxiliary camera lens (1021~1024) of the auxiliary camera group 102 or the main camera lens (1011~1014) of the main camera group 101 can be configured to the left based on the lens optical axis Or the main camera group 101 and the auxiliary camera group 102, which are rotated to the right by a certain angle and are located at the same reference line, the lens optical axes of them can be respectively oriented in the same direction or different directions before the camera calibration is completed; (8) In a preferred embodiment of the present invention, it may further include an inertial sensor 104 (IMU) coupled to the computing unit 103 ), which is used to transmit motion information and posture information of multiple degrees of freedom (DOF) to more accurately track how the subject and the shooting scene move in the real world. Among them, the degrees of freedom can be divided into two types: translation And rotation, translation can include X axis (front/rear), Y axis (left/right), Z axis (up/down), rotation can include pitch (Pitch), roll (Roll) and yaw (Yaw) , But not limited to these degrees of freedom.

請參閱「第3圖」,其為本創作的相機配置流程圖,並請搭配參閱「第2圖」、「第4圖」~「第6圖」,本創作提出一種全向立體視覺的相機配置方法S,包括: (1)於基準線配置相機組(步驟S10):如「第4圖」所示,將一主相機組101的至少四個主攝像鏡頭(1011~1014)配置於可構成一矩形的四個基準線(L1~L4),將一輔相機組102的至少四個副攝像鏡頭(1021~1024)分別配置於各基準線(L1~L4),使各主攝像鏡頭(1011~1014)與各副攝像鏡頭(1021~1024)沿著四個基準線(L1~L4)於同一平面間隔排列設置為矩形; (2)校正相機(步驟S20):旋轉輔相機組102之至少一副攝像鏡頭(1021~1024)、或旋轉主相機組101之主攝像鏡頭(1011~1014)的鏡頭光軸,使排列於基準線(L1~L4)之主攝像鏡頭(1011~1014)與副攝像鏡頭(1021~1024)的鏡頭光軸彼此平行,以及使排列於基準線(L1~L4)之主攝像鏡頭(1011~1014)與副攝像鏡頭(1021~1024)的鏡頭光軸皆與對應的基準線(L1~L4)垂直,以完成相機校正,其中,該等攝像鏡頭(1011~1014、1021~1024)係可被組態為以鏡頭光軸為基準,分別二部分向左或向右旋轉一特定角度,即如「第4圖」所示,並請搭配參閱「第2圖」,圖中所示位於基準線L1的主攝像鏡頭1011,係可以其光軸為基準向左或向右旋轉θ度,而位於主攝像鏡頭1011之左右兩側的副攝像鏡頭(1021、1022),由於其與主攝像鏡頭1011的鏡心皆排列於相同的基準線L1,故可將副攝像鏡頭(1021、1022)的光軸旋轉θ度,以讓位於相同基準線L1的攝像鏡頭(1011、1021、1022)在偏移矩陣T中的y、z分量(Ty、Tz)趨近於0,讓攝像鏡頭(1011、1021、1022)的鏡心座標形成彼此零旋轉,而只保留X軸向的偏移量(Tx); (3)承上,同樣地,位於基準線L2的主攝像鏡頭1012,係可以其光軸為基準向左或向右旋轉Φ度,而位於主攝像鏡頭1012之左右兩側的副攝像鏡頭(1022、1023),由於其與主攝像鏡頭1012的鏡心皆排列於相同的基準線L2,故可將副攝像鏡頭(1022、1023)的光軸旋轉Φ度,以讓位於相同基準線L2的攝像鏡頭(1012、1022、1023)在偏移矩陣T中的y、z分量(Ty、Tz)趨近於0,讓攝像鏡頭(1012、1022、1023)的鏡心座標形成彼此零旋轉,而只保留X軸向的偏移量(Tx),至於基準線L3與基準線L4的相機校正作法,則可分別以旋轉角度φ、Ω旋轉攝像鏡頭的光軸,與前述對於基準線L1、L2的作法類同,以此類推,於此不再贅述,其中,

Figure 02_image003
(4)承上,由於本創作於執行步驟S20時,並未改變各攝像鏡頭(1011~1014、1022~1024)的位置,故主相機組101與輔相機組102之攝像頭所接收到的光線都是相同的,依此,可透過旋轉鏡心成功模擬鏡頭光軸所面對的方向; (5)產生深度資訊(步驟S30):由設置於同一基準線、且完成相機校正的主相機組101與輔相機組102所擷取的影像(例如「第5圖」所示的各拍攝影像,而本示意圖僅為舉例,並不以此些影像數量為限),供一運算單元103演算出不同角度的至少兩深度圖,且各深度圖經運算單元103執行影像融合演算後,可消除各深度圖彼此之間的遮擋區域,而生成一去遮擋深度圖,請搭配參閱「第6圖」,圖中左側的深度圖為運算單元103基於位於基準線L1的主攝像鏡頭101與副攝像鏡頭1021所擷取之影像,而生成之深度圖DL,而中間的深度圖為運算單元103基於位於基準線L1的主攝像鏡頭101與副攝像鏡頭1022所擷取之影像,而生成之深度圖DR,從圖中的深度圖DL與深度圖DR皆可發現部分的遮擋區域,而右側的深度圖D即為深度圖DL與深度圖DR作影像融合演算而生成的去遮擋深度圖; (6)產生全向深度圖(步驟S40):運算單元103對於各基準線所演算出的各去遮擋深度圖,進行一影像拼接而獲取一全向深度圖。 Please refer to "Picture 3", which is the flow chart of the camera configuration of this creation, and please refer to "Picture 2", "Picture 4" ~ "Picture 6", this creation proposes an omnidirectional stereo vision camera The configuration method S includes: (1) arranging the camera group on the baseline (step S10): as shown in "Figure 4", arranging at least four main camera lenses (1011~1014) of a main camera group 101 in Four reference lines (L1~L4) forming a rectangle, at least four auxiliary imaging lenses (1021~1024) of an auxiliary camera group 102 are respectively arranged on each reference line (L1~L4), so that each main imaging lens ( 1011~1014) and each sub-camera lens (1021~1024) are arranged in a rectangle along the four reference lines (L1~L4) at the same plane interval; (2) correct the camera (step S20): rotate the auxiliary camera group 102 At least one auxiliary camera lens (1021~1024), or rotate the optical axis of the main camera lens (1011~1014) of the main camera group 101, so that the main camera lens (1011~1014) arranged on the reference line (L1~L4) The lens optical axes of the sub-camera lenses (1021~1024) are parallel to each other, and the optical axes of the main camera lenses (1011~1014) and the sub-camera lenses (1021~1024) arranged on the reference line (L1~L4) are both It is perpendicular to the corresponding reference line (L1~L4) to complete the camera calibration. Among them, the camera lenses (1011~1014, 1021~1024) can be configured to use the lens optical axis as the reference, and the two parts are to the left Or rotate it to the right by a specific angle, as shown in "Figure 4", and please refer to "Figure 2". The main camera lens 1011 at the reference line L1 shown in the figure can use its optical axis as the reference direction Rotate θ degrees to the left or right, and the secondary imaging lenses (1021, 1022) located on the left and right sides of the main imaging lens 1011 are arranged on the same reference line L1 as the lens centers of the main imaging lens 1011. The optical axis of the sub-camera lens (1021, 1022) is rotated by θ degrees, so that the y-z component (Ty, Tz) of the imaging lens (1011, 1021, 1022) located at the same reference line L1 in the offset matrix T tends to Close to 0, let the center coordinates of the camera lens (1011, 1021, 1022) form a zero rotation with each other, but only retain the X-axis offset (Tx); (3) Up, similarly, located at the reference line L2 The main camera lens 1012 is rotated by Φ degrees to the left or right based on its optical axis, and the auxiliary camera lenses (1022, 1023) located on the left and right sides of the main camera lens 1012 are The lens centers are arranged on the same reference line L2, so the optical axis of the sub-camera lens (1022, 1023) can be rotated by Φ degrees, so that the camera lens (1012, 1022, 1023) on the same reference line L2 is shifted Y, z components in matrix T (Ty, Tz) Close to 0, let the lens center coordinates of the camera lens (1012, 1022, 1023) form zero rotation with each other, and only retain the X-axis offset (Tx), as for the camera correction method of the reference line L3 and the reference line L4 , You can rotate the optical axis of the camera lens by the rotation angle φ and Ω respectively, which is the same as the above method for the reference lines L1 and L2, and so on, and will not be repeated here. Among them,
Figure 02_image003
(4) Continued, because the author did not change the position of each camera lens (1011 ~ 1014, 1022 ~ 1024) when performing step S20, so the cameras received by the cameras of the main camera group 101 and the auxiliary camera group 102 The light is the same, so the direction of the lens optical axis can be successfully simulated by rotating the lens center; (5) Depth information is generated (step S30): the main camera is set on the same reference line and completes the camera calibration The images captured by the group 101 and the auxiliary camera group 102 (such as the captured images shown in "Figure 5", and this schematic diagram is only an example, and is not limited to the number of these images), for the calculation unit 103 to calculate At least two depth maps at different angles are produced, and after each depth map is subjected to image fusion calculation by the arithmetic unit 103, the occlusion area between the depth maps can be eliminated, and an unoccluded depth map is generated. Please refer to "Figure 6" The depth map on the left in the figure is the depth map DL generated by the operation unit 103 based on the images captured by the main camera lens 101 and the sub camera lens 1021 at the reference line L1, and the middle depth map is based on the operation unit 103 The depth map DR generated from the images captured by the main camera lens 101 and the sub camera lens 1022 at the reference line L1 can be found from the depth map DL and the depth map DR in the figure. Figure D is the unoccluded depth map generated by the image fusion calculation of the depth map DL and the depth map DR; (6) Generate an omnidirectional depth map (step S40): the computing unit 103 calculates each deocclusion calculated for each reference line Depth map, an image stitching to obtain an omnidirectional depth map.

請參閱「第7圖」,其為本創作之另一實施例(一),並請搭配參閱「第2圖」,本實施例與「第2圖」~「第6圖」之技術類同,主要差異在於,由於在不靠投射主動光源的情況下,3D資訊的品質與影像中物體的紋理程度有關,例如毛毯、報紙即為富含紋理的物體,而白紙、單色牆壁即為不具紋理的物體,另外若是光源不足的情況,例如夜晚或在室內但未開燈的情況下,也會影響3D資訊的品質,依此,本實施例的全向立體視覺的相機配置系統10更可包含耦接於運算單元103的一繞射光學元件105(Diffractive Optical Element,DOE),且繞射光學元件105可分別搭載於主相機組101的各個主攝像鏡頭(1011~1024),而繞射光學元件105主要用於投射光點於物體表面,以輔助判斷被攝物體與拍攝場景的三維深度,意即,於步驟S30執行時,若環境光源不足或被攝物體之紋理特徵不明顯時,耦接於運算單元103的繞射光學元件105係可投射光點於物體表面,以給予被攝物體紋理與光源,達到於特定位置或空間產生所需的圖案,以輔助判斷被攝物體與拍攝場景的三維深度。Please refer to "Figure 7", which is another embodiment of the creation (1), and please refer to "Figure 2", this embodiment is similar to the technology of "Figure 2" to "Figure 6" The main difference is that, without relying on the projection of an active light source, the quality of 3D information is related to the texture level of the objects in the image. For example, blankets and newspapers are texture-rich objects, while white paper and monochrome walls are not. Textured objects, in addition, if the light source is insufficient, such as at night or indoors but not turned on, will also affect the quality of 3D information. Therefore, the omnidirectional stereo vision camera configuration system 10 of this embodiment is more It includes a diffractive optical element (DOE) 105 coupled to the arithmetic unit 103, and the diffractive optical element 105 can be respectively mounted on each main imaging lens (1011~1024) of the main camera group 101, and diffracted The optical element 105 is mainly used to project the light spot on the surface of the object to assist in determining the three-dimensional depth of the subject and the shooting scene, which means that when step S30 is executed, if the ambient light source is insufficient or the texture feature of the subject is not obvious, The diffractive optical element 105 coupled to the computing unit 103 can project a light spot on the surface of the object to give the texture and light source of the object to produce a desired pattern at a specific position or space to assist in determining the object and shooting The three-dimensional depth of the scene.

請參閱「第8圖」,其為本創作之另一實施例(二),並請搭配參閱「第2圖」,本實施例與「第2圖」~「第7圖」之技術類同,主要差異在於,本實施例的全向立體視覺的相機配置系統10更包含耦接於運算單元103的一光達模組106(Lidar),供以透過測量發送和接受到的脈衝訊號(例如脈衝雷射)的時間間隔,來計算被攝物體與拍攝場景的深度資訊,而深度資訊的格式可例如為點雲(Point Cloud,其中資訊可包括水平角度、垂直角度、距離、強度、line、id、時間戳(Laser Timestamp)意即,執行步驟S30時,運算單元103可透過與其耦接的光達模組106,測量發送和接受到的脈衝訊號的時間間隔來確定一飛行時間(ToF),進而演算被攝物體與拍攝場景的深度資訊,並且,在光達模組106尚未回傳深度資訊至運算單元103前,運算單元103可先對步驟S30執行時所產生的各深度圖以及該等攝像鏡頭(1011~1014、1021~1024)所擷取的影像進行影像分割(image segmentation),以供運算單元103獲取全向深度圖時,可針對較遠距離的被攝物體或拍攝場景生成較精確的深度資訊,以彌補若僅採用光達模組106來偵測較遠距離之被攝物體或拍攝場景的深度,其回傳的深度資訊會有x、y資訊不夠緻密的問題;承上,本實施例的全向立體視覺的相機配置系統10更可包含耦接於運算單元103的一雷達模組107(Radar),供以於步驟S30執行時接收空間內存在物體所反射之無線電波,以供運算單元103計算出被攝物體與拍攝場景的深度資訊,且雷達模組107可例如為一毫米波雷達(mmWave Rader),換言之,本實施例透過使用光達模組106與雷達模組107,可解決本創作透過該等攝像鏡頭(1011~1014、1021~1024)達成全向立體視覺時,針對較遠距離的被攝物體或場景之深度資訊可能會有不夠準確的問題。Please refer to "Figure 8", which is another embodiment of the creation (2), and please refer to "Figure 2", this embodiment is similar to the technology of "Figure 2" ~ "Figure 7" The main difference is that the omnidirectional stereo vision camera configuration system 10 of this embodiment further includes a light module 106 (Lidar) coupled to the computing unit 103 for pulse signals sent and received through measurement (for example Pulse laser) to calculate the depth information of the subject and the shooting scene, and the format of the depth information can be, for example, Point Cloud, where the information can include horizontal angle, vertical angle, distance, intensity, line, id, timestamp (Laser Timestamp) means that when step S30 is executed, the computing unit 103 can determine the time of flight (ToF) by measuring the time interval of the pulse signal sent and received through the light module 106 coupled thereto , And then calculate the depth information of the subject and the shooting scene, and before the light module 106 returns the depth information to the computing unit 103, the computing unit 103 can first perform each of the depth maps generated during step S30 and the When the images captured by the camera lens (1011~1014, 1021~1024) are image segmentation (image segmentation) for the computing unit 103 to obtain an omnidirectional depth map, it can be generated for a distant object or shooting scene More accurate depth information to make up for the problem that if only the light module 106 is used to detect the depth of a distant object or shooting scene, the returned depth information will have insufficient x and y information; Above, the omni-directional stereo vision camera configuration system 10 of this embodiment may further include a radar module 107 (Radar) coupled to the computing unit 103 for receiving the wireless reflection from the objects in the receiving space when step S30 is executed The radio wave is used by the computing unit 103 to calculate the depth information of the subject and the shooting scene, and the radar module 107 can be, for example, a millimeter wave radar (mmWave Rader). In other words, in this embodiment, the light module 106 and the radar are used The module 107 can solve the problem that the depth information of a distant object or scene may not be accurate when the creation of omnidirectional stereo vision through these camera lenses (1011~1014, 1021~1024).

以上所述者,僅為本創作之較佳之實施例而已,並非用以限定本創作實施之範圍;任何熟習此技藝者,在不脫離本創作之精神與範圍下所作之均等變化與修飾,皆應涵蓋於本創作之專利範圍內。The above are only the preferred embodiments of this creation, and are not intended to limit the scope of the implementation of this creation; anyone who is familiar with this skill will make equal changes and modifications without departing from the spirit and scope of this creation. It should be covered by the patent scope of this creation.

綜上所述,本創作係具有「產業利用性」、「新穎性」與「進步性」等專利要件;申請人爰依專利法之規定,向 鈞局提起新型專利之申請。In summary, this creative department has patent requirements such as "industrial utility", "novelty" and "progressiveness"; the applicant has filed an application for a new type of patent with the Bureau of Law in accordance with the provisions of the Patent Law.

O:鏡心 b12:基線 b12’:基線 b13:基線 b13’:基線 10:全向立體視覺的相機配置系統 101:主相機組 101:主相機組 1011:主攝影鏡頭 1012:主攝影鏡頭 1013:主攝影鏡頭 1014:主攝影鏡頭 102:輔相機組 1021:副攝影鏡頭 1022:副攝影鏡頭 1023:副攝影鏡頭 1024:副攝影鏡頭 103:運算單元 104:慣性傳感器 105:繞射光學元件 106:光達模組 107:雷達模組 L1~L4:基準線 D:深度圖 DL:深度圖 DR:深度圖 S:全向立體視覺的相機配置方法 S10:於基準線配置相機組 S20:校正相機 S30:產生深度資訊 S40:產生全向深度圖 O: mirror heart b12: baseline b12’: Baseline b13: Baseline b13’: Baseline 10: Camera configuration system for omnidirectional stereo vision 101: Main camera group 101: Main camera group 1011: Main camera lens 1012: Main camera lens 1013: Main photographic lens 1014: main photographic lens 102: auxiliary camera group 1021: Vice photography lens 1022: Vice photography lens 1023: Vice photography lens 1024: secondary camera lens 103: arithmetic unit 104: Inertial sensor 105: Diffractive optics 106: Light module 107: Radar module L1~L4: baseline D: depth map DL: Depth map DR: depth map S: Camera configuration method for omnidirectional stereo vision S10: Configure the camera group on the baseline S20: calibrate the camera S30: Generate in-depth information S40: Generate omnidirectional depth map

第1圖,為習知相機校正示意圖。 第2圖,為本創作的相機配置系統架構圖。 第3圖,為本創作的相機配置流程圖。 第4圖,為本創作的之實施示意圖(一)。 第5圖,為本創作的之實施示意圖(二)。 第6圖,為本創作的之實施示意圖(三)。 第7圖,為本創作之另一實施例(一)。 第8圖,為本創作之另一實施例(二)。 Figure 1 is a schematic diagram of conventional camera calibration. Figure 2 is the architecture diagram of the camera configuration system created for this. Figure 3 is a flow chart of the camera configuration for this creation. Figure 4 is a schematic diagram of the implementation of this creation (1). Figure 5 is a schematic diagram of the implementation of this creation (2). Figure 6 is a schematic diagram of the implementation of this creation (3). Figure 7 is another embodiment (1) of this creation. Figure 8 is another embodiment (2) of this creation.

10:全向立體視覺的相機配置系統 10: Camera configuration system for omnidirectional stereo vision

101:主相機組 101: Main camera group

101:主相機組 101: Main camera group

1011:主攝影鏡頭 1011: Main camera lens

1012:主攝影鏡頭 1012: Main camera lens

1013:主攝影鏡頭 1013: Main photographic lens

1014:主攝影鏡頭 1014: main photographic lens

102:輔相機組 102: auxiliary camera group

1021:副攝影鏡頭 1021: Vice photography lens

1022:副攝影鏡頭 1022: Vice photography lens

1023:副攝影鏡頭 1023: Vice photography lens

1024:副攝影鏡頭 1024: secondary camera lens

103:運算單元 103: arithmetic unit

104:慣性傳感器 104: Inertial sensor

Claims (7)

一種全向立體視覺的相機配置系統,包括: 一主相機組,至少包括四個主攝像鏡頭,各該主攝像鏡頭分別排列於可構成一矩形的四個基準線; 一輔相機組,至少包括四個副攝像鏡頭,且各該主攝像鏡頭與各該副攝像鏡頭,沿著該四個基準線於同一平面間隔排列設置為該矩形; 一運算單元,分別與各該主攝像鏡頭及各該副攝像鏡頭呈資訊連結; 各該副攝像鏡頭或各該主攝像鏡頭的光軸被組態為可旋轉,使排列於該基準線之該副攝像鏡頭與該主攝像鏡頭的光軸彼此平行,以及使排列於該基準線之該副攝像鏡頭與該主攝像鏡頭的光軸皆與對應的該基準線垂直,以完成相機校正; 位於同一基準線且完成相機校正的該等攝像鏡頭,其所擷取的多個影像供該運算單元演算出至少兩深度圖,且該至少二深度圖供該運算單元執行影像融合演算,以消除各該深度圖彼此之間的遮擋區域,而生成一去遮擋深度圖;以及 該運算單元對於各該基準線所演算出的各該去遮擋深度圖,進行一影像拼接而獲取一全向深度圖。 An omnidirectional stereo vision camera configuration system, including: A main camera group including at least four main camera lenses, each of which is arranged on four reference lines that can form a rectangle; An auxiliary camera group, including at least four auxiliary camera lenses, and each of the main camera lens and each of the auxiliary camera lenses is arranged along the four datum lines on the same plane at intervals to form the rectangle; An arithmetic unit, which is connected to each of the main camera lens and the auxiliary camera lens to present information link; The optical axis of each of the sub imaging lenses or each of the main imaging lenses is configured to be rotatable, so that the optical axes of the sub imaging lens and the main imaging lens arranged on the reference line are parallel to each other, and arranged on the reference line The optical axes of the auxiliary camera lens and the main camera lens are both perpendicular to the corresponding reference line to complete the camera calibration; For the camera lenses that are located on the same reference line and have completed camera calibration, the captured multiple images are used by the arithmetic unit to calculate at least two depth maps, and the at least two depth maps are used by the arithmetic unit to perform image fusion calculations to eliminate Each occlusion area between the depth maps to generate an unoccluded depth map; and The computing unit performs image stitching on each of the unoccluded depth maps calculated by the reference lines to obtain an omnidirectional depth map. 如申請專利範圍第1項的全向立體視覺的相機配置系統,其中,更包含耦接於該運算單元的一繞射光學元件,供以投射光點於物體表面,以輔助判斷被攝物體與拍攝場景的三維深度。For example, the omnidirectional stereo vision camera configuration system of the first patent application scope, which further includes a diffractive optical element coupled to the computing unit, is used to project the light spot on the surface of the object, to assist in determining the subject and The three-dimensional depth of the shooting scene. 如申請專利範圍第1項的全向立體視覺的相機配置系統,其中,該輔相機組的各該副攝像鏡頭,或該主相機組的該主攝像鏡頭,係被組態為以光軸為基準向左或向右旋轉。An omnidirectional stereo vision camera configuration system as claimed in item 1 of the patent scope, wherein each auxiliary camera lens of the auxiliary camera group, or the main camera lens of the main camera group, is configured to use the optical axis as The datum rotates to the left or right. 如申請專利範圍第1項或第2項的全向立體視覺的相機配置系統,更包括耦接於該運算單元的一光達模組,供以透過測量發送和接受到的脈衝訊號的時間間隔,來計算被攝物體與拍攝場景的深度資訊。For example, the omnidirectional stereo vision camera configuration system of the first or second patent application scope further includes a light module coupled to the computing unit for the time interval of the pulse signal sent and received by the measurement To calculate the depth information of the subject and shooting scene. 如申請專利範圍第4項的全向立體視覺的相機配置系統,其中,該光達模組尚未回傳深度資訊至該運算單元前,該運算單元亦供以對各該深度圖以及該等攝像鏡頭所擷取的影像進行影像分割。For example, the omnidirectional stereo vision camera configuration system of the fourth patent application, in which the light module has not yet returned the depth information to the arithmetic unit, the arithmetic unit is also used for the depth map and the cameras The image captured by the lens performs image segmentation. 如申請專利範圍第1項的全向立體視覺的相機配置系統,其中,更包括耦接於該運算單元的一慣性傳感器,供以回傳多個自由度的運動資訊與姿態資訊。For example, the omnidirectional stereo vision camera configuration system of the first item of the patent scope, which further includes an inertial sensor coupled to the computing unit, is used to return multiple degrees of freedom of motion information and posture information. 如申請專利範圍第1項的全向立體視覺的相機配置系統,其中,更包括耦接於該運算單元的一雷達模組,供以接收空間內存在物體所反射之無線電波,以供該運算單元計算出被攝物體與拍攝場景的深度資訊。The omnidirectional stereo vision camera configuration system as claimed in item 1 of the patent scope further includes a radar module coupled to the computing unit for receiving radio waves reflected by objects present in the space for the computing The unit calculates the depth information of the subject and shooting scene.
TW108215064U 2019-11-14 2019-11-14 Camera configuration system with omnidirectional stereo vision TWM594322U (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
TW108215064U TWM594322U (en) 2019-11-14 2019-11-14 Camera configuration system with omnidirectional stereo vision

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
TW108215064U TWM594322U (en) 2019-11-14 2019-11-14 Camera configuration system with omnidirectional stereo vision

Publications (1)

Publication Number Publication Date
TWM594322U true TWM594322U (en) 2020-04-21

Family

ID=71134049

Family Applications (1)

Application Number Title Priority Date Filing Date
TW108215064U TWM594322U (en) 2019-11-14 2019-11-14 Camera configuration system with omnidirectional stereo vision

Country Status (1)

Country Link
TW (1) TWM594322U (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112135021A (en) * 2020-09-15 2020-12-25 杭州海康威视数字技术股份有限公司 Photographing apparatus

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112135021A (en) * 2020-09-15 2020-12-25 杭州海康威视数字技术股份有限公司 Photographing apparatus
CN112135021B (en) * 2020-09-15 2022-03-01 杭州海康威视数字技术股份有限公司 Photographing apparatus

Similar Documents

Publication Publication Date Title
TWI555379B (en) An image calibrating, composing and depth rebuilding method of a panoramic fish-eye camera and a system thereof
WO2018076154A1 (en) Spatial positioning calibration of fisheye camera-based panoramic video generating method
CN110044300B (en) Amphibious three-dimensional vision detection device and detection method based on laser
WO2018068719A1 (en) Image stitching method and apparatus
CN106127745B (en) The combined calibrating method and device of structure light 3 D vision system and line-scan digital camera
US20170127045A1 (en) Image calibrating, stitching and depth rebuilding method of a panoramic fish-eye camera and a system thereof
CN111243002A (en) Monocular laser speckle projection system calibration and depth estimation method applied to high-precision three-dimensional measurement
JP2017108387A (en) Image calibrating, stitching and depth rebuilding method of panoramic fish-eye camera and system thereof
JP5070435B1 (en) Three-dimensional relative coordinate measuring apparatus and method
KR20160116075A (en) Image processing apparatus having a function for automatically correcting image acquired from the camera and method therefor
CN102243432A (en) Panoramic three-dimensional photographing device
JP2007192832A (en) Calibrating method of fish eye camera
US20210364900A1 (en) Projection Method of Projection System for Use to Correct Image Distortion on Uneven Surface
CN111854636B (en) Multi-camera array three-dimensional detection system and method
WO2018032841A1 (en) Method, device and system for drawing three-dimensional image
JP2010276433A (en) Imaging device, image processor, and distance measuring device
CN108205799B (en) Image splicing method and device
WO2021093804A1 (en) Omnidirectional stereo vision camera configuration system and camera configuration method
TWM594322U (en) Camera configuration system with omnidirectional stereo vision
CN110766752B (en) Virtual reality interactive glasses with light reflecting mark points and space positioning method
JP5648159B2 (en) Three-dimensional relative coordinate measuring apparatus and method
CN114993207B (en) Three-dimensional reconstruction method based on binocular measurement system
WO2023201578A1 (en) Extrinsic parameter calibration method and device for monocular laser speckle projection system
TWI725620B (en) Omnidirectional stereo vision camera configuration system and camera configuration method
WO2021068723A1 (en) Sensor calibration method and sensor calibration apparatus