CN107403451B - Adaptive binary feature monocular visual odometry method and computer and robot - Google Patents
Adaptive binary feature monocular visual odometry method and computer and robot Download PDFInfo
- Publication number
- CN107403451B CN107403451B CN201710459019.6A CN201710459019A CN107403451B CN 107403451 B CN107403451 B CN 107403451B CN 201710459019 A CN201710459019 A CN 201710459019A CN 107403451 B CN107403451 B CN 107403451B
- Authority
- CN
- China
- Prior art keywords
- feature
- image
- visual odometry
- search
- features
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 104
- 230000000007 visual effect Effects 0.000 title claims abstract description 52
- 230000003044 adaptive effect Effects 0.000 title claims abstract description 47
- 238000001514 detection method Methods 0.000 claims abstract description 33
- 238000005286 illumination Methods 0.000 claims abstract description 8
- 230000008569 process Effects 0.000 claims description 51
- 238000004364 calculation method Methods 0.000 claims description 13
- 238000006073 displacement reaction Methods 0.000 claims description 6
- 230000001629 suppression Effects 0.000 claims description 5
- 238000004891 communication Methods 0.000 claims description 4
- 239000011159 matrix material Substances 0.000 claims description 4
- 230000004044 response Effects 0.000 claims description 4
- 230000009467 reduction Effects 0.000 claims description 3
- 238000010845 search algorithm Methods 0.000 claims description 3
- FFBHFFJDDLITSX-UHFFFAOYSA-N benzyl N-[2-hydroxy-4-(3-oxomorpholin-4-yl)phenyl]carbamate Chemical compound OC1=C(NC(=O)OCC2=CC=CC=C2)C=CC(=C1)N1CCOCC1=O FFBHFFJDDLITSX-UHFFFAOYSA-N 0.000 claims description 2
- 230000002159 abnormal effect Effects 0.000 abstract description 2
- 238000010586 diagram Methods 0.000 description 8
- 230000008901 benefit Effects 0.000 description 4
- 230000008520 organization Effects 0.000 description 2
- 230000001360 synchronised effect Effects 0.000 description 2
- 230000003190 augmentative effect Effects 0.000 description 1
- 238000010276 construction Methods 0.000 description 1
- 238000003066 decision tree Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000007613 environmental effect Effects 0.000 description 1
- 238000003709 image segmentation Methods 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000008092 positive effect Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/73—Determining position or orientation of objects or cameras using feature-based methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/70—Denoising; Smoothing
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/70—Determining position or orientation of objects or cameras
- G06T7/77—Determining position or orientation of objects or cameras using statistical methods
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Life Sciences & Earth Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Bioinformatics & Computational Biology (AREA)
- Evolutionary Biology (AREA)
- Probability & Statistics with Applications (AREA)
- Image Analysis (AREA)
Abstract
本发明属于计算机视觉技术领域,公开了一种自适应二值特征单目视觉里程计方法及计算机、机器人,根据环境中的光照变化情况自适应地调整特征检测阈值,在不同的光照条件下稳定检测到恒定的特征数目;同时通过构造四叉树的方式组织检测到的特征,以深度优先搜索和广度优先搜索相结合的方法随机选择特征,使得最终输出特征均匀地分布在整幅图像上。本发明最终输出特征均匀地分布在整幅图像上,减少了动态目标上的特征异常值对位姿估算精度的影响。
The invention belongs to the technical field of computer vision, and discloses an adaptive binary feature monocular visual odometry method, a computer and a robot. The feature detection threshold is adaptively adjusted according to the illumination changes in the environment, and is stable under different illumination conditions. A constant number of features is detected; at the same time, the detected features are organized by constructing a quad-tree, and the features are randomly selected by a combination of depth-first search and breadth-first search, so that the final output features are evenly distributed on the entire image. The final output feature of the invention is evenly distributed on the whole image, which reduces the influence of the abnormal value of the feature on the dynamic target on the pose estimation accuracy.
Description
技术领域technical field
本发明属于计算机视觉技术领域,尤其涉及一种自适应二值特征单目视觉里程计方法及计算机、机器人。The invention belongs to the technical field of computer vision, and in particular relates to an adaptive binary feature monocular visual odometry method, a computer and a robot.
背景技术Background technique
在计算机视觉中,视觉里程计(Visual Odometry,VO)是指利用单个或多个相机的输入视频帧来估算移动机器人的位置和方向。不仅仅是移动机器人自动导航的核心技术之一,同时还在新兴的可穿戴设备计算、3D重构、虚拟现实、增强现实和自动驾驶等领域有着巨大的应用前景。视觉里程计方法从所使用的相机的种类和数量上,大体可以分为如下三类方法:第一类是基于双目相机的视觉里程计方法,突出优点是在每个时刻可以同时获得某一位置的不同视点的两幅图像,可以通过这两幅图像之间的已知视差关系直接计算出该时刻相机的深度信息,从而可以较为容易的知道场景中地图点所在的实际位置;存在的一些缺点是双目相机的安装配置不便,以及双目相机之间的视差固定而造成的大尺度户外环境和小尺度室内环境之间的切换不便。第二类是基于RGB-D相机的视觉里程计方法,RGB-D相机除了可以获得场景的图像之外,还集成有其它传感器可以直接获得场景的稠密深度地图,该类相机的代表是Kinect深度相机,缺点是硬件成本较高,且传感器探测的有效深度距离太短,从而适用的环境有限。第三类是基于单目相机的视觉里程计方法,只使用单个相机便可以进行定位与地图构建。但是,单目相机在每个时刻只能得到场景的一幅图像,场景中地图点的深度信息获取需要通过多幅图像之间的关联对应关系计算得到,因此深度信息的获取是单目视觉里程计中的一个难点。由于单目视觉中深度信息的计算是通过不同视点的图像之间的对应关系得到,所以计算出来地图点的深度值与实际深度值之间存在一个比例因子,深度的不确定性使得单目视觉里程计方法可以在大尺度户外环境和小尺度室内环境之间自由切换。因此,单目视觉里程计方法,具有硬件成本低和适用的环境范围广等优点。当相机运动场景中的环境状况良好时,现有的视觉里程计方法估算相机运动位姿轨迹,可以达到较好的实时性能和鲁棒性能。但是,现有的视觉里程计方法还存在如下一些缺点:当相机运动场景中环境光照太亮或太暗时,在此种光照条件下采集到的视频帧上,现有的视觉里程计方法往往不能提取到足够的特征,从而相机位姿轨迹估算就此中断,或者位姿估算的精度因异常点比例过多而大大降低;当相机运动场景中存在动态运动目标时,动态目标上检测到的特征会造成位姿估计错误;当相机的采集视频帧的帧率和视觉里程方法处理的视频帧速度不同步时,会出现跳帧现象,从而极易导致视觉里程方法轨迹估算中断。In computer vision, Visual Odometry (VO) refers to the use of input video frames from single or multiple cameras to estimate the position and orientation of a mobile robot. It is not only one of the core technologies of automatic navigation of mobile robots, but also has huge application prospects in the fields of emerging wearable device computing, 3D reconstruction, virtual reality, augmented reality and autonomous driving. The visual odometry method can be roughly divided into the following three categories according to the type and number of cameras used: the first category is the visual odometry method based on the binocular camera, the outstanding advantage is that a certain The depth information of the camera at the moment can be directly calculated from the two images of different viewpoints of the two images through the known disparity relationship between the two images, so that the actual location of the map point in the scene can be easily known; there are some The disadvantage is that the installation and configuration of the binocular cameras are inconvenient, and the switching between the large-scale outdoor environment and the small-scale indoor environment caused by the fixed parallax between the binocular cameras is inconvenient. The second type is the visual odometry method based on RGB-D camera. In addition to the image of the scene, the RGB-D camera also integrates other sensors to directly obtain the dense depth map of the scene. The representative of this type of camera is the Kinect depth map. The disadvantage of the camera is that the hardware cost is high, and the effective depth distance detected by the sensor is too short, so the applicable environment is limited. The third category is the visual odometry method based on a monocular camera, which can perform localization and map construction using only a single camera. However, the monocular camera can only obtain one image of the scene at each moment, and the acquisition of depth information of map points in the scene needs to be calculated through the correlation between multiple images, so the acquisition of depth information is the monocular visual mileage. a difficulty in the calculation. Since the calculation of depth information in monocular vision is obtained by the correspondence between images of different viewpoints, there is a scale factor between the calculated depth value of the map point and the actual depth value, and the uncertainty of depth makes monocular vision The odometry method can freely switch between large-scale outdoor environments and small-scale indoor environments. Therefore, the monocular visual odometry method has the advantages of low hardware cost and wide range of applicable environments. When the environmental conditions in the camera motion scene are good, the existing visual odometry methods estimate the camera motion pose trajectory, which can achieve better real-time performance and robust performance. However, the existing visual odometry methods still have the following shortcomings: when the ambient light in the camera motion scene is too bright or too dark, the existing visual odometry methods are often unable to capture video frames under such lighting conditions. Sufficient features are extracted, so that the camera pose trajectory estimation is interrupted, or the accuracy of pose estimation is greatly reduced due to the excessive proportion of outliers; when there are dynamic moving objects in the camera motion scene, the detected features on the dynamic objects will be Causes pose estimation errors; when the frame rate of the captured video frame of the camera and the frame rate of the video processed by the visual odometry method are not synchronized, frame skipping will occur, which can easily lead to the interruption of the trajectory estimation of the visual odometry method.
综上所述,现有技术存在的问题是:现有的视觉里程计方法存在不能提取到足够特征,相机位姿轨迹估算就此中断或者位姿估算的精度因异常点比例过多大大降低;当相机运动场景中存在动态运动目标时,动态目标上检测到的特征会造成位姿估计错误;当相机的采集视频帧的帧率和视觉里程方法处理的视频帧速度不同步时,会出现跳帧现象,导致视觉里程方法轨迹估算中断。To sum up, the problems existing in the prior art are: the existing visual odometry methods cannot extract enough features, and the camera pose trajectory estimation is interrupted or the accuracy of pose estimation is greatly reduced due to the excessive proportion of abnormal points; When there is a dynamic moving target in the camera motion scene, the detected features on the dynamic target will cause pose estimation errors; when the frame rate of the captured video frame of the camera and the frame rate of the video processed by the visual odometry method are not synchronized, frame skipping will occur. phenomenon, leading to the interruption of trajectory estimation by the visual odometry method.
发明内容SUMMARY OF THE INVENTION
针对现有技术存在的问题,本发明提供了一种自适应二值特征单目视觉里程计方法及计算机、机器人。Aiming at the problems existing in the prior art, the present invention provides an adaptive binary feature monocular visual odometry method, a computer and a robot.
本发明是这样实现的,一种自适应二值特征单目视觉里程计方法,所述自适应二值特征单目视觉里程计方法根据环境中的光照变化情况自适应地调整特征检测阈值,在不同的光照条件下稳定检测到恒定的特征数目;同时通过构造四叉树的方式组织检测到的特征,以深度优先搜索和广度优先搜索相结合的方法随机选择特征,使得最终输出特征均匀地分布在整幅图像上。The present invention is implemented in this way, an adaptive binary feature monocular visual odometry method, wherein the adaptive binary feature monocular visual odometry method adaptively adjusts the feature detection threshold according to the illumination changes in the environment, A constant number of features is stably detected under different lighting conditions; at the same time, the detected features are organized by constructing a quad-tree, and the features are randomly selected by a combination of depth-first search and breadth-first search, so that the final output features are evenly distributed. on the entire image.
进一步,所述自适应二值特征单目视觉里程计方法包括以下步骤:Further, the adaptive binary feature monocular visual odometry method includes the following steps:
步骤一,创建两个进程,分别记为Process I和Process II;进程Process I用于相机的视频帧采集,进程Process II用于视觉里程计进行相机位姿估计;In step 1, two processes are created, denoted as Process I and Process II respectively; the process Process I is used for the video frame acquisition of the camera, and the process Process II is used for the visual odometry to estimate the camera pose;
步骤二,开辟一块共享内存区域,用于进程Process I和Process II之间的数据通信;记这块共享内存区域为帧序列池,在帧序列池上维护一个先进先出的视频帧队列;Step 2, open up a shared memory area for data communication between Process I and Process II; record this shared memory area as a frame sequence pool, and maintain a first-in, first-out video frame queue on the frame sequence pool;
步骤三,在进程Process I中打开相机,采集视频帧;如果视频帧的分辨率过大,按照比例关系对视频帧降分辨率,将视频帧压入到帧序列池中;Step 3, open camera in process Process 1, collect video frame; If the resolution of video frame is too large, reduce resolution to video frame according to proportional relationship, video frame is pressed into frame sequence pool;
步骤四,进程Process II从共享内存池中按序取出视频帧图像;Step 4, the process Process II sequentially fetches the video frame images from the shared memory pool;
步骤五,计算灰度图像的直方图统计,每个直方图条块的宽度为16个像素;利用每个直方图条块的像素个数除以图像的总个数,得到每个直方图条块的统计概率;Step 5: Calculate the histogram statistics of the grayscale image. The width of each histogram block is 16 pixels; divide the number of pixels of each histogram block by the total number of images to obtain each histogram block. the statistical probability of the block;
步骤六,根据不同的图像熵大小设置不同的角点检测阈值,检测到不同的特征点个数;Step 6: Set different corner detection thresholds according to different image entropy sizes, and detect different numbers of feature points;
步骤七,使用的高斯平均算子对图像进行滤波,滤除噪声,将整幅图像分为小图像块,利用根据环境的情况自适应设置的阈值,通过AGAST角点检测算法分别在小图像块上进行角点特征检测;In step 7, the Gaussian average operator is used to filter the image, filter out noise, divide the whole image into small image blocks, and use the threshold value adaptively set according to the environment to separate the small image blocks through the AGAST corner detection algorithm. corner feature detection;
步骤八,利用非极大值抑制算法,获得局部图像块内的极大值最大响应的AGAST角点;Step 8: Use the non-maximum suppression algorithm to obtain the AGAST corner points with the maximum response in the local image block;
步骤九,如果非极大值抑制后,特征点的数目亦然大于指定要选取的特征点数目,随机均匀的选取指定数目的特征点集;Step 9, if after the non-maximum value is suppressed, the number of feature points is also greater than the specified number of feature points to be selected, randomly and uniformly select the specified number of feature point sets;
步骤十,计算得到的AGAST角点特征的方向;Step 10: Calculate the direction of the AGAST corner feature;
步骤十一,通过rBRIEF算法生成二值特征描述符,得到图像上检测到的自适应均匀二值特征;Step 11: Generate binary feature descriptors through the rBRIEF algorithm to obtain adaptive uniform binary features detected on the image;
步骤十二,通过自适应均匀二值特征之间的汉明距离和方向距离差,在当前帧和前一帧上搜索匹配的自适应均匀二值特征对应点集;Step 12: Search for a matching point set corresponding to the adaptive uniform binary feature on the current frame and the previous frame through the Hamming distance and the directional distance difference between the adaptive uniform binary features;
步骤十三,通过搜索到的前后帧之间的匹配特征点对应集,计算出相机的位姿矩阵。Step 13: Calculate the pose matrix of the camera through the searched corresponding sets of matching feature points between the front and rear frames.
进一步,所述步骤三中视频帧降分辨率的计算公式为:Further, the calculation formula of video frame resolution reduction in the step 3 is:
其中,(w,h)是原始图像的大小,(w′,h′)表示降分辨率后的图像大小。Among them, (w, h) is the size of the original image, and (w', h') represents the size of the image after downscaling.
进一步,所述步骤四中进程Process II从共享内存池中按序取出视频帧图像;图像是彩色图像,转换为灰度图像,记为Igray,对图像Igray执行自适应均匀二值特征检测算法,检测Igray上的自适应均匀二值特征。Further, in the described step 4, the process Process II takes out the video frame image in sequence from the shared memory pool; the image is a color image, and is converted into a grayscale image, which is denoted as I gray , and the adaptive uniform binary feature detection is performed to the image I gray . Algorithm to detect adaptive uniform binary features on I gray .
进一步,所述步骤五中计算Igray的图像熵Entropy(Igray),计算灰度图像的直方图统计,每个直方图条块的宽度为16个像素;利用每个直方图条块的像素个数除以图像的总个数,得到每个直方图条块的统计概率pi,图像熵便可按照如下公式计算:Further, in the described
进一步,所述步骤九中随机均匀选取指定数目特征点的方法具体包括:Further, the method for randomly and uniformly selecting a specified number of feature points in the step 9 specifically includes:
将图像上所有的特征点组织为一棵四叉树,使用广度优先搜索和深度优先搜索相结合的搜索算法来随机均匀的选取特征点,将图像按照长宽中轴递归地等分为左上、左下、右上和右下四块,直到图像块内没有特征点为止;利用BFS算法来随机搜索选择四叉树上的结点,如果某个结点是叶子结点且该结点只有一个特征点,那么选择该特征点;否则,继续随机搜索其它结点;如果一个结点不是叶子结点,需要在结点表示的图像区域内选取特征点,随机找到一个包含之前未选择的特征点的叶子结点;迭代,直到选择到指定数量的特征点数目。All feature points on the image are organized into a quadtree, and the search algorithm combining breadth-first search and depth-first search is used to randomly and uniformly select feature points, and the image is recursively divided into upper left and The lower left, upper right and lower right four blocks, until there are no feature points in the image block; use the BFS algorithm to randomly search and select the nodes on the quadtree, if a node is a leaf node and the node has only one feature point , then select the feature point; otherwise, continue to search for other nodes randomly; if a node is not a leaf node, you need to select a feature point in the image area represented by the node, and randomly find a leaf containing the feature point that was not selected before. Nodes; iterate until the specified number of feature points is selected.
进一步,所述步骤十中计算得到的AGAST角点特征的方向θ,计算公式如下:Further, the direction θ of the AGAST corner feature calculated in the tenth step, the calculation formula is as follows:
其中,Mij表示特征所在图像块的几何矩,按照Mij=∑x,yxiyj·I(x,y)计算得到;xc与yc表示特征所在图像块的质心,按照xc=M10/M00和公式yc=M01/M00计算得到;u′ij表示特征所在图像块的中心矩,由图像块的几何矩Mij和质心xc、yc计算出来。Among them, M ij represents the geometric moment of the image block where the feature is located, which is calculated according to M ij =∑ x, y x i y j ·I(x, y); x c and y c represent the centroid of the image block where the feature is located, according to x c = M 10 /M 00 and the formula y c =M 01 /M 00 are calculated; u′ ij represents the central moment of the image block where the feature is located, and is calculated from the geometric moment Mi ij of the image block and the centroids x c , y c .
进一步,所述步骤十二具体包括:Further, the step 12 specifically includes:
首先,设置搜索窗口的大小,每个特征点将在所在的邻域窗口内寻找匹配点;First, set the size of the search window, each feature point will find a matching point in the neighborhood window where it is located;
然后,寻找查询特征点pi所在邻域窗口内的所有测试特征点集{p′i},计算pi与{p′i}中的所有特征点描述符之间的汉明距离,记下最小的汉明距离与次小的汉明距离,以及它们所对应的测试特征点;当最小的汉明距离小于指定阈值tm,且最小的汉明距离与次小的汉明距离之间的比率小于某个比率rm,那么所对应的测试特征点便是与查询特征点相匹配的特征点;Then, find all the test feature point sets {p′ i } in the neighborhood window where the query feature point p i is located, calculate the Hamming distance between p i and all feature point descriptors in {p′ i }, and write down The smallest Hamming distance and the next smallest Hamming distance, and their corresponding test feature points; when the smallest Hamming distance is less than the specified threshold t m , and the smallest Hamming distance and the next smallest Hamming distance If the ratio is less than a certain ratio rm , then the corresponding test feature point is the feature point that matches the query feature point;
最后,利用Delaunay三角网格生成算法在当前帧的匹配点上生成三角网格,剔除三角网格中面积明显比大多数网格大的那些三角形,以及三角上远离大多数网格的顶点;计算三角网格中每个三角像素顶点相对于上一帧上匹配点的位移值,当位移值大于某一阈值时,剔除该匹配点;最终找到的前后两帧之间的匹配对应点集。Finally, use the Delaunay triangle mesh generation algorithm to generate a triangle mesh at the matching points of the current frame, and remove those triangles in the triangle mesh whose area is significantly larger than most meshes, as well as the vertices on the triangle that are far away from most meshes; calculate The displacement value of each triangle pixel vertex in the triangular mesh relative to the matching point on the previous frame, when the displacement value is greater than a certain threshold, the matching point is eliminated; the matching corresponding point set between the two frames before and after is finally found.
本发明的另一目的在于提供一种应用所述自适应二值特征单目视觉里程计方法的计算机。Another object of the present invention is to provide a computer applying the adaptive binary feature monocular visual odometry method.
本发明的另一目的在于提供一种应用所述自适应二值特征单目视觉里程计方法的机器人。Another object of the present invention is to provide a robot applying the adaptive binary feature monocular visual odometry method.
本发明的优点及积极效果为:可以根据环境中的光照变化情况自适应地调整特征检测阈值,可以在不同的光照条件下稳定检测到恒定的特征数目。从图11的(b)中可以看出,在(a)图相机运动轨迹场景光照变化剧烈的情况下,本方法提出的自适应均匀二值特征相比于ORB特征而言可以稳定的检测到指定数目的特征,并且两者的实时性能保持在同一数量级上。另外,通过构造四叉树的方式组织检测到的特征,以深度优先搜索和广度优先搜索相结合的方法随机选择特征,使得最终输出特征均匀地分布在整幅图像上,减少动态目标上的特征异常值对位姿估算精度的影响。从图10中可以看出,本方法提出的基于自适应均匀二值特征的单目视觉里程计方法估算出的相机运动轨迹与实际相机运动轨迹之间的拟合精度,要比基于ORB特征的单目视觉里程计方法估算出的相机运动轨迹高。The advantages and positive effects of the present invention are: the feature detection threshold can be adaptively adjusted according to the illumination changes in the environment, and a constant number of features can be stably detected under different illumination conditions. It can be seen from (b) of Fig. 11 that in the case of severe illumination changes in the camera motion trajectory scene in (a), the adaptive uniform binary feature proposed by this method can be stably detected compared to the ORB feature. A specified number of features, and the real-time performance of both remains on the same order of magnitude. In addition, the detected features are organized by constructing a quad-tree, and the features are randomly selected by a combination of depth-first search and breadth-first search, so that the final output features are evenly distributed on the entire image, reducing the features on dynamic targets. The effect of outliers on the accuracy of pose estimation. It can be seen from Figure 10 that the fitting accuracy between the camera motion trajectory estimated by the monocular visual odometry method based on adaptive uniform binary features and the actual camera motion trajectory proposed by this method is better than that based on ORB features. The camera motion trajectory estimated by the monocular visual odometry method is high.
附图说明Description of drawings
图1是本发明实施例提供的自适应二值特征单目视觉里程计方法流程图。FIG. 1 is a flowchart of a method for adaptive binary feature monocular visual odometry provided by an embodiment of the present invention.
图2是本发明实施例提供的自适应二值特征单目视觉里程计方法的实现流程图。FIG. 2 is a flowchart of the implementation of the method for self-adaptive binary feature monocular visual odometry provided by an embodiment of the present invention.
图3是本发明实施例提供的帧序列池的内存组织形式示意图。FIG. 3 is a schematic diagram of a memory organization form of a frame sequence pool provided by an embodiment of the present invention.
图4是本发明实施例提供的自适应均匀二值特征检测算法流程图。FIG. 4 is a flowchart of an adaptive uniform binary feature detection algorithm provided by an embodiment of the present invention.
图5是本发明实施例提供的图像对比度与图像熵之间的关系示意图。FIG. 5 is a schematic diagram of the relationship between image contrast and image entropy provided by an embodiment of the present invention.
图6是本发明实施例提供的图像未分块(a)和图像分块(b)后的AGAST角点检测结果示意图。FIG. 6 is a schematic diagram of an AGAST corner detection result after the image is not segmented (a) and the image is segmented (b) according to an embodiment of the present invention.
图7是本发明实施例提供的特征点随机均匀选择后的结果示意图。FIG. 7 is a schematic diagram of a result of random and uniform selection of feature points according to an embodiment of the present invention.
图8是本发明实施例提供的自适应均匀二值特征检测结果示意图。FIG. 8 is a schematic diagram of an adaptive uniform binary feature detection result provided by an embodiment of the present invention.
图9是本发明实施例提供的两帧之间搜索匹配的对应特征点集示意图。FIG. 9 is a schematic diagram of a corresponding feature point set for searching and matching between two frames according to an embodiment of the present invention.
图10是本发明实施例提供的基于自适应均匀二值特征与ORB特征的单目视觉里程计方法的位姿轨迹计算结果对比示意图;10 is a schematic diagram showing the comparison of the pose and trajectory calculation results of the monocular visual odometry method based on adaptive uniform binary features and ORB features provided by an embodiment of the present invention;
其中,(a)表示相机运动的实际轨迹;(b)中实线表示基于自适应均匀二值特征的单目视觉里程计方法,虚线表示基于ORB特征的单目视觉里程计方法。从(b)中可以看出基于自适应均匀二值特征的单目视觉里程计方法计算出来的相机运动轨迹,其拟合(a)图中的实际相机运动轨迹的精度要优于基于ORB特征的单目视觉里程计方法。Among them, (a) represents the actual trajectory of the camera motion; (b) the solid line represents the monocular visual odometry method based on adaptive uniform binary features, and the dotted line represents the monocular visual odometry method based on ORB features. It can be seen from (b) that the camera motion trajectory calculated by the monocular visual odometry method based on adaptive uniform binary features is more accurate in fitting the actual camera motion trajectory in (a) than based on the ORB feature. The monocular visual odometry method.
图11是本发明实施例提供的自适应均匀二值特征与ORB特征的性能对比图。FIG. 11 is a performance comparison diagram between an adaptive uniform binary feature and an ORB feature provided by an embodiment of the present invention.
具体实施方式Detailed ways
为了使本发明的目的、技术方案及优点更加清楚明白,以下结合实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。In order to make the objectives, technical solutions and advantages of the present invention clearer, the present invention will be further described in detail below with reference to the embodiments. It should be understood that the specific embodiments described herein are only used to explain the present invention, but not to limit the present invention.
下面结合附图对本发明的应用原理作详细的描述。The application principle of the present invention will be described in detail below with reference to the accompanying drawings.
如图1所示,本发明实施例提供的自适应二值特征单目视觉里程计方法包括以下步骤:As shown in FIG. 1 , the adaptive binary feature monocular visual odometry method provided by the embodiment of the present invention includes the following steps:
S101:创建两个进程,分别记为Process I和Process II;其中,进程Process I用于相机的视频帧采集,进程Process II用于视觉里程计进行相机位姿估计;S101: Create two processes, denoted as Process I and Process II respectively; wherein, the process Process I is used for video frame acquisition of the camera, and the process Process II is used for visual odometry to estimate the camera pose;
S102:开辟一块共享内存区域,用于进程Process I和Process II之间的数据通信。记这块共享内存区域为帧序列池,在帧序列池上维护了一个先进先出的视频帧队列;S102: Open up a shared memory area for data communication between Process I and Process II. Remember this shared memory area as the frame sequence pool, and maintain a first-in, first-out video frame queue on the frame sequence pool;
S103:在进程Process I中打开相机,采集视频帧。如果视频帧的分辨率过大,按照一定的比例关系对视频帧降分辨率,然后将视频帧压入到帧序列池中;S103: Turn on the camera in the process Process I to capture video frames. If the resolution of the video frame is too large, reduce the resolution of the video frame according to a certain proportional relationship, and then push the video frame into the frame sequence pool;
S104:进程Process II从共享内存池中按序取出视频帧图像;S104: The process Process II sequentially fetches video frame images from the shared memory pool;
S105:计算灰度图像的直方图统计,每个直方图条块的宽度为16个像素;利用每个直方图条块的像素个数除以图像的总个数,得到每个直方图条块的统计概率;S105: Calculate the histogram statistics of the grayscale image, the width of each histogram block is 16 pixels; divide the number of pixels of each histogram block by the total number of images to obtain each histogram block The statistical probability of ;
S106:根据不同的图像熵大小设置不同的角点检测阈值,以检测到不同的特征点个数;S106: Set different corner detection thresholds according to different image entropy sizes to detect different numbers of feature points;
S107:使用的高斯平均算子对图像进行滤波,滤除噪声,将整幅图像分为小图像块,利用根据环境的情况自适应设置的阈值,通过AGAST角点检测算法分别在这些小图像块上进行角点特征检测;S107: Use the Gaussian average operator to filter the image, filter out noise, divide the whole image into small image blocks, and use the threshold value adaptively set according to the environment to separate the small image blocks through the AGAST corner detection algorithm. corner feature detection;
S108:利用非极大值抑制算法,获得局部图像块内的极大值最大响应的AGAST角点;S108: Using the non-maximum value suppression algorithm, obtain the AGAST corner point of the maximum value maximum response in the local image block;
S109:如果非极大值抑制后,特征点的数目亦然大于指定要选取的特征点数目,那么随机均匀的选取指定数目的特征点集;S109: If the number of feature points is also greater than the specified number of feature points to be selected after the non-maximum value is suppressed, then randomly and uniformly select the specified number of feature point sets;
S110:计算得到的AGAST角点特征的方向;S110: the direction of the calculated AGAST corner feature;
S111:通过rBRIEF算法生成二值特征描述符,得到图像上检测到的自适应均匀二值特征;S111: Generate a binary feature descriptor through the rBRIEF algorithm to obtain an adaptive uniform binary feature detected on the image;
S112:通过自适应均匀二值特征之间的汉明距离和方向距离差,在当前帧和前一帧上搜索匹配的自适应均匀二值特征对应点集;S112: Search for a matching point set corresponding to the adaptive uniform binary feature on the current frame and the previous frame by using the Hamming distance and the directional distance difference between the adaptive uniform binary features;
S113:通过搜索到的前后帧之间的匹配特征点对应集,计算出相机的位姿矩阵。S113: Calculate the pose matrix of the camera through the searched corresponding sets of matching feature points between the front and rear frames.
下面结合附图对本发明的应用原理作进一步的描述。The application principle of the present invention will be further described below with reference to the accompanying drawings.
如图2所示,本发明实施例提供的自适应二值特征单目视觉里程计方法具体包括以下步骤:As shown in FIG. 2 , the adaptive binary feature monocular visual odometry method provided by the embodiment of the present invention specifically includes the following steps:
步骤1:创建两个进程,分别记为Process I和Process II。其中,进程Process I用于相机的视频帧采集,进程Process II用于视觉里程计进行相机位姿估计。Step 1: Create two processes, denoted as Process I and Process II. Among them, Process I is used for camera video frame acquisition, and Process II is used for visual odometry to estimate camera pose.
步骤2:开辟一块共享内存区域,用于进程Process I和Process II之间的数据通信。记这块共享内存区域为帧序列池,在帧序列池上维护了一个先进先出的视频帧队列,帧序列池的内存组织形式如图3。Step 2: Open up a shared memory area for data communication between Process I and Process II. Remember this shared memory area as the frame sequence pool. A first-in, first-out video frame queue is maintained on the frame sequence pool. The memory organization of the frame sequence pool is shown in Figure 3.
其中,帧序列池信息头中保留了一些标志位、队列头尾位置和图像的大小等信息。b表示队列中是否有图像,s表示队列头所在的位置,e表示队列尾所在的位置,w表示图像的宽,h表示图像的高,c表示的图像的通道数。Among them, some information such as flag bits, queue head and tail positions, and image size are reserved in the frame sequence pool information header. b indicates whether there is an image in the queue, s indicates the position of the queue head, e indicates the position of the queue tail, w indicates the width of the image, h indicates the height of the image, and c indicates the number of channels of the image.
步骤3:在进程Process I中打开相机,采集视频帧。如果视频帧的分辨率过大,按照一定的比例关系对视频帧降分辨率,然后将视频帧压入到帧序列池中。视频帧降分辨率的计算公式如下:Step 3: Open the camera in Process I to capture video frames. If the resolution of the video frame is too large, reduce the resolution of the video frame according to a certain proportional relationship, and then push the video frame into the frame sequence pool. The formula for calculating the video frame resolution reduction is as follows:
其中,(w,h)是原始图像的大小,(w′,h′)表示降分辨率后的图像大小。Among them, (w, h) is the size of the original image, and (w', h') represents the size of the image after downscaling.
步骤4:进程Process II从共享内存池中按序取出视频帧图像。如果图像是彩色图像,将其转换为灰度图像,记为Igray。首先,对图像Igray执行自适应均匀二值特征检测算法,检测Igray上的自适应均匀二值特征,如步骤5至步骤11,自适应均匀二值特征检测算法的流程图如图4。然后,搜索当前帧和前一帧之间的匹配对应特征点集,如步骤12。最后,通过前后帧之间匹配的对应点集计算出相机的位姿,如步骤13。Step 4: Process Process II sequentially fetches video frame images from the shared memory pool. If the image is a color image, convert it to a grayscale image, denoted as I gray . First, the adaptive uniform binary feature detection algorithm is performed on the image I gray to detect the adaptive uniform binary feature on the I gray , as shown in
步骤5:计算Igray的图像熵Entropy(Igray)。首先计算灰度图像的直方图统计,每个直方图条块的宽度为16个像素。然后利用每个直方图条块的像素个数除以图像的总个数,得到每个直方图条块的统计概率pi。那么,图像熵便可按照如下公式计算出来:Step 5: Calculate the image entropy Entropy (I gray ) of I gray . First calculate the histogram statistics of the grayscale image, each histogram bin is 16 pixels wide. Then divide the number of pixels of each histogram block by the total number of images to obtain the statistical probability p i of each histogram block. Then, the image entropy can be calculated according to the following formula:
步骤6:根据不同的图像熵大小设置不同的角点检测阈值,以检测到不同的特征点个数。熵较大的图像说明图像的对比度较高,反应了环境的光照状况良好,此时设置较高的角点检测阈值。熵较小的图像说明图像的对比度较小,反应了环境的光照太亮或太暗,或者是由相机的曝光引起的图像太暗或太亮,此时设置较低的角点检测阈值。由于划分了16个直方图条块,所以熵值的可能区间是[0,4]。图像熵与对比度之间的关系如图5。Step 6: Set different corner detection thresholds according to different image entropy values to detect different numbers of feature points. An image with a larger entropy indicates that the contrast of the image is higher, which reflects the good lighting conditions of the environment. At this time, a higher corner detection threshold is set. An image with smaller entropy indicates that the contrast of the image is smaller, reflecting that the ambient light is too bright or too dark, or the image is too dark or too bright caused by the exposure of the camera, and a lower corner detection threshold is set at this time. Since 16 histogram bins are divided, the possible interval for the entropy value is [0, 4]. The relationship between image entropy and contrast is shown in Figure 5.
在实际场景环境中计算相机采集的视频帧的图像熵,统计发现当熵值在[0,2.7]时图像的对比度很小,本发明将此时的角点检测阈值设置在10,当熵值在(2.7,3.0]时图像的对比度良好,本发明将此时的角点检测阈值设置在30,当熵值在(3.0,4]时将角点检测的阈值设置在40。也可以划分更小的图像熵值区间和阈值对应关系,但是没有必要建立图像熵值和特征检测阈值之间的连续函数关系,因为相差两三个像素的阈值对AGAST角点检测的结果影响并不是很大。The image entropy of the video frame collected by the camera is calculated in the actual scene environment, and it is found that the contrast of the image is very small when the entropy value is [0, 2.7]. The present invention sets the corner detection threshold at this time at 10, when the entropy value is When the image contrast is good at (2.7, 3.0], the present invention sets the corner detection threshold at 30, and when the entropy value is at (3.0, 4], the corner detection threshold is set at 40. It is also possible to divide more There is a small image entropy value interval and the corresponding relationship between the threshold value, but it is not necessary to establish a continuous functional relationship between the image entropy value and the feature detection threshold value, because the difference between the threshold value by two or three pixels does not greatly affect the results of AGAST corner detection.
步骤7:使用3×3的高斯平均算子对图像Igray进行滤波,滤除噪声。将整幅图像分为小图像块,利用步骤6根据环境的情况自适应设置的阈值,通过AGAST角点检测算法分别在这些小图像块上进行角点特征检测。Step 7: Use a 3×3 Gaussian average operator to filter the image I gray to filter out noise. Divide the whole image into small image blocks, and use the threshold value adaptively set according to the environment in step 6, and perform corner feature detection on these small image blocks through the AGAST corner detection algorithm.
当在整幅图像进行AGAST角点检测时,检测到的角点有在图像出现聚集的情况,图像未分块时检测的角点主要集中在左右的两颗树上,如图6(a)。当将整幅图像分为小图像块,分别在这些图像块上进行角点检测,可以使得各个区域都有合适数量的角点分布,可以在一定程度上抵消图像中的动态目标上的角点与其它异常值点的影响。When AGAST corner detection is performed on the entire image, the detected corners are sometimes clustered in the image. When the image is not divided into blocks, the detected corners are mainly concentrated on the left and right trees, as shown in Figure 6(a) . When the entire image is divided into small image blocks, and corner detection is performed on these image blocks, each area can have a suitable number of corner points, which can offset the corner points on the dynamic target in the image to a certain extent. and other outliers.
图像分块后检测的AGAST特征点数目更多了,主要是因为AGAST角点检测算法使用二叉决策树来判断选择最终的角点,图像分块后每一个小块上都有一个二叉决策树,相应的整幅图像上检测的特征点就更多了。同时,分块后每个区域上都检测到了特征,如图6(b)。The number of AGAST feature points detected after the image is divided into blocks is more, mainly because the AGAST corner detection algorithm uses a binary decision tree to determine the final corner point. After the image is divided into blocks, there is a binary decision on each small block. Tree, the corresponding feature points detected on the entire image are more. At the same time, features are detected on each region after segmentation, as shown in Figure 6(b).
图像块的大小建议不要太小,太小会导致图像分割的图像块过多而导致计算量增加,也不要太大,太大起不到分割图像块的效果。本发明在图像上实验测试选择的图像块大小是50像素,因为此时特征计算的时间最接近于对整幅图像计算特征的时间,而且特征点基本分布在整个图像区域上。It is recommended that the size of the image block should not be too small, which will cause too many image blocks for image segmentation and increase the amount of calculation. The size of the image block selected by the present invention in the experimental test on the image is 50 pixels, because the feature calculation time is the closest to the feature calculation time for the entire image, and the feature points are basically distributed on the entire image area.
步骤8:利用非极大值抑制算法,获得局部图像块内的极大值最大响应的AGAST角点。Step 8: Use the non-maximum suppression algorithm to obtain the AGAST corner points with the maximum response in the local image block.
步骤9:如果步骤8非极大值抑制后,特征点的数目亦然大于指定要选取的特征点数目,那么随机均匀的选取指定数目的特征点集,随机选择结果如图7。随机均匀选取指定数目特征点的方法如下:Step 9: If the number of feature points is also greater than the specified number of feature points to be selected after the non-maximum value suppression in step 8, then randomly and uniformly select the specified number of feature point sets, and the random selection result is shown in Figure 7. The method of randomly and uniformly selecting a specified number of feature points is as follows:
将图像上所有的特征点组织为一棵四叉树,使用广度优先搜索(Breath FirstSearch,BFS)和深度优先搜索(Depth First Search,DFS)相结合的搜索算法来随机均匀的选取特征点。首先,将图像按照长宽中轴递归地等分为左上、左下、右上和右下四块,直到该图像块内没有特征点为止;然后利用BFS算法来随机搜索选择四叉树上的结点,如果某个结点是叶子结点且该结点只有一个特征点,那么选择该特征点。否则,继续使用BFS算法来随机搜索其它结点。如果一个结点不是叶子结点,但是需要在结点表示的图像区域内选取特征点,那么使用DFS算法来随机找到一个包含之前未选择的特征点的叶子结点。迭代以上过程,直到选择到指定数量的特征点数目。All feature points on the image are organized into a quad-tree, and a search algorithm combining breadth first search (BFS) and depth first search (DFS) is used to randomly and uniformly select feature points. First, the image is recursively divided into upper left, lower left, upper right and lower right four blocks according to the central axis of length and width, until there are no feature points in the image block; then the BFS algorithm is used to randomly search and select the nodes on the quadtree , if a node is a leaf node and the node has only one feature point, then select the feature point. Otherwise, continue to use the BFS algorithm to randomly search for other nodes. If a node is not a leaf node, but a feature point needs to be selected in the image area represented by the node, then the DFS algorithm is used to randomly find a leaf node that contains a feature point that was not selected before. Iterate the above process until the specified number of feature points is selected.
步骤10:计算步骤9得到的AGAST角点特征的方向θ,计算公式如下:Step 10: Calculate the direction θ of the AGAST corner feature obtained in Step 9, and the calculation formula is as follows:
其中,Mij表示特征所在图像块的几何矩,按照公式Mij=∑x,yxiyj·I(x,y)计算得到。xc与yc表示特征所在图像块的质心,按照公式xc=M10/M00和公式yc=M01/M00计算得到。Wherein, M ij represents the geometric moment of the image block where the feature is located, and is calculated according to the formula M ij =∑ x, y x i y j ·I(x, y). x c and y c represent the centroid of the image block where the feature is located, and are calculated according to the formula x c =M 10 /M 00 and the formula y c =M 01 /M 00 .
步骤11:通过rBRIEF算法生成二值特征描述符,得到Igray上检测到的自适应均匀二值特征,如图8。Step 11: Generate a binary feature descriptor through the rBRIEF algorithm to obtain an adaptive uniform binary feature detected on I gray , as shown in Figure 8.
步骤12:通过自适应均匀二值特征之间的汉明距离和方向距离差,在当前帧和前一帧上搜索匹配的自适应均匀二值特征对应点集,如图9。Step 12: Search for a matching point set corresponding to the adaptive uniform binary feature on the current frame and the previous frame by using the Hamming distance and the directional distance difference between the adaptive uniform binary features, as shown in Figure 9.
首先,设置搜索窗口的大小,每个特征点将在它所在的邻域窗口内寻找匹配点。First, set the size of the search window, each feature point will find matching points within its neighborhood window.
然后,寻找查询特征点pi所在邻域窗口内的所有测试特征点集{p′i}。计算pi与{p′i}中的所有特征点描述符之间的汉明距离,记下最小的汉明距离与次小的汉明距离,以及它们所对应的测试特征点。当最小的汉明距离小于指定阈值tm,且最小的汉明距离与次小的汉明距离之间的比率小于某个比率rm,那么所对应的测试特征点便是与查询特征点相匹配的特征点。如果通过汉明距离没有找到pi的匹配点,那么再按照相同的方法利用特征的方向寻找一次。其中,阈值tm和比率rm根据图像熵Entropy(Igray)设置。当图像熵较小时,阈值tm和比率rm设置较大。当图像熵较大时,阈值tm和比率rm设置较小。Then, find all the test feature point sets {p′ i } in the neighborhood window where the query feature point p i is located. Calculate the Hamming distance between p i and all feature point descriptors in {p' i }, note the smallest Hamming distance and the next smallest Hamming distance, and their corresponding test feature points. When the smallest Hamming distance is less than the specified threshold t m , and the ratio between the smallest Hamming distance and the next smallest Hamming distance is less than a certain ratio rm , then the corresponding test feature point is the same as the query feature point. matching feature points. If the matching point of pi is not found through the Hamming distance, then use the direction of the feature to find it again in the same way. Among them, the threshold t m and the ratio rm are set according to the image entropy Entropy (I gray ). When the image entropy is small, the threshold t m and the ratio rm are set larger. When the image entropy is larger, the threshold t m and the ratio rm are set smaller.
最后,利用Delaunay三角网格生成算法在当前帧的匹配点上生成三角网格,剔除三角网格中面积明显比大多数网格大的那些三角形,以及该三角上远离大多数网格的顶点。然后计算三角网格中每个三角像素顶点相对于上一帧上匹配点的位移值,当位移值大于某一阈值时,剔除该匹配点。最终找到的前后两帧之间的匹配对应点集。Finally, the Delaunay triangle mesh generation algorithm is used to generate a triangle mesh at the matching points of the current frame, and those triangles in the triangle mesh whose area is significantly larger than most meshes are culled, as well as the vertices on the triangle that are far away from most meshes. Then, the displacement value of each triangle pixel vertex in the triangular mesh relative to the matching point on the previous frame is calculated, and when the displacement value is greater than a certain threshold, the matching point is eliminated. The matching corresponding point set between the two frames before and after is finally found.
步骤13:通过步骤12搜索到的前后帧之间的匹配特征点对应集,计算出相机的位姿矩阵。相机位姿的计算分为两种情况:当系统刚初始化时,通过前后帧之间的匹配特征点对应集,采用基于RANSAC异常值拒绝的八点算法计算相机的位姿,并三角化重构出特征点所对应的3D场景结构点;当系统初始化完成后,找到重构的3D场景结构点与当前帧上的特征点关联关系,采用基于代数异常值拒绝的PnP算法计算相机的位姿,并三角化重构出特征点所对应的3D场景结构点。Step 13: Calculate the pose matrix of the camera through the corresponding sets of matching feature points between the front and rear frames searched in Step 12. The calculation of the camera pose is divided into two cases: when the system is just initialized, the camera pose is calculated by the eight-point algorithm based on RANSAC outlier rejection based on the corresponding set of matching feature points between the front and rear frames, and the triangulation reconstruction is performed. The 3D scene structure points corresponding to the feature points are obtained; when the system is initialized, the relationship between the reconstructed 3D scene structure points and the feature points on the current frame is found, and the PnP algorithm based on algebraic outlier rejection is used to calculate the pose of the camera. And the 3D scene structure points corresponding to the feature points are reconstructed by triangulation.
基于自适应均匀二值特征的单目视觉里程计方法和基于ORB特征的单目视觉里程计方法的位姿计算结果对比结果,如图10。The comparison results of the pose calculation results of the monocular visual odometry method based on the adaptive uniform binary feature and the monocular visual odometry method based on the ORB feature are shown in Figure 10.
以上所述仅为本发明的较佳实施例而已,并不用以限制本发明,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,均应包含在本发明的保护范围之内。The above descriptions are only preferred embodiments of the present invention and are not intended to limit the present invention. Any modifications, equivalent replacements and improvements made within the spirit and principles of the present invention shall be included in the protection of the present invention. within the range.
Claims (9)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710459019.6A CN107403451B (en) | 2017-06-16 | 2017-06-16 | Adaptive binary feature monocular visual odometry method and computer and robot |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710459019.6A CN107403451B (en) | 2017-06-16 | 2017-06-16 | Adaptive binary feature monocular visual odometry method and computer and robot |
Publications (2)
Publication Number | Publication Date |
---|---|
CN107403451A CN107403451A (en) | 2017-11-28 |
CN107403451B true CN107403451B (en) | 2020-11-10 |
Family
ID=60404767
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710459019.6A Active CN107403451B (en) | 2017-06-16 | 2017-06-16 | Adaptive binary feature monocular visual odometry method and computer and robot |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN107403451B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108010081B (en) * | 2017-12-01 | 2021-12-17 | 中山大学 | RGB-D visual odometer method based on Census transformation and local graph optimization |
CN108182438B (en) * | 2018-01-17 | 2020-09-25 | 清华大学 | Graph binary feature learning method and device based on deep reinforcement learning |
CN109146972B (en) * | 2018-08-21 | 2022-04-12 | 南京师范大学镇江创新发展研究院 | Visual navigation method based on rapid feature point extraction and gridding triangle constraint |
CN109827591B (en) * | 2019-01-23 | 2022-05-17 | 西安微电子技术研究所 | Odometer signal detection circuit and method suitable for multiple input signals |
CN110033499B (en) * | 2019-04-21 | 2021-01-15 | 南京林业大学 | Landsat OLI image-based water body mapping method |
CN114066824B (en) * | 2021-10-28 | 2024-05-14 | 华南理工大学 | Binocular vision odometer method with dynamic target detection function |
Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101937078A (en) * | 2009-06-30 | 2011-01-05 | 深圳市气象局 | Nowcasting method and system of thunder cloud cluster based on boundary recognition and tracer technique |
CN102982561A (en) * | 2012-12-31 | 2013-03-20 | 哈尔滨工业大学 | Method for detecting binary robust scale invariable feature of color of color image |
CN103112015A (en) * | 2013-01-29 | 2013-05-22 | 山东电力集团公司电力科学研究院 | Operating object position and posture recognition method applicable to industrial robot |
CN103258202A (en) * | 2013-05-02 | 2013-08-21 | 电子科技大学 | Method for extracting textural features of robust |
CN105426841A (en) * | 2015-11-19 | 2016-03-23 | 北京旷视科技有限公司 | Human face detection based monitor camera self-positioning method and apparatus |
CN105975971A (en) * | 2016-04-22 | 2016-09-28 | 安徽大学 | Low-memory image feature description algorithm |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
EP3038017A1 (en) * | 2014-12-23 | 2016-06-29 | Thomson Licensing | Method of generating descriptor for interest point in image and apparatus implementing the same |
-
2017
- 2017-06-16 CN CN201710459019.6A patent/CN107403451B/en active Active
Patent Citations (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101937078A (en) * | 2009-06-30 | 2011-01-05 | 深圳市气象局 | Nowcasting method and system of thunder cloud cluster based on boundary recognition and tracer technique |
CN102982561A (en) * | 2012-12-31 | 2013-03-20 | 哈尔滨工业大学 | Method for detecting binary robust scale invariable feature of color of color image |
CN103112015A (en) * | 2013-01-29 | 2013-05-22 | 山东电力集团公司电力科学研究院 | Operating object position and posture recognition method applicable to industrial robot |
CN103258202A (en) * | 2013-05-02 | 2013-08-21 | 电子科技大学 | Method for extracting textural features of robust |
CN105426841A (en) * | 2015-11-19 | 2016-03-23 | 北京旷视科技有限公司 | Human face detection based monitor camera self-positioning method and apparatus |
CN105975971A (en) * | 2016-04-22 | 2016-09-28 | 安徽大学 | Low-memory image feature description algorithm |
Also Published As
Publication number | Publication date |
---|---|
CN107403451A (en) | 2017-11-28 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN107403451B (en) | Adaptive binary feature monocular visual odometry method and computer and robot | |
CN111798475A (en) | A method for constructing 3D semantic map of indoor environment based on point cloud deep learning | |
CN111724439A (en) | A visual positioning method and device in a dynamic scene | |
CN103824070B (en) | A kind of rapid pedestrian detection method based on computer vision | |
CN113674416B (en) | Three-dimensional map construction method and device, electronic equipment and storage medium | |
CN110097553A (en) | The semanteme for building figure and three-dimensional semantic segmentation based on instant positioning builds drawing system | |
CN112364865B (en) | A detection method for moving small objects in complex scenes | |
CN103886107B (en) | Robot localization and map structuring system based on ceiling image information | |
CN110599522B (en) | Method for detecting and removing dynamic target in video sequence | |
CN110287907B (en) | Object detection method and device | |
CN112446882B (en) | A robust visual SLAM method based on deep learning in dynamic scenes | |
CN110097584A (en) | The method for registering images of combining target detection and semantic segmentation | |
CN113223045A (en) | Vision and IMU sensor fusion positioning system based on dynamic object semantic segmentation | |
CN110599545B (en) | Feature-based dense map construction system | |
CN114140527B (en) | Dynamic environment binocular vision SLAM method based on semantic segmentation | |
CN110688905A (en) | A Keyframe-Based 3D Object Detection and Tracking Method | |
CN112207821B (en) | Target searching method of visual robot and robot | |
CN104182968B (en) | The fuzzy moving-target dividing method of many array optical detection systems of wide baseline | |
CN103729620B (en) | A kind of multi-view pedestrian detection method based on multi-view Bayesian network | |
CN108280844A (en) | A kind of video object localization method based on the tracking of region candidate frame | |
CN105809108B (en) | Pedestrian positioning method and system based on distributed vision | |
CN111951297A (en) | A target tracking method based on structured pixel-by-pixel target attention mechanism | |
CN108961385B (en) | SLAM composition method and device | |
CN113012196B (en) | Positioning method based on information fusion of binocular camera and inertial navigation sensor | |
CN108229281B (en) | Neural network generation method, face detection device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |