WO2022253027A1 - Underwater sonar image matching method based on gaussian distribution clustering - Google Patents

Underwater sonar image matching method based on gaussian distribution clustering Download PDF

Info

Publication number
WO2022253027A1
WO2022253027A1 PCT/CN2022/094444 CN2022094444W WO2022253027A1 WO 2022253027 A1 WO2022253027 A1 WO 2022253027A1 CN 2022094444 W CN2022094444 W CN 2022094444W WO 2022253027 A1 WO2022253027 A1 WO 2022253027A1
Authority
WO
WIPO (PCT)
Prior art keywords
sonar
image
map
target
steps
Prior art date
Application number
PCT/CN2022/094444
Other languages
French (fr)
Chinese (zh)
Inventor
王慧
邱海洋
智鹏飞
董苗
Original Assignee
江苏科技大学
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 江苏科技大学 filed Critical 江苏科技大学
Publication of WO2022253027A1 publication Critical patent/WO2022253027A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/751Comparing pixel values or logical combinations thereof, or feature values having positional relevance, e.g. template matching
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • G06F18/232Non-hierarchical techniques
    • G06F18/2321Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions
    • G06F18/23213Non-hierarchical techniques using statistics or function optimisation, e.g. modelling of probability density functions with fixed number of clusters, e.g. K-means clustering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/70Denoising; Smoothing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/44Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/70Arrangements for image or video recognition or understanding using pattern recognition or machine learning
    • G06V10/74Image or video pattern matching; Proximity measures in feature spaces
    • G06V10/75Organisation of the matching processes, e.g. simultaneous or sequential comparisons of image or video features; Coarse-fine approaches, e.g. multi-scale approaches; using context analysis; Selection of dictionaries
    • G06V10/757Matching configurations of points or features

Definitions

  • Step A extract features, establish the matching relationship between features, and obtain the height information of the map; specifically, it includes the following three steps:
  • Step A1 collecting sonar data and performing feature detection
  • Step A2 image registration, generating a Gaussian map
  • Step A3 calculating the scene elevation map
  • Step B perform pose estimation, and update the feature map to generate a three-dimensional space map; specifically, the following steps are included:
  • Step B1 pose estimation
  • performing feature extraction includes the following 4 steps:
  • Step A1-1 collecting sonar data, selecting an area with obvious pixels from the sonar data image as the target processing area;
  • Step A1-2 use the gray level distribution information of the target feature area and analyze the scene to divide the target area into three categories, distinguish the target pixel area and other non-target areas, and eliminate the pixels caused by noise according to the gray level distribution information Region, to realize the description of feature points.
  • step A-1-2 includes the following two steps:
  • Step A1-2a using the gray distribution information of the target feature area and through scene analysis, the relatively obvious gray value in the sonar image is divided into three categories: bright small spots that form 3D targets or structures, and adjacent to 3D targets The shaded area cast by the surface of , a flat surface between the first two grayscale values;
  • Step A1-2b through scene analysis, take the bright small spots that make up the 3D target or structure as the target area, distinguish the target pixels from other non-target areas, remove the pixel area due to noise, and reduce the image error.
  • image registration in step A2 generating a Gaussian map includes the following steps:
  • Step A2-2 using low-pass filtering to remove noise and reduce errors in feature extraction.
  • step A2-1 includes the following steps:
  • step A2-1a the target feature points obtained in step A1-2b are gridded, and based on the points in the grid, each gridded target pixel is clustered in the grid through the K-means clustering method;
  • Step A2-1b use Gaussian distribution to represent the target feature points clustered by k-means, calculate the eigenvalues and eigenvectors of the covariance matrix to reflect the direction and smoothness of the surface, and use the cluster mean and covariance to represent any Clustering, generating a Gaussian map.
  • the mean vector and covariance matrix correspond to the location, size and orientation of each 2D image region.
  • Step A2-1b-1 select 1%-2% of the pixels with high brightness, remove the pixels less than 8, select the k value so that any sub-region is not larger than 32 pixels, and divide the region larger than 32 pixels into smaller regions, Get a suitable feature area;
  • Step A2-1b-2 calculating the mean value vector and covariance matrix corresponding to the position, size and direction of each two-dimensional image region, and generating a Gaussian map.
  • step A3 calculating the scene elevation map in step A3 includes the following steps:
  • step B1 pose estimation includes the following steps:
  • step B1-1 in the scene elevation map obtained in step A3-4, select the first frame of sonar image as the reference image, and the second frame of image as the image to be matched, and each image point is moved to the new sonar image after rigid sonar motion.
  • the position of the two sonar images is registered by calculating the frame-to-frame motion parameters;
  • step B2 After the parameters are optimized according to step B2, the calculation speed is accelerated to obtain the optimal solution, and the sonar trajectory relative to the initial position is calculated to complete accurate three-dimensional reconstruction.
  • Fig. 4 is an image processed by seabed sonar image and k-means clustering in the embodiment of the present invention.
  • Fig. 6 is an elevation map calculated by using elevation angle information in an embodiment of the present invention.
  • Step B Perform pose estimation and update the feature map to generate a three-dimensional space map.
  • step A1 collects sonar data and performs feature detection
  • step A2 image registration to generate a Gaussian map
  • step A3 calculates the scene elevation map.
  • Step A1-2 use the gray level distribution information of the target feature area and analyze the scene to divide the target area into three categories, distinguish the target pixel area and other non-target areas, and eliminate the pixels caused by noise according to the gray level distribution information Region, to realize the description of feature points.
  • Step A2-1b use Gaussian distribution to represent the target feature points clustered by k-means, calculate the eigenvalues and eigenvectors of the covariance matrix by formula 1 and formula 2 to reflect information such as the direction and smoothness of the surface, and use the cluster mean and covariance represent any clustering, yielding a Gaussian map.
  • the mean vector and covariance matrix correspond to the location, size and orientation of each 2D image region.
  • Feature region j is represented by Gaussian mean value ⁇ j and variance ⁇ j as shown in Figure 3.
  • the cluster-based Gaussian map (statistical mean, covariance) representation eliminates the need to maintain a cumbersome grid structure, while Reduce computation time during optimization without losing precision.
  • Step A3-1 calculate according to the method of planarity assumption, estimate the elevation angle of the scene relative to the flat part and the point on the three-dimensional object.
  • the planarity assumption basically estimates the elevation angles of points on relatively flat parts of the scene and on 3-D objects (based on the shadows they cast).
  • R is the rotation matrix
  • W is the rotation velocity vector
  • the new position of the 3D scene point is calculated, and two sonar views are obtained.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Evolutionary Computation (AREA)
  • Data Mining & Analysis (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Computing Systems (AREA)
  • Databases & Information Systems (AREA)
  • Medical Informatics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Probability & Statistics with Applications (AREA)
  • Evolutionary Biology (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

An underwater sonar image matching method based on Gaussian distribution clustering. Accurate three-dimensional reconstruction is performed on a two-dimensional sonar image by means of performing image registration and optimizing sonar three-dimensional motion parameters. The method comprises the following steps: step A: extracting features, establishing a matching relationship between the features, and obtaining elevation information of a map; and step B: performing position and posture estimation, and updating a feature map, so as to generate a three-dimensional space map. By means of the present invention, feature extraction to reconstruction of an environment map covering elevation information are realized for a sonar image, and certain motion posture estimation information is also provided, which can be used in the field of underwater robot sonar image processing and mapping.

Description

基于高斯分布聚类的水下声呐图像匹配方法Underwater sonar image matching method based on Gaussian distribution clustering 技术领域technical field
本发明涉及船舶与海洋技术和声纳图像处理领域,特别涉及水下探测与感知以及目标三维重建的方法,具体提出一种基于高斯分布聚类的水下声呐图像匹配方法。The invention relates to the fields of ship and marine technology and sonar image processing, in particular to methods for underwater detection and perception and three-dimensional reconstruction of targets, and specifically proposes an underwater sonar image matching method based on Gaussian distribution clustering.
背景技术Background technique
在过去的十年中,各种基于特征、基于模板、基于区域以及基于傅立叶的方法的二维FS声纳图像配准技术已经被探索,基于各种技术的高分辨率二维(2D)多波束前向扫描(FS)声纳视频系统已经商业化,可用于浑浊水域的操作和成像。但是以往的研究一般基于单个的特征点,然后图像配准是制定作为一个优化问题,采用正态分布转换。但是网格大小的选择涉及分辨率和计算时间之间的权衡,并不一定是自动的。点的分布不服从单变量高斯分布,以往的高斯分布计算会导致真实分布的错误表示,复杂网格结构计算优化时间过长。生成高程图时3D投射到2D声纳图像上仰角会丢失,使重建的目标误差大,以及优化方法的不理想影响图像配准和精确的三维重建。In the past decade, various feature-based, template-based, region-based, and Fourier-based methods for two-dimensional FS sonar image registration have been explored, and high-resolution two-dimensional (2D) Forward-scanning beam (FS) sonar video systems are commercially available for operation and imaging in turbid waters. But previous studies are generally based on single feature points, and then image registration is formulated as an optimization problem, using a normal distribution transformation. But the choice of grid size involves a trade-off between resolution and computation time and is not necessarily automatic. The distribution of points does not obey the univariate Gaussian distribution. The previous calculation of Gaussian distribution will lead to the wrong representation of the real distribution, and the optimization time of complex grid structure calculation is too long. When the elevation map is generated, the elevation angle will be lost when the 3D is projected onto the 2D sonar image, which will cause a large error in the reconstruction target, and the imperfection of the optimization method will affect image registration and accurate 3D reconstruction.
发明内容Contents of the invention
本发明的目的是开发一种基于高斯分布聚类的水下声呐图像匹配方法,通过图像配准和优化声纳三维运动参数,通过二维声纳图像进行水下环境三维重建。The purpose of the present invention is to develop an underwater sonar image matching method based on Gaussian distribution clustering, through image registration and optimization of sonar three-dimensional motion parameters, to carry out three-dimensional underwater environment reconstruction through two-dimensional sonar images.
一种基于高斯分布聚类的水下声呐图像匹配方法,包括以下步骤:A method for matching underwater sonar images based on Gaussian distribution clustering, comprising the following steps:
步骤A,提取特征,建立特征间的匹配关系,获得地图的高度信息;具体包括以下3个步骤:Step A, extract features, establish the matching relationship between features, and obtain the height information of the map; specifically, it includes the following three steps:
步骤A1,采集声纳数据,进行特征检测;Step A1, collecting sonar data and performing feature detection;
步骤A2,图像配准,生成高斯地图;Step A2, image registration, generating a Gaussian map;
步骤A3,计算出场景高程地图;Step A3, calculating the scene elevation map;
步骤B,进行位姿估计,将特征地图更新,从而生成三维空间地图;具体包括以下步骤:Step B, perform pose estimation, and update the feature map to generate a three-dimensional space map; specifically, the following steps are included:
步骤B1,位姿估计;Step B1, pose estimation;
步骤B2,参数优化;Step B2, parameter optimization;
步骤B3,生成三维地图。Step B3, generating a three-dimensional map.
进一步地,步骤A1中采集声纳数据,进行特征提取包括以下4个步骤:Further, collecting sonar data in step A1, performing feature extraction includes the following 4 steps:
步骤A1-1,采集声纳数据,从声纳数据图像中选取像素明显的区域为目标处理区域;Step A1-1, collecting sonar data, selecting an area with obvious pixels from the sonar data image as the target processing area;
步骤A1-2,利用目标特征区域灰度分布信息并通过场景分析,将目标区域分为三个类别,分辨出目标像素区域和其他非目标区域,根据灰度分布信息,消除由于噪声产生的像素区域,实现特征点的描述。Step A1-2, use the gray level distribution information of the target feature area and analyze the scene to divide the target area into three categories, distinguish the target pixel area and other non-target areas, and eliminate the pixels caused by noise according to the gray level distribution information Region, to realize the description of feature points.
进一步地,其特征在于,步骤A-1-2包括以下2个步骤:Further, it is characterized in that step A-1-2 includes the following two steps:
步骤A1-2a,利用目标特征区域灰度分布信息并通过场景分析,将声纳图像中相对明显的灰度值分为三种类别:组成3D目标或结构的明亮小斑点、与3D目标相邻的表面投射的阴影区、介于前两个灰度值之间的平坦表面;Step A1-2a, using the gray distribution information of the target feature area and through scene analysis, the relatively obvious gray value in the sonar image is divided into three categories: bright small spots that form 3D targets or structures, and adjacent to 3D targets The shaded area cast by the surface of , a flat surface between the first two grayscale values;
步骤A1-2b,通过场景分析,将组成3D目标或结构的明亮小斑点作为目标区域,将目标像素和其他非目标区域分辨出,去除由于噪声产生的像素区域,减少图像误差。Step A1-2b, through scene analysis, take the bright small spots that make up the 3D target or structure as the target area, distinguish the target pixels from other non-target areas, remove the pixel area due to noise, and reduce the image error.
进一步地,步骤A2中图像配准,生成高斯地图包括以下步骤:Further, image registration in step A2, generating a Gaussian map includes the following steps:
步骤A2-1,通过K-means聚类方法将像素强度较强的目标像素聚类;Step A2-1, clustering target pixels with strong pixel intensity by K-means clustering method;
步骤A2-2,采用低通滤波去除噪声,减少对特征提取的误差。Step A2-2, using low-pass filtering to remove noise and reduce errors in feature extraction.
进一步地,步骤A2-1包括以下步骤:Further, step A2-1 includes the following steps:
步骤A2-1a,对步骤A1-2b得到的目标特征点网格化,基于网格内的点,通过K-means聚类法,将每个网格化目标像素进行网格内聚类;In step A2-1a, the target feature points obtained in step A1-2b are gridded, and based on the points in the grid, each gridded target pixel is clustered in the grid through the K-means clustering method;
步骤A2-1b,用高斯分布表示k-means聚类过的目标特征点,计算协方差矩阵的特征值和特征向量反应表面的方向和光滑性等信息,用聚类均值和协方差表示任一聚类,生成高斯地图。均值向量和协方差矩阵对应于每个二维图像区域的位置、大小和方向。Step A2-1b, use Gaussian distribution to represent the target feature points clustered by k-means, calculate the eigenvalues and eigenvectors of the covariance matrix to reflect the direction and smoothness of the surface, and use the cluster mean and covariance to represent any Clustering, generating a Gaussian map. The mean vector and covariance matrix correspond to the location, size and orientation of each 2D image region.
进一步地,步骤A2-1b包括以下2个步骤:Further, step A2-1b includes the following two steps:
步骤A2-1b-1,选取1%-2%具有高亮度的像素,剔除小于8的像素,选取k值使任何子区域不大于32像素,将大于32像素的区域划分为较小的区域,得到合适的特征区域;Step A2-1b-1, select 1%-2% of the pixels with high brightness, remove the pixels less than 8, select the k value so that any sub-region is not larger than 32 pixels, and divide the region larger than 32 pixels into smaller regions, Get a suitable feature area;
步骤A2-1b-2,计算均值向量和协方差矩阵对应于每个二维图像区域的位置、大小和方向,生成高斯地图。Step A2-1b-2, calculating the mean value vector and covariance matrix corresponding to the position, size and direction of each two-dimensional image region, and generating a Gaussian map.
进一步地,步骤A3中计算场景高程图包括以下步骤:Further, calculating the scene elevation map in step A3 includes the following steps:
步骤A3-1,根据平面性假设的方法,计算扫描方位的投射阴影点的仰角的测量值;Step A3-1, according to the planarity assumption method, calculate the measurement value of the elevation angle of the projected shadow point of the scanning azimuth;
步骤A3-2,使用最小和最大距离值的图像点来固定平面,即前端点和后缘点的三维坐标是固定的,在不同方位扫描出类似的点对,通过投射阴影确定遮挡轮廓点建立目标物体仰角与阴影的对应关系,每个三维对象从前端到后端的高程值由线性插值填充;Step A3-2, use the image points of the minimum and maximum distance values to fix the plane, that is, the three-dimensional coordinates of the front point and the trailing edge point are fixed, similar point pairs are scanned in different directions, and the occlusion contour points are determined by projecting shadows to establish The corresponding relationship between the elevation angle of the target object and the shadow, the elevation value of each 3D object from the front end to the back end is filled by linear interpolation;
步骤A3-3,通过低通滤波对平滑后的数据进行处理,以减小图像噪声对局部峰值的影响;通过k-means分割,建立背景、对象和阴影区域的阈值,即聚类后根据灰度值分为三种类别,划分图像类别定位物体到阴影和阴影到地面的转换;Step A3-3, process the smoothed data through low-pass filtering to reduce the influence of image noise on local peaks; through k-means segmentation, establish thresholds for background, objects and shadow areas, that is, after clustering, according to gray The degree value is divided into three categories, dividing the image category to locate the conversion of objects to shadows and shadows to the ground;
步骤A3-4,通过后缘点的仰角和遮挡轮廓的角度估算出三维物体的大小,利用下式计算出场景高程图:In step A3-4, the size of the 3D object is estimated by the elevation angle of the trailing edge point and the angle of the occlusion contour, and the scene elevation map is calculated using the following formula:
Figure PCTCN2022094444-appb-000001
Figure PCTCN2022094444-appb-000001
式中,Hs为声纳探测到海底的距离,Rs为斜距,Ls为阴影长度,Ht为目标物高度,Lt为目标物长度。In the formula, Hs is the distance from the sonar to the bottom, Rs is the slant distance, Ls is the shadow length, Ht is the height of the target, and Lt is the length of the target.
进一步地,步骤B1位姿估计包括以下步骤:Further, step B1 pose estimation includes the following steps:
步骤B1-1,在步骤A3-4得到的场景高程图中,选取第一帧声纳图像作为参考图像,第二帧图像作为待匹配图像,每个图像点都通过声纳刚性运动后到新的位置,通过计算帧到帧的运动参数来对两张声纳图像进行配准;In step B1-1, in the scene elevation map obtained in step A3-4, select the first frame of sonar image as the reference image, and the second frame of image as the image to be matched, and each image point is moved to the new sonar image after rigid sonar motion. The position of the two sonar images is registered by calculating the frame-to-frame motion parameters;
步骤B1-2,计算声纳运动参数,寻求最适合在转换特性从参考图像到待匹配图像,用空间转换函数进行位姿估计,对两幅图像的变换即旋转平移等对比,得到最好的变换参数;然后每次将前一次得到的配准图像作为参考,对下一帧声纳图像进行配准,直至整个图像序列所有待匹配声纳图像配准完成,将所有相邻帧的配准统一到相邻的参考系中,减少成对配准造成的积累误差。Step B1-2, calculate the sonar motion parameters, seek the most suitable transformation characteristics from the reference image to the image to be matched, use the space transformation function to estimate the pose, and compare the transformation of the two images, such as rotation and translation, to obtain the best Transform parameters; then each time the previous registration image is used as a reference to register the next frame of sonar images until the registration of all sonar images to be matched in the entire image sequence is completed, and the registration of all adjacent frames Uniform into adjacent frames of reference to reduce accumulated errors caused by pairwise registrations.
进一步地,步骤B2参数优化包括以下步骤:Further, step B2 parameter optimization includes the following steps:
步骤B2-1,由步骤A3-4得到的高程图每个图像点都投影到相应的空间点,经过声纳刚性运动后计算出新的位置三维场景点,得到两个声纳视图;In step B2-1, each image point of the elevation map obtained in step A3-4 is projected to the corresponding space point, and after sonar rigid motion, the new position 3D scene point is calculated to obtain two sonar views;
步骤B2-2,对两个声纳视图参数优化,将第二个视图的三维点变换到第一个视图的坐标系中,最佳配准评估所有转换的特征高斯分布。Step B2-2, optimize the parameters of the two sonar views, transform the 3D points of the second view into the coordinate system of the first view, and evaluate the characteristic Gaussian distribution of all transformations for the best registration.
进一步地,步骤B3生成三维地图包括以下步骤:Further, generating the three-dimensional map in step B3 includes the following steps:
根据步骤B2优化参数后,计算速度加快得到最优解,计算相对于初始位置的声纳轨 迹,完成精确的三维重建。After the parameters are optimized according to step B2, the calculation speed is accelerated to obtain the optimal solution, and the sonar trajectory relative to the initial position is calculated to complete accurate three-dimensional reconstruction.
本发明达到的有益效果为:实现了对声纳图像从特征提取到涵盖高程信息的环境地图的重建,同时提供一定的运动姿态估计信息,可用于水下机器人声纳图像处理与建图领域。The beneficial effects achieved by the present invention are: realizing feature extraction of sonar images to reconstruction of environmental maps covering elevation information, and providing certain motion attitude estimation information, which can be used in the field of sonar image processing and mapping of underwater robots.
附图说明Description of drawings
图1为本发明实施例中整个声纳图像处理的流程图。Fig. 1 is a flow chart of the whole sonar image processing in the embodiment of the present invention.
图2为本发明实施例中声纳数据图像中像素明显的目标处理区域。Fig. 2 is a target processing area with obvious pixels in the sonar data image in the embodiment of the present invention.
图3为本发明实施例中通过场景分析得到的三个级别像素区。FIG. 3 is a three-level pixel area obtained through scene analysis in an embodiment of the present invention.
图4为本发明实施例中海底声纳图和k-means聚类处理过的图像。Fig. 4 is an image processed by seabed sonar image and k-means clustering in the embodiment of the present invention.
图5为本发明实施例中用高斯分布表示k-means聚类处理目标像素。FIG. 5 shows the Gaussian distribution used to represent the target pixels of k-means clustering processing in the embodiment of the present invention.
图6为本发明实施例中利用仰角信息计算高程图。Fig. 6 is an elevation map calculated by using elevation angle information in an embodiment of the present invention.
图7为本发明实施例中声纳视图1和通过刚性运动得到的视图2。Fig. 7 is a sonar view 1 and a view 2 obtained by rigid motion in an embodiment of the present invention.
具体实施方式Detailed ways
下面结合说明书附图对本发明的技术方案做进一步的详细说明。The technical solution of the present invention will be further described in detail below in conjunction with the accompanying drawings.
本发明的目的是开发一种基于高斯分布聚类的水下声呐图像匹配方法,通过图像配准和优化声纳三维运动参数,通过二维声纳图像进行水下环境三维重建,整体步骤如图1所示,包括以下步骤:The purpose of the present invention is to develop an underwater sonar image matching method based on Gaussian distribution clustering, through image registration and optimization of sonar three-dimensional motion parameters, three-dimensional reconstruction of underwater environment through two-dimensional sonar images, the overall steps are shown in the figure 1, including the following steps:
步骤A:提取特征,建立特征间的匹配关系,获得地图的高度信息;Step A: Extract features, establish matching relationships between features, and obtain map height information;
步骤B:进行位姿估计,将特征地图更新,从而生成三维空间地图。Step B: Perform pose estimation and update the feature map to generate a three-dimensional space map.
步骤A特征检测包括以下步骤:步骤A1采集声纳数据,进行特征检测;步骤A2图像配准,生成高斯地图;步骤A3计算出场景高程地图。The feature detection in step A includes the following steps: step A1 collects sonar data and performs feature detection; step A2 image registration to generate a Gaussian map; step A3 calculates the scene elevation map.
步骤A1中采集声纳数据,进行特征提取包括以下2个步骤:Collecting sonar data in step A1, performing feature extraction includes the following two steps:
步骤A1-1,采集声纳数据,从声纳数据图像中选取像素明显的区域为目标处理区域,如图2所示。Step A1-1, collect sonar data, and select an area with obvious pixels from the sonar data image as the target processing area, as shown in FIG. 2 .
步骤A1-2,利用目标特征区域灰度分布信息并通过场景分析,将目标区域分为三个类别,分辨出目标像素区域和其他非目标区域,根据灰度分布信息,消除由于噪声产生的像素区域,实现特征点的描述。Step A1-2, use the gray level distribution information of the target feature area and analyze the scene to divide the target area into three categories, distinguish the target pixel area and other non-target areas, and eliminate the pixels caused by noise according to the gray level distribution information Region, to realize the description of feature points.
步骤A1-2包括以下2个步骤:Step A1-2 includes the following 2 steps:
步骤A1-2a,利用目标特征区域灰度分布信息并通过场景分析,将声纳图像中相对明显的灰度值分为三种类别,如图3所示:1.组成3D目标或结构的明亮小斑点;2.与3D目标相邻的表面投射的阴影区;3.介于前两个灰度值之间的平坦表面。Step A1-2a, using the gray distribution information of the target feature area and analyzing the scene, the relatively obvious gray values in the sonar image are divided into three categories, as shown in Figure 3: Small spots; 2. Shadow areas cast by surfaces adjacent to the 3D object; 3. Flat surfaces between the first two grayscale values.
步骤A1-2b,通过场景分析和像素强度,将目标像素和其他非目标区域分辨出,去除由于噪声产生的像素区域,减少图像误差。Step A1-2b, through scene analysis and pixel intensity, distinguish target pixels from other non-target areas, remove pixel areas due to noise, and reduce image errors.
步骤A2中图像配准,生成高斯地图包括以下步骤:Image registration in step A2, generating a Gaussian map includes the following steps:
步骤A2-1,通过K-MEANS聚类方法将像素强度较强的目标像素聚类,如图4所示。In step A2-1, the target pixels with strong pixel intensity are clustered by the K-MEANS clustering method, as shown in FIG. 4 .
步骤A2-2,采用低通滤波去除噪声,减少对特征提取的误差。Step A2-2, using low-pass filtering to remove noise and reduce errors in feature extraction.
作为本发明的进一步限定,步骤A2-1包括以下步骤:As a further definition of the present invention, step A2-1 includes the following steps:
步骤A2-1a,对步骤A1-2b得到的目标特征点网格化,基于网格内的点,通过K-means聚类法,将每个网格化目标像素进行网格内聚类,如图5,选取1%-2%具有高亮度的像素,剔除小于8的像素,选取k值使任何子区域不大于32像素,将大于32像素的区域划分为较小的区域,得到合适的特征区域。In step A2-1a, the target feature points obtained in step A1-2b are gridded, and based on the points in the grid, each gridded target pixel is clustered in the grid through the K-means clustering method, such as Figure 5, select 1%-2% of the pixels with high brightness, eliminate the pixels less than 8, select the k value so that any sub-region is not greater than 32 pixels, divide the region greater than 32 pixels into smaller regions, and obtain suitable features area.
步骤A2-1b,用高斯分布表示k-means聚类过的目标特征点,由公式1和公式2计算协方差矩阵的特征值和特征向量反应表面的方向和光滑性等信息,用聚类均值和协方差表示任一聚类,生成高斯地图。均值向量和协方差矩阵对应于每个二维图像区域的位置、大小和方向。Step A2-1b, use Gaussian distribution to represent the target feature points clustered by k-means, calculate the eigenvalues and eigenvectors of the covariance matrix by formula 1 and formula 2 to reflect information such as the direction and smoothness of the surface, and use the cluster mean and covariance represent any clustering, yielding a Gaussian map. The mean vector and covariance matrix correspond to the location, size and orientation of each 2D image region.
Figure PCTCN2022094444-appb-000002
Figure PCTCN2022094444-appb-000002
Figure PCTCN2022094444-appb-000003
Figure PCTCN2022094444-appb-000003
其中,示一个网格内所有扫描点,通过上下标以及取值范围表示区域j内Nj点的集合(共R个区域)中的第i个点。特征区域j用均值μj和方差Σj高斯表示如图3所示不同于正态分布,基于集群的高斯映射(统计均值,协方差)表示,消除了保持一个繁琐的网格结构的必要性,而不失去精度,减少优化过程中的计算时间。Among them, all scanning points in a grid are shown, and the i-th point in the set of Nj points in area j (a total of R areas) is indicated by subscripts, superscripts, and value ranges. Feature region j is represented by Gaussian mean value μj and variance Σj as shown in Figure 3. Different from the normal distribution, the cluster-based Gaussian map (statistical mean, covariance) representation eliminates the need to maintain a cumbersome grid structure, while Reduce computation time during optimization without losing precision.
步骤A3中计算场景高程图包括以下步骤:Calculating the scene elevation map in step A3 includes the following steps:
步骤A3-1,根据平面性假设的方法进行计算,估计场景相对平坦部分和三维物体上的 点的仰角。平面性假设主要是估算场景中相对平坦部分和3-D物体上的点的仰角(基于它们投射的阴影)。Step A3-1, calculate according to the method of planarity assumption, estimate the elevation angle of the scene relative to the flat part and the point on the three-dimensional object. The planarity assumption basically estimates the elevation angles of points on relatively flat parts of the scene and on 3-D objects (based on the shadows they cast).
步骤A3-2,使用最小和最大距离值的图像点来固定平面,即前端点和后缘点的三维坐标是固定的,如图6所示,在不同方位扫描出类似的点对,通过投射阴影确定遮挡轮廓点建立目标物体仰角与阴影的对应关系,每个三维对象从前端到后端的高程值由线性插值填充。Step A3-2, use the image points of the minimum and maximum distance values to fix the plane, that is, the three-dimensional coordinates of the front point and the trailing edge point are fixed, as shown in Figure 6, similar point pairs are scanned in different orientations, and are projected The shadow determines the occlusion contour point to establish the corresponding relationship between the elevation angle of the target object and the shadow, and the elevation value of each 3D object from the front end to the back end is filled by linear interpolation.
步骤A3-3,通过低通滤波对平滑后的数据进行处理,以减小图像噪声对局部峰值的影响;通过k-means分割,建立背景、对象和阴影区域的阈值(在步骤A1-2中建立)定位物体到阴影和阴影到地面的转换。Step A3-3, process the smoothed data by low-pass filtering to reduce the influence of image noise on local peaks; through k-means segmentation, establish thresholds for background, objects and shadow regions (in step A1-2 Create) positioning object to shadow and shadow to ground transformations.
步骤A3-4,利用公式3通过后缘点的仰角和遮挡轮廓的角度估算出三维物体的大小,计算出场景高程图。小物体可能没有明显的阴影,对于这些物体,将物体高度为设为零。Step A3-4, using Formula 3 to estimate the size of the three-dimensional object through the elevation angle of the trailing edge point and the angle of the occlusion contour, and calculate the scene elevation map. Small objects may not have obvious shadows, for these objects, set the object height to zero.
Figure PCTCN2022094444-appb-000004
Figure PCTCN2022094444-appb-000004
式中,Hs:声纳探测到海底的距离,Rs:斜距,Ls:阴影长度,Rh:平距,Ht:目标物高度,Lt:目标物长度。In the formula, Hs: distance detected by sonar to the bottom, Rs: slant distance, Ls: shadow length, Rh: horizontal distance, Ht: target height, Lt: target length.
步骤B包括以下步骤:步骤B1位姿估计;步骤B2参数优化;步骤B3生成三维地图。Step B includes the following steps: step B1 pose estimation; step B2 parameter optimization; step B3 generate a three-dimensional map.
步骤B1位姿估计包括以下步骤:Step B1 pose estimation includes the following steps:
步骤B1-1,在步骤A3-4得到的场景高程图中,选取第一帧声纳图像作为参考图像,第二帧图像作为待匹配图像,每个图像点都可以通过声纳刚性运动后到新的位置,如图7,通过计算帧到帧的运动参数来对两张声纳图像进行配准。Step B1-1, in the scene elevation map obtained in step A3-4, select the first frame of sonar image as the reference image, and the second frame of image as the image to be matched, and each image point can be moved to The new position, as shown in Figure 7, is used to register the two sonar images by calculating frame-to-frame motion parameters.
三维世界投影到二维声纳图像上丢失仰角,可表示为三维场景点P s沿三维点的距离和方位角的零高程(X s,Y s)平面上的映射h(P s),计算公式为: The elevation angle is lost when the 3D world is projected onto the 2D sonar image, which can be expressed as the mapping h(P s ) on the zero-elevation (X s , Y s ) plane of the 3D scene point P s along the distance and azimuth of the 3D point, and calculate The formula is:
Figure PCTCN2022094444-appb-000005
Figure PCTCN2022094444-appb-000005
式中,R为声纳束到反射目标的距离,Φ为方位角。In the formula, R is the distance from the sonar beam to the reflected target, and Φ is the azimuth angle.
其中声纳坐标系中点P的三维坐标系P s=(X s Y s Z s) T可以用(R,θ,Φ) T表示,笛卡尔和球面声纳坐标换算公式: The three-dimensional coordinate system P s = (X s Y s Z s ) T of the point P in the sonar coordinate system can be expressed by (R,θ,Φ) T , and the conversion formula of Cartesian and spherical sonar coordinates is:
Figure PCTCN2022094444-appb-000006
Figure PCTCN2022094444-appb-000006
Figure PCTCN2022094444-appb-000007
Figure PCTCN2022094444-appb-000007
用6个分量描述声纳的三维运动,T=[t x,t y,t z] T和W=[w x,w y,w z] T,二维图像中不包括转动分量[w x,w y,],通过计算帧到帧的运动参数[t x,t y,t z,w z]对两幅图进行配准。 Use 6 components to describe the three-dimensional motion of the sonar, T=[t x , ty ,t z ] T and W=[w x ,w y ,w z ] T , the two-dimensional image does not include the rotation component [w x ,w y ,], register the two images by calculating the frame-to-frame motion parameters [t x ,t y ,t z ,w z ].
Figure PCTCN2022094444-appb-000008
Figure PCTCN2022094444-appb-000008
步骤B1-2,计算声纳运动参数,寻求最适合在转换特性从参考图像到待匹配图像,用空间转换函数进行位姿估计,利用公式7将第一幅图投影到第二幅图S′,对两幅图像的变换即旋转平移等对比.声纳位置运动产生两个场景图(在同一范围R内,不同高度角上)图像配准通过场景特征的仰角改变通过公式8估计两帧声纳的运动参数[t x,t y,t z,w z],求出最小S和S′(公式9),得到最好的变换参数;然后每次将前一次得到的配准图像作为参考,对下一帧声纳图像进行配准,直至整个图像序列所有待匹配声纳图像配准完成,将所有相邻帧的配准统一到相邻的参考系中,减少成对配准造成的积累误差。 Step B1-2, calculate the sonar motion parameters, seek the most suitable transformation characteristics from the reference image to the image to be matched, use the spatial transformation function to estimate the pose, and use the formula 7 to project the first image to the second image S′ , the transformation of the two images, that is, the comparison of rotation and translation. The movement of the sonar position produces two scene images (in the same range R, at different elevation angles). Accept the motion parameters [t x , t y , t z , w z ], find the minimum S and S' (formula 9), and get the best transformation parameters; then use the previous registration image as a reference each time , to register the next frame of sonar images until the registration of all sonar images to be matched in the entire image sequence is completed, and the registration of all adjacent frames is unified into the adjacent reference frame to reduce the error caused by pairwise registration accumulate errors.
Figure PCTCN2022094444-appb-000009
Figure PCTCN2022094444-appb-000009
S′=HS   (9)S'=HS (9)
式中,H是一个包含平移和平面内旋转的变换矩阵。where H is a transformation matrix including translation and in-plane rotation.
步骤B2参数优化包括以下步骤:Step B2 parameter optimization comprises the following steps:
步骤B2-1,由步骤A3-4得到的高程图每个图像点都可以用公式5投影到相应的空间 点P′ s在分量m=(T,W)声纳刚性运动后,新的位置P′三维场景点由公式10和公式11确定。 Step B2-1, each image point of the elevation map obtained by step A3-4 can be projected to the corresponding space point P 's by using formula 5. After the component m=(T, W) sonar rigid motion, the new position The three-dimensional scene point P' is determined by Equation 10 and Equation 11.
Figure PCTCN2022094444-appb-000010
Figure PCTCN2022094444-appb-000010
Figure PCTCN2022094444-appb-000011
Figure PCTCN2022094444-appb-000011
式中,R为旋转矩阵,W为旋转速度向量。In the formula, R is the rotation matrix, and W is the rotation velocity vector.
声纳图像对应的S=h(P s)和S′=h(P′ s)由公式9和公式12确定。 S=h(P s ) and S'=h(P' s ) corresponding to the sonar image are determined by Formula 9 and Formula 12.
Figure PCTCN2022094444-appb-000012
Figure PCTCN2022094444-appb-000012
经过声纳刚性运动后计算出新的位置三维场景点,得到两个声纳视图。After the sonar rigid motion, the new position of the 3D scene point is calculated, and two sonar views are obtained.
步骤B2-2包括以下步骤:Step B2-2 includes the following steps:
对步骤B2-1得到的两个声纳视图由公式13进行参数优化,将第二个视图的三维点变换到第一个视图的坐标系中,最佳配准在评估所有转换的特征高斯分布时产生最大函数值。The parameters of the two sonar views obtained in step B2-1 are optimized by Equation 13, and the 3D points of the second view are transformed into the coordinate system of the first view. The best registration evaluates the characteristic Gaussian distribution of all transformations produces the maximum function value.
Figure PCTCN2022094444-appb-000013
Figure PCTCN2022094444-appb-000013
式中,G j(s)表示第一个视图中的高斯分布(对应第j个特征),G′ j(s)表示第二个视图中高斯分布(对应第j个特征),P si和P′ si表示两个视图特征点集合对应的三维场景,M -1将第二个视图的三维点变换到第一个视图的坐标系中。 In the formula, G j (s) represents the Gaussian distribution in the first view (corresponding to the jth feature), G′ j (s) represents the Gaussian distribution in the second view (corresponding to the jth feature), P si and P′ si represents the 3D scene corresponding to the feature point sets of the two views, and M -1 transforms the 3D points of the second view into the coordinate system of the first view.
步骤B3生成三维地图包括以下步骤:Step B3 generating a three-dimensional map includes the following steps:
根据步骤B2优化参数后,从第一幅图象直接映射到S′的第二幅视图中,利用公式8中的变换矩阵H=(M,R,Φ)计算速度和优化作为关键的情况下,cosΦ≈1,S=h(MP s)和S′=h(M -1P s)由公式14计算。 After optimizing the parameters according to step B2, directly map from the first image to the second view of S′, and use the transformation matrix H=(M, R, Φ) in the formula 8 to calculate the speed and optimize as the key situation , cosΦ≈1, S=h(MP s ) and S′=h(M −1 P s ) are calculated by Equation 14.
Figure PCTCN2022094444-appb-000014
Figure PCTCN2022094444-appb-000014
齐次变换M,根据公式15递归计算计算相对于初始位置的声纳轨迹,完成精确的三维重建。The homogeneous transformation M is recursively calculated according to formula 15 to calculate the sonar trajectory relative to the initial position, and complete accurate three-dimensional reconstruction.
kM 0kM k-1 k-1M 0    (15) k M 0 = k M k-1 k-1 M 0 (15)
式中,k表示帧数(时间指数)。In the formula, k represents the number of frames (time index).
以上所述仅为本发明的较佳实施方式,本发明的保护范围并不以上述实施方式为限,但凡本领域普通技术人员根据本发明所揭示内容所作的等效修饰或变化,皆应纳入权利要求书中记载的保护范围内。The above descriptions are only preferred embodiments of the present invention, and the scope of protection of the present invention is not limited to the above embodiments, but all equivalent modifications or changes made by those of ordinary skill in the art according to the disclosure of the present invention should be included within the scope of protection described in the claims.

Claims (8)

  1. 一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:所述方法包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering, it is characterized in that: described method comprises the following steps:
    步骤A,提取特征,建立特征间的匹配关系,获得地图的高程信息;具体包括以下3个步骤:Step A, extract features, establish the matching relationship between features, and obtain the elevation information of the map; specifically, it includes the following three steps:
    步骤A1,采集声纳数据,进行特征提取;Step A1, collecting sonar data and performing feature extraction;
    步骤A2,图像配准,生成高斯地图;Step A2, image registration, generating a Gaussian map;
    步骤A3,计算出场景高程地图;Step A3, calculating the scene elevation map;
    步骤A3中计算场景高程图包括以下步骤:Calculating the scene elevation map in step A3 includes the following steps:
    步骤A3-1,根据平面性假设的方法,估计场景相对平坦部分和三维物体上的点的仰角;基于投射的阴影估算场景中相对平坦部分和3-D物体上的点的仰角;Step A3-1, according to the method of planarity assumption, estimating the elevation angle of the point on the relatively flat part of the scene and the three-dimensional object; estimating the elevation angle of the point on the relatively flat part and the 3-D object in the scene based on the cast shadow;
    步骤A3-2,使用最小和最大距离值的图像点来固定平面,即前端点和后缘点的三维坐标是固定的,在不同方位扫描出类似的点对,通过投射阴影确定遮挡轮廓点建立目标物体仰角与阴影的对应关系,每个三维对象从前端到后端的高程值由线性插值填充;Step A3-2, use the image points of the minimum and maximum distance values to fix the plane, that is, the three-dimensional coordinates of the front point and the trailing edge point are fixed, similar point pairs are scanned in different directions, and the occlusion contour points are determined by projecting shadows to establish The corresponding relationship between the elevation angle of the target object and the shadow, the elevation value of each 3D object from the front end to the back end is filled by linear interpolation;
    步骤A3-3,通过低通滤波对平滑后的数据进行处理,以减小图像噪声对局部峰值的影响;通过k-means分割,建立背景、对象和阴影区域的阈值,定位物体到阴影和阴影到地面的转换;Step A3-3, process the smoothed data through low-pass filtering to reduce the influence of image noise on local peaks; through k-means segmentation, establish thresholds for background, objects and shadow areas, and locate objects to shadows and shadows conversion to ground;
    步骤A3-4,通过后缘点的仰角和遮挡轮廓的角度估算出三维物体的大小,利用下式计算出场景高程图:In step A3-4, the size of the 3D object is estimated by the elevation angle of the trailing edge point and the angle of the occlusion contour, and the scene elevation map is calculated using the following formula:
    Figure PCTCN2022094444-appb-100001
    Figure PCTCN2022094444-appb-100001
    式中,Hs为声纳探测到海底的距离,Rs为斜距,Ls为阴影长度,Ht为目标物高度,Lt为目标物长度;In the formula, Hs is the distance from the sonar to the seabed, Rs is the slant distance, Ls is the shadow length, Ht is the height of the target, and Lt is the length of the target;
    步骤B,进行位姿估计,将特征地图更新,从而生成三维空间地图;具体包括以下步骤:Step B, perform pose estimation, and update the feature map to generate a three-dimensional space map; specifically, the following steps are included:
    步骤B1,位姿估计;Step B1, pose estimation;
    步骤B2,参数优化;Step B2, parameter optimization;
    步骤B3,生成三维地图,包括以下步骤:Step B3, generating a three-dimensional map, includes the following steps:
    根据步骤B2优化参数后,计算速度加快得到最优解,计算相对于初始位置的声纳轨迹,完成精确的三维重建。After the parameters are optimized according to step B2, the calculation speed is accelerated to obtain the optimal solution, and the sonar trajectory relative to the initial position is calculated to complete accurate three-dimensional reconstruction.
  2. 根据权利要求1所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤A1中采集声纳数据,进行特征提取包括以下2个步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 1, characterized in that: collecting sonar data in step A1, performing feature extraction comprises the following 2 steps:
    步骤A1-1,采集声纳数据,从声纳数据图像中选取像素明显的区域为目标处理区域;Step A1-1, collecting sonar data, selecting an area with obvious pixels from the sonar data image as the target processing area;
    步骤A1-2,利用目标特征区域灰度分布信息并通过场景分析,将目标区域分为三个类别,分辨出目标像素区域和其他非目标区域,根据灰度分布信息,消除由于噪声产生的像素区域,实现特征点的描述。Step A1-2, use the gray level distribution information of the target feature area and analyze the scene to divide the target area into three categories, distinguish the target pixel area and other non-target areas, and eliminate the pixels caused by noise according to the gray level distribution information Region, to realize the description of feature points.
  3. 根据权利要求2所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤A1-2包括以下2个步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 2, characterized in that: step A1-2 comprises the following two steps:
    步骤A1-2a,利用目标特征区域灰度分布信息并通过场景分析,将声纳图像中相对明显的灰度值分为三种类别:组成3D目标或结构的明亮小斑点、与3D目标相邻的表面投射的阴影区、介于前两个灰度值之间的平坦表面;Step A1-2a, using the gray distribution information of the target feature area and through scene analysis, the relatively obvious gray value in the sonar image is divided into three categories: bright small spots that form 3D targets or structures, and adjacent to 3D targets The shaded area cast by the surface of , a flat surface between the first two grayscale values;
    步骤A1-2b,通过场景分析和像素强度,将目标像素区域和其他非目标区域分辨出,去除由于噪声产生的像素区域,减少图像误差。Step A1-2b, through scene analysis and pixel intensity, distinguish the target pixel area from other non-target areas, remove the pixel area due to noise, and reduce the image error.
  4. 根据权利要求1所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤A2中图像配准,生成高斯地图包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 1, it is characterized in that: image registration in step A2, generating Gaussian map comprises the following steps:
    步骤A2-1,通过K-means聚类方法将像素强度较强的目标像素聚类;Step A2-1, clustering target pixels with strong pixel intensity by K-means clustering method;
    步骤A2-2,采用低通滤波去除噪声,减少对特征提取的误差。Step A2-2, using low-pass filtering to remove noise and reduce errors in feature extraction.
  5. 根据权利要求4所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤A2-1包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 4, it is characterized in that: step A2-1 comprises the following steps:
    步骤A2-1a,对得到的目标特征点网格化,基于网格内的点,通过K-means聚类法,将每个网格化目标像素进行网格内聚类;Step A2-1a, grid the obtained target feature points, based on the points in the grid, use the K-means clustering method to cluster each grid target pixel in the grid;
    步骤A2-1b,用高斯分布表示k-means聚类过的目标特征点,计算协方差矩阵的特征值和特征向量反应表面的方向和光滑性信息,用聚类均值和协方差表示任一聚类,生成高斯地图;均值向量和协方差矩阵对应于每个二维图像区域的位置、大小和方向。Step A2-1b, use Gaussian distribution to represent the target feature points clustered by k-means, calculate the eigenvalues and eigenvectors of the covariance matrix to reflect the direction and smoothness information of the surface, and use the cluster mean and covariance to represent any cluster Class that generates a Gaussian map; the mean vector and covariance matrix correspond to the location, size, and orientation of each 2D image region.
  6. 根据权利要求5所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤A2-1a包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 5, it is characterized in that: step A2-1a comprises the following steps:
    选取1%-2%具有高亮度的像素,剔除小于8的像素,使任何子区域不大于32像素,将大于32像素的区域划分为较小的区域,得到合适的特征区域。Select 1%-2% of the pixels with high brightness, eliminate the pixels smaller than 8, make any sub-region not larger than 32 pixels, divide the region larger than 32 pixels into smaller regions, and obtain suitable feature regions.
  7. 根据权利要求1所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤B1位姿估计包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 1, it is characterized in that: step B1 pose estimation comprises the following steps:
    步骤B1-1,在得到的场景高程图中,选取第一帧声纳图像作为参考图像,第二帧图像作为待匹配图像,每个图像点都通过声纳刚性运动后到新的位置,通过计算帧到帧的运动参数来对两张声纳图像进行配准;Step B1-1, in the obtained scene elevation map, select the first frame of sonar image as the reference image, and the second frame of image as the image to be matched. Compute frame-to-frame motion parameters to register two sonar images;
    步骤B1-2,计算声纳运动参数,寻求最适合在转换特性从参考图像到待匹配图像,用空间转换函数进行位姿估计,对两幅图像的变换即旋转平移对比,得到最好的变换参数;然后每次将前一次得到的配准图像作为参考,对下一帧声纳图像进行配准,直至整个图像序列所有待匹配声纳图像配准完成,将所有相邻帧的配准统一到相邻的参考系中,减少成对配准造成的积累误差。Step B1-2, calculate the sonar motion parameters, find the most suitable transformation characteristics from the reference image to the image to be matched, use the space transformation function to estimate the pose, and compare the transformation of the two images, that is, the rotation and translation, to obtain the best transformation parameters; then each time the previous registration image is used as a reference to register the next frame of sonar images until the registration of all sonar images to be matched in the entire image sequence is completed, and the registration of all adjacent frames is unified into adjacent frames of reference to reduce the cumulative error caused by pairwise registration.
  8. 根据权利要求1所述的一种基于高斯分布聚类的水下声呐图像匹配方法,其特征在于:步骤B2参数优化包括以下步骤:A kind of underwater sonar image matching method based on Gaussian distribution clustering according to claim 1, is characterized in that: step B2 parameter optimization comprises the following steps:
    步骤B2-1,由步骤A3-4得到的高程图每个图像点都投影到相应的空间点,经过声纳刚性运动后计算出新的位置三维场景点,得到两个声纳视图;In step B2-1, each image point of the elevation map obtained in step A3-4 is projected to the corresponding space point, and after sonar rigid motion, the new position 3D scene point is calculated to obtain two sonar views;
    步骤B2-2,对两个声纳视图参数优化,将第二个视图的三维点变换到第一个视图的坐标系中,最佳配准评估所有转换的特征高斯分布。Step B2-2, optimize the parameters of the two sonar views, transform the 3D points of the second view into the coordinate system of the first view, and evaluate the characteristic Gaussian distribution of all transformations for the best registration.
PCT/CN2022/094444 2021-05-31 2022-05-23 Underwater sonar image matching method based on gaussian distribution clustering WO2022253027A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202110601626.8 2021-05-31
CN202110601626.8A CN113313172B (en) 2021-05-31 2021-05-31 Underwater sonar image matching method based on Gaussian distribution clustering

Publications (1)

Publication Number Publication Date
WO2022253027A1 true WO2022253027A1 (en) 2022-12-08

Family

ID=77376534

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/094444 WO2022253027A1 (en) 2021-05-31 2022-05-23 Underwater sonar image matching method based on gaussian distribution clustering

Country Status (2)

Country Link
CN (1) CN113313172B (en)
WO (1) WO2022253027A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116012539A (en) * 2023-03-27 2023-04-25 埃尔法(山东)仪器有限公司 Calculation method for three-dimensional imaging of air mass by combining unmanned aerial vehicle with laser detection
CN116342965A (en) * 2023-05-26 2023-06-27 中国电建集团江西省电力设计院有限公司 Water level measurement error analysis and control method and system
CN117824664A (en) * 2024-03-05 2024-04-05 河海大学 Active SLAM method of autonomous unmanned system based on multi-beam sounding sonar
CN118071754A (en) * 2024-04-24 2024-05-24 胜利油田中心医院 Thyroid ultrasound image data processing method

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113313172B (en) * 2021-05-31 2022-03-18 江苏科技大学 Underwater sonar image matching method based on Gaussian distribution clustering
CN113744337B (en) * 2021-09-07 2023-11-24 江苏科技大学 Synchronous positioning and mapping method integrating vision, IMU and sonar
CN114596408A (en) * 2022-02-08 2022-06-07 武汉大学 Micro-parallel three-dimensional reconstruction method based on continuous two-dimensional metal distribution image

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112581610A (en) * 2020-10-16 2021-03-30 武汉理工大学 Robust optimization method and system for establishing map from multi-beam sonar data
CN112802195A (en) * 2020-12-30 2021-05-14 浙江大学 Underwater robot continuous occupying and mapping method based on sonar
CN113313172A (en) * 2021-05-31 2021-08-27 江苏科技大学 Underwater sonar image matching method based on Gaussian distribution clustering

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105405146A (en) * 2015-11-17 2016-03-16 中国海洋大学 Feature density clustering and normal distribution transformation based side-scan sonar registration method
CN109544609A (en) * 2018-10-11 2019-03-29 天津大学 A kind of sidescan-sonar image matching process based on SIFT algorithm
US11892298B2 (en) * 2019-09-17 2024-02-06 FLIR Belgium BVBA Navigational danger identification and feedback systems and methods

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112581610A (en) * 2020-10-16 2021-03-30 武汉理工大学 Robust optimization method and system for establishing map from multi-beam sonar data
CN112802195A (en) * 2020-12-30 2021-05-14 浙江大学 Underwater robot continuous occupying and mapping method based on sonar
CN113313172A (en) * 2021-05-31 2021-08-27 江苏科技大学 Underwater sonar image matching method based on Gaussian distribution clustering

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
M. D. AYKIN ; S. NEGAHDARIPOUR: "On feature extraction and region matching for forward scan sonar imaging", OCEANS 2012, 14 October 2012 (2012-10-14), pages 1 - 9, XP032299951, ISBN: 978-1-4673-0829-8, DOI: 10.1109/OCEANS.2012.6404983 *

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN116012539A (en) * 2023-03-27 2023-04-25 埃尔法(山东)仪器有限公司 Calculation method for three-dimensional imaging of air mass by combining unmanned aerial vehicle with laser detection
CN116012539B (en) * 2023-03-27 2023-06-06 埃尔法(山东)仪器有限公司 Calculation method for three-dimensional imaging of air mass by combining unmanned aerial vehicle with laser detection
CN116342965A (en) * 2023-05-26 2023-06-27 中国电建集团江西省电力设计院有限公司 Water level measurement error analysis and control method and system
CN116342965B (en) * 2023-05-26 2023-11-24 中国电建集团江西省电力设计院有限公司 Water level measurement error analysis and control method and system
CN117824664A (en) * 2024-03-05 2024-04-05 河海大学 Active SLAM method of autonomous unmanned system based on multi-beam sounding sonar
CN117824664B (en) * 2024-03-05 2024-05-28 河海大学 Active SLAM method of autonomous unmanned system based on multi-beam sounding sonar
CN118071754A (en) * 2024-04-24 2024-05-24 胜利油田中心医院 Thyroid ultrasound image data processing method

Also Published As

Publication number Publication date
CN113313172A (en) 2021-08-27
CN113313172B (en) 2022-03-18

Similar Documents

Publication Publication Date Title
WO2022253027A1 (en) Underwater sonar image matching method based on gaussian distribution clustering
CN107833270B (en) Real-time object three-dimensional reconstruction method based on depth camera
JP6681729B2 (en) Method for determining 3D pose of object and 3D location of landmark point of object, and system for determining 3D pose of object and 3D location of landmark of object
RU2713611C2 (en) Three-dimensional space simulation method
US10469828B2 (en) Three-dimensional dense structure from motion with stereo vision
CN112001926B (en) RGBD multi-camera calibration method, system and application based on multi-dimensional semantic mapping
CN111862201B (en) Deep learning-based spatial non-cooperative target relative pose estimation method
CN110866969B (en) Engine blade reconstruction method based on neural network and point cloud registration
CN111899328B (en) Point cloud three-dimensional reconstruction method based on RGB data and generation countermeasure network
CN108597009B (en) Method for detecting three-dimensional target based on direction angle information
JP6349418B2 (en) Object positioning by high-precision monocular movement
CN108225319B (en) Monocular vision rapid relative pose estimation system and method based on target characteristics
Aykin et al. On feature extraction and region matching for forward scan sonar imaging
CN112132876B (en) Initial pose estimation method in 2D-3D image registration
CN113483747A (en) Improved AMCL (advanced metering library) positioning method based on semantic map with corner information and robot
CN109801360B (en) Image-based gastrointestinal three-dimensional reconstruction and visualization method
CN111127613B (en) Image sequence three-dimensional reconstruction method and system based on scanning electron microscope
CN108960267A (en) System and method for model adjustment
Yu et al. A stereovision method for obstacle detection and tracking in non-flat urban environments
CN117422753A (en) High-precision scene real-time three-dimensional reconstruction method combining optics and SAR (synthetic aperture radar) images
CN109064473B (en) 2.5D ultrasonic panoramic image segmentation method
CN112037282B (en) Aircraft attitude estimation method and system based on key points and skeleton
CN116630423A (en) ORB (object oriented analysis) feature-based multi-target binocular positioning method and system for micro robot
CN111709269A (en) Human hand segmentation method and device based on two-dimensional joint information in depth image
Ilg et al. Reconstruction of rigid body models from motion distorted laser range data using optical flow

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22815074

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 22815074

Country of ref document: EP

Kind code of ref document: A1