CN106997591A - A kind of super voxel dividing method of RGB D image mutative scales - Google Patents

A kind of super voxel dividing method of RGB D image mutative scales Download PDF

Info

Publication number
CN106997591A
CN106997591A CN201710168730.6A CN201710168730A CN106997591A CN 106997591 A CN106997591 A CN 106997591A CN 201710168730 A CN201710168730 A CN 201710168730A CN 106997591 A CN106997591 A CN 106997591A
Authority
CN
China
Prior art keywords
point
voxel
distance
points
rgb
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710168730.6A
Other languages
Chinese (zh)
Inventor
袁夏
徐鹏
周宏扬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Science and Technology
Original Assignee
Nanjing University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Science and Technology filed Critical Nanjing University of Science and Technology
Priority to CN201710168730.6A priority Critical patent/CN106997591A/en
Publication of CN106997591A publication Critical patent/CN106997591A/en
Pending legal-status Critical Current

Links

Landscapes

  • Image Analysis (AREA)

Abstract

The invention discloses the super voxel dividing method of mutative scale in one towards RGB D view data, seed point is selected in data using the sampling of Poisson dish, then cluster is iterated according to the color distance and space length between data point and each seed point, obtain initial super voxel segmentation result, the super voxel obtained using initial segmentation is summit, the syntople of super voxel is that non-directed graph is set up on side, super voxel is carried out using the method based on graph theory to merge, the super voxel segmentation result that scale size differs is obtained in same RGB D images, the super voxel segmentation of mutative scale is realized.The inventive method belongs to data prediction, splits obtained super voxel and obtains the super voxel of large scale in the high region of data consistency, obtains the super voxel of small yardstick in the place of data consistency difference, more meets human vision cognitive features.

Description

RGB-D image variable-scale hyper-voxel segmentation method
Technical Field
The invention relates to a superpixel segmentation method, in particular to a variable-scale superpixel segmentation method suitable for RGB-D images.
Background
With the progress of sensor technology, the acquisition cost of RGB-D images becomes lower and lower, and how to more effectively preprocess RGB-D images is an important research content of computer vision in recent years. In order to fully utilize the three-dimensional geometric information in the RGB-D image, similar to the concept of two-dimensional image super-pixel over-segmentation, the RGB-D image is over-segmented into super-voxels, so that an effective preprocessing mode is provided, and the data volume processed by a subsequent algorithm can be effectively reduced.
The currently common voxel size of the voxel division algorithm is high in consistency of the sizes of the voxels after the scale parameters are determined, if multi-scale analysis is needed subsequently, the sizes of the voxels are controlled by setting different scale factors, so that a voxel division result of one scale is obtained by calculation when each scale is analyzed, and the calculation amount is increased. Based on the thought of variable-scale analysis, when the superpixel segmentation is carried out, the region with high data consistency is segmented into the superpixels with large scale, the region with poor data consistency is segmented into the superpixels with small scale, and therefore the local distribution of the data is realizedThe RGB-D data are adaptively segmented into the superpixels with different scales, so that a subsequent algorithm can perform variable-scale calculation on over-segmented data, and the calculation amount is reduced.
Disclosure of Invention
The invention aims to provide a variable-scale superpixel segmentation method for an RGB-D image, which can better perform segmentation pretreatment on the RGB-D image.
The technical solution for realizing the purpose of the invention is as follows: a variable-scale hyper-voxel segmentation method for an RGB-D image comprises the following steps:
step 1, selecting seed points, setting a frame of RGB-D image data as P, setting the resolution as m rows and n columns, wherein each data point comprises 3 color channels (r, g and b) and 1 depth channel (D). Randomly selecting a point P from P0Setting a radius threshold value R as the minimum distance between seed points to be generated as initial seed points, and sampling in P by using a Poisson disc sampling algorithm to obtain a seed point set; the seed point selection specifically comprises the following steps:
step 1-1, randomly selecting a point P in a frame of RGB-D image data P0As an initial seed point, an active sampling point queue L1Initialized to null, p0Adding L1Queue the inactive sampling points L2Initialization is null;
step 1-2, judging an active sampling point queue L1Whether or not it is empty, if L is1If not, then from L1A point p of intermediate dequeueiWith piRandomly selecting candidate sampling points in a concentric circle region with the radius of R and 2R as the circle center, and adding the candidate sampling points into L if the distance between the candidate sampling points and the existing seed points is greater than R1(ii) a If K times of trying still no qualified candidate sampling point, p is addediFrom L1Is deleted and L is added2(ii) a Wherein R adopts a pixel coordinate unit, and K is a preset numerical value;
step 1-3, L2The point in (1) is the selected seed point.
Step 2, voxel pre-segmentation, namely converting the color space of each point in the data from RGB to Lab, converting the depth value d to a three-dimensional space coordinate (x, y, z), taking each seed point as a clustering center, iteratively calculating the distance between a non-seed point and a seed point by integrating the color distance and the three-dimensional space distance, and clustering the closest point to the seed point into a class, thereby obtaining an initial voxel segmentation result; the pre-segmentation of the hyper-voxels specifically comprises the following steps:
step 2-1, converting the depth value of the point in the RGB-D image into a three-dimensional space coordinate, and setting the point of the ith row and the jth column in P as PijDepth value of dijThe depth values are converted to three-dimensional space coordinates using equation (1):
where f is the focal length of the camera and (c)x,cy) For the center coordinates of the image, the RGB colors of each point in P are converted to Lab colors using a standard color space conversion formula, and the points in P are all represented as 6-dimensional vectors [ l, a, b, x, y, z];
Step 2-2, taking the seed points obtained in the step 1 as initial clustering centers, carrying out area search clustering, calculating the distance between each point in the neighborhood range of 2R x 2R of each clustering center and the clustering center, classifying each non-clustering center point into the clustering center with the minimum characteristic distance to complete the first clustering process, wherein the two points P in P are used as the clustering centersiAnd pjThe feature distance metric between the following equations:
in the formula (2), dlabIs the color distance, dxyzλ is a weight for determining color information and spatial distance information, and the spatial distance is more important as λ is larger, and the color distance is more important as λ is smaller;
step 2-3, an iterative clustering process, wherein the clustering center of each class is recalculated according to the result of the first clustering in the step 2-2; the new clustering center characteristic value is the average value of all the point characteristics of each class, then the point closest to the new clustering center characteristic value is searched in one class to be used as a new clustering center point, and the class of each non-clustering center point is recalculated according to the characteristic distance calculation and classification method in 2-2; ending the iteration for k times; after the iterative computation is finished, the points in each class form an initial hyper-voxel.
And 3, carrying out hyper-voxel fusion, namely taking the initial hyper-voxels as vertexes, establishing an edge structure undirected graph G (V, E) between adjacent hyper-voxels, measuring the difference between the adjacent hyper-voxels by using the Lab color space distance and the normal vector direction included angle of each initial hyper-voxel, and fusing the initial hyper-voxels by using the intra-class difference minimization and inter-class difference maximization ideas to obtain the scale-variable hyper-voxels. The hyper-voxel fusion specifically comprises the following steps:
step 3-1, setting the obtained super voxel set of the initial segmentation as C, and each super voxel as CiTo super voxel ciIs a vertex viEstablishing a set of vertices V, neighboring superpixels V with a common boundaryi,vjBetween establishes the edge eijForming an edge set E, and constructing an undirected graph G by using V and E; calculating the normal vector of each point by using a standard k neighbor point coordinate covariance matrix principal component decomposition method according to the three-dimensional space coordinate of the point P, and then assigning a weight w (v) to each edge of G by using an equation (5)i,vj)
Wherein,lmaxare respectively hyper-voxel ciAverage and maximum luminance of all points in, thetaijIs a hyper-voxel ciAnd cjThe normal quantity of the hyper-voxel is taken as the normal vector mean value of all points in the hyper-voxel, and α is a weight factor;
step 3-2, merging the initial superpixels, and initializing each superpixel before merging into a region Ai,vkAnd vlIs a region AiDefine the internal dissimilarity as the maximum edge of the minimum spanning tree of the region, as shown in equation (7)
Int(Ai)=maxw(vk,vl),vk,vl∈Ai,(vk,vl)∈E (7)
Two regions A are defined by the formula (8)iAnd AjOf (a) external dissimilarity of, wherein vmIs a region AiOf (5) supra voxel vnIs a region AjSuper voxel in (1)
Dif(Ai,Aj)=minw(vm,vn),vm∈Ai,vn∈Aj,(vm,vn)∈E (8)
MInt(Ai,Aj) Is AiAnd AjThe minimum internal dissimilarity of these two regions is then calculated using equation (9)
MInt(Ai,Aj)=min((Int(Ai)+τ(Ai)),(Int(Aj+τ(Aj))) (9)
Wherein τ (A)i)=e/|Ai|,|AiI is the region AiThe number of the included points, e is a set constant;
then comparing the external dissimilarity and the internal dissimilarity of the two regions, merging the two regions if the formula (10) is satisfied, otherwise not merging
MInt(Ai,Aj)<Dif(Ai,Aj) (10)
The region merging process is carried out until the regions in P can not be merged, and the variable-scale hyper-voxel is obtained.
Compared with the prior art, the invention has the following remarkable advantages: (1) the initial seed points are selected by adopting a Poisson disc sampling method, and compared with an even grid sampling method, the method is more in line with the distribution characteristics of cone cells on the retina and is beneficial to even sampling in an uneven three-dimensional data point set; (2) the method integrates geometric information such as depth values, normal vectors and the like on the basis of colors, and is more suitable for processing RGB-D type data; (3) the method of the invention obtains the superpixels with different scales by segmenting the same frame of RGB-D data through one-time calculation, is obviously different from the over-segmentation method which can only obtain the superpixels with more consistent scales on the same frame of data through one-time calculation, and is more beneficial to reducing the calculated amount of the subsequent algorithm and multi-scale analysis.
The invention is further described below with reference to the accompanying drawings.
Drawings
FIG. 1 shows the results of the variable scale voxel segmentation after the algorithm pre-segmentation and fusion.
FIG. 2 is a flow chart of the RGB-D image scale-variable hyper-voxel segmentation method of the present invention.
Detailed Description
The invention discloses a variable-scale superpixel segmentation method of an RGB-D image, which comprises the following steps of:
step 1, selecting seed points, setting a frame of RGB-D image data as P, setting the resolution as m rows and n columns, wherein the information of each data point comprises 3 color channels (r, g and b) and 1 depth channel (D). Randomly selecting a point P from P0And as an initial seed point, setting a radius threshold value R as the minimum distance between seed points to be generated, and sampling in P by using a Poisson disc sampling algorithm to obtain a seed point set. The method specifically comprises the following steps:
step 1-1, randomly selecting a point P in P0As an initial seed point, an active sampling point queue L1Initialized to null, p0Adding L1Queue the inactive sampling points L2Initialization is null;
step 1-2, if L1If not, then from L1A point p of intermediate dequeueiWith piTaking R and 2R (R adopts pixel coordinate unit) as the circle center, randomly selecting candidate sampling points in a concentric circle region with the radius of R and 2R respectively, and adding the candidate sampling points into L if the distance between the candidate sampling points and the existing seed points is more than R1. If K times of trying still no qualified candidate sampling point, p is addediFrom L1Is deleted and L is added2
Step 1-3, L2The point in (1) is the selected seed point.
And 2, performing presorting on the voxels, converting the color space of each point in the P into Lab from RGB by using a standard color space conversion method, converting the depth value d into a three-dimensional space coordinate (x, y, z), and iteratively calculating the distance between a non-seed point and a seed point by taking each seed point as a clustering center and integrating the color distance and the three-dimensional space distance to obtain an initial voxel presorting result. The method specifically comprises the following steps:
step 2-1, converting the depth value of the point in the RGB-D image into a three-dimensional space coordinate, and setting the point of the ith row and the jth column in P as PijDepth value of dijConverting the depth value to three using equation (1)Dimensional space coordinates:
where f is the focal length of the camera and (c)x,cy) For the center coordinates of the image, the RGB color of each point in P is converted to Lab color using a standard color space conversion formula. By color space and coordinate space transformation, the points in P are both represented as 6-dimensional vectors [ l, a, b, x, y, z];
Step 2-2, taking the various sub-points obtained in the step 1 as initial clustering centers, carrying out area search clustering, calculating the distance between the point in the neighborhood range of 2R x 2R of each clustering center and the clustering center, classifying each non-clustering center point into the clustering center with the minimum characteristic distance to complete the first clustering process, wherein the two points P in P are used as the clustering centersiAnd pjThe feature distance metric between the following equations:
in the formula (2), dlabIs the color distance, dxyzFor the spatial distance, λ is a weight for determining the color information and the spatial distance information, the spatial distance is more important when λ is larger, and the color distance is more important when λ is smaller, specifically set according to the specific application requirements.
And 2-3, performing iterative clustering process, and recalculating the clustering center of each class according to the result of the first clustering in the step 2-2. And the new clustering center characteristic value is the average value of the characteristics of all the points in the class, then the point closest to the new clustering center characteristic value is searched in the class as a new clustering center point, and the class of each non-clustering center point is recalculated according to the characteristic distance calculation and classification method in 2-2. And ending the iteration k times. After the iterative computation is finished, the points in each class form an initial hyper-voxel.
And 3, fusing the hyper-voxels, namely constructing an undirected graph G (V, E) for the edges by taking the initial hyper-voxels as vertexes and the adjacency relation among the hyper-voxels, measuring the difference between the adjacent hyper-voxels by using the Lab color space distance and the normal vector direction included angle of each initial hyper-voxel, and fusing the initial hyper-voxels by using the intra-class difference minimization and the inter-class difference maximization concepts to obtain the scale-variable hyper-voxels. The method specifically comprises the following steps:
step 3-1, setting the obtained super voxel set of the initial segmentation as C, and each super voxel as CiTo super voxel ciIs a vertex viEstablishing a set of vertices V, neighboring superpixels V with a common boundaryi,vjBetween establishes the edge eijAnd forming an edge set E, and constructing an undirected graph G by using V and E. Calculating the normal vector of each point by using a standard k neighbor point coordinate covariance matrix principal component decomposition method according to the three-dimensional space coordinate of the point P, and then assigning a weight w (v) to each edge of G by using an equation (5)i,vj)
Wherein,lmaxare respectively hyper-voxel ciAverage and maximum luminance of all points in, thetaijIs a hyper-voxel ciAnd cjAngle of normal vector, of voxels of hyper-voxelsThe normal quantity is taken as the normal vector mean of all points in the hyper-voxel, and α is a weight factor.
Step 3-2, the initial hyper-voxels are merged, and each hyper-voxel is initialized to a region A before mergingi,vkAnd vlIs a region AiDefine the internal dissimilarity as the maximum edge of the minimum spanning tree of the region, as shown in equation (7)
Int(Ai)=maxw(vk,vl),vk,vl∈Ai,(vk,vl)∈E (7)
The external dissimilarity of the two regions is defined by equation (8)
Dif(Ai,Aj)=minw(vm,vn),vm∈Ai,vn∈Aj,(vm,vn)∈E (8)
MINt (Ai, Aj) is AiAnd AjThe minimum internal difference between these two regions is then calculated using equation (9)
MInt(Ai,Aj)=min((Int(Ai)+τ(Ai)),(Int(Aj+τ(Aj))) (9)
Wherein τ (A)i)=e/|Ai|,|AiI is the region AiThe number of the included points, e is a set constant;
then comparing the external dissimilarity and the internal dissimilarity of the two regions, merging the two regions if the formula (10) is satisfied, otherwise not merging
MInt(Ai,Aj)<Dif(Ai,Aj) (10)
The region merging process is carried out until the regions in P can not be merged, and the variable-scale hyper-voxel is obtained.
The initial seed points are selected by adopting a Poisson disc sampling method, and compared with a uniform grid sampling method, the method is more consistent with the distribution characteristic of cone cells on the retina and is beneficial to uniform sampling in a non-uniform three-dimensional data point set.
The present invention is further illustrated by the following specific examples.
Example 1
FIG. 1 is a schematic diagram of the process of pre-segmentation and scale-variable segmentation of an RGB-D image collected by a frame of Kinect depth camera and the result thereof.
The method for segmenting the RGB-D image of one frame by using the variable-scale hyper-voxel over-segmentation method comprises the following steps:
step 1, selecting seed points, wherein one frame of RGB-D image data is P, the resolution is 480 lines and 640 columns, and the information of each data point comprises 3 color channels (r, g and b) and 1 depth channel (D). Randomly selecting a point P from P0And setting a radius threshold value R to be 20 as the minimum distance between the seed points to be generated as the initial seed points, and sampling in P by using a Poisson disc sampling algorithm to obtain a seed point set. The method specifically comprises the following steps:
step 1-1, randomly selecting a point P in P0As an initial seed point, an active sampling point queue L1Initialized to null, p0Adding L1Queue the inactive sampling points L2Initialization is null;
step 1-2, if L1If not, then from L1A point p of intermediate dequeueiWith piRandomly selecting candidate sampling points in a concentric circle region with the radius of R and 2R as the circle center, and adding the candidate sampling points into L if the distance between the candidate sampling points and the existing seed points is greater than R1. If no qualified candidate sample point exists after K is 30 times of trials, p is addediFrom L1Is deleted and L is added2
Steps 1-3,L2The point in (1) is the selected seed point
And 2, performing voxel pre-segmentation, namely converting the color space of each point in the P into Lab from RGB by using a standard color space conversion formula, converting the depth value d into a three-dimensional space coordinate (x, y, z), and iteratively calculating the distance between a non-seed point and a seed point by taking each seed point as a clustering center and integrating the color distance and the three-dimensional space distance to obtain an initial voxel segmentation result. The method specifically comprises the following steps:
step 2-1, converting the depth value of the point in the RGB-D image into a three-dimensional space coordinate, and setting the point of the ith row and the jth column in P as PijDepth value of dijThe depth values are converted to three-dimensional space coordinates using equation (1):
according to the Kinect depth camera's development manual, f-570.3, cx=320,cy240. The RGB colors of each point in P are converted to Lab colors using standard color space conversion formulas. By color space and coordinate space transformation, the points in P are both represented as 6-dimensional vectors [ l, a, b, x, y, z];
Step 2-2, taking the seed points obtained in the step 1 as initial clustering centers, carrying out area search clustering, calculating the distance between each point in the neighborhood range of 2R x 2R of each clustering center and the clustering center, classifying each non-clustering center point into the clustering center with the minimum characteristic distance to complete a first clustering process, wherein two points P in P are used as the clustering centersiAnd pjThe feature distance metric between the following equations:
in the formula (2), dlabIs the color distance, dxyzλ is 1 as the spatial distance.
And 2-3, performing iterative clustering process, and recalculating the clustering center of each class according to the result of the first clustering in the step 2-2. And the new clustering center characteristic value is the average value of all the point characteristics of each class, then the point closest to the new clustering center characteristic value is searched in one class to be used as a new clustering center point, and the class of each non-clustering center point is recalculated according to the characteristic distance calculation and classification method in 2-2. The iteration k ends 10 times. After the iterative computation is completed, the points in each class form an initial hyper-voxel, as shown by the pre-segmentation result in fig. 1.
And 3, fusing the hyper-voxels, namely constructing an undirected graph G (V, E) for the edges by taking the initial hyper-voxels as vertexes and the adjacency relation among the hyper-voxels, measuring the difference between the adjacent hyper-voxels by using the Lab color space distance and the normal vector direction included angle of each initial hyper-voxel, and fusing the initial hyper-voxels by using the intra-class difference minimization and the inter-class difference maximization concepts to obtain the scale-variable hyper-voxels. The method specifically comprises the following steps:
step 3-1, setting the obtained super voxel set of the initial segmentation as C, and each super voxel as CiTo super voxel ciIs a vertex viEstablishing a set of vertices V, neighboring superpixels V with a common boundaryi,vjBetween establishes the edge eijAnd forming an edge set E, and constructing an undirected graph G by using V and E. Calculating the normal vector of each point by using a standard k neighbor point coordinate covariance matrix principal component decomposition method according to the three-dimensional space coordinate of the point P, and then assigning a weight w (v) to each edge of G by using an equation (5)i,vj)
Wherein,lmaxare respectively hyper-voxel ciAverage and maximum luminance of all points in, thetaijIs a hyper-voxel ciAnd cjThe normal quantity of the hyper-voxel is the normal vector mean of all points in the hyper-voxel, α is 0.5.
Step 3-2, initial hyper-voxel combination, wherein each hyper-voxel is initialized to a region A before combinationi,vkAnd vlIs a region AiDefine the internal dissimilarity as the maximum edge of the minimum spanning tree of the region, as shown in equation (7)
Int(Ai)=maxw(vk,vl),vk,vl∈Ai,(vk,vl)∈E (7)
Two regions A are defined by the formula (8)iAnd AjOf (a) external dissimilarity of, wherein vmIs a region AiOf (5) supra voxel vnIs a region AjSuper voxel in (1)
Dif(Ai,Aj)=minw(vm,vn),vm∈Ai,vn∈Aj,(vm,vn)∈E (8)
MInt(Ai,Aj) Is AiAnd AjThe minimum internal dissimilarity of these two regions is then calculated using equation (9)
MInt(Ai,Aj)=min((Int(Ai)+τ(Ai)),(Int(Aj+τ(Aj))) (9)
Wherein τ (A)i)=e/|Ai|,|AiI is the region AiThe number of dots included, e equals 100.
Then comparing the external dissimilarity and the internal dissimilarity of the two regions, merging the two regions if the formula (10) is satisfied, otherwise not merging
MInt(Ai,Aj)<Dif(Ai,Aj) (10)
The region merging process is performed until the regions in P can not be merged, and each finally obtained region after merging is a variable-scale superpixel obtained by segmentation, as shown in the variable-scale superpixel segmentation result in fig. 1.
The method of the invention obtains the superpixels with different scales by segmenting the same frame of RGB-D data through one-time calculation, is obviously different from the over-segmentation method which can only obtain the superpixels with more consistent scales on the same frame of data through one-time calculation, and is more beneficial to reducing the calculated amount of the subsequent algorithm and multi-scale analysis.

Claims (4)

1. A variable-scale hyper-voxel segmentation method for an RGB-D image is characterized by comprising the following steps of:
step 1, selecting seed points, setting a frame of RGB-D image data as P, setting the resolution as m rows and n columns, wherein the information of each data point comprises 3 color channels (r, g and b) and 1 depth channel (D); randomly selecting a point P from P0(the number 0 is a subscript) is used as an initial seed point, a radius threshold value R is set as the minimum distance between seed points to be generated, and a Poisson disc sampling algorithm is used for sampling in the P to obtain a seed point set;
step 2, pre-dividing the voxel, converting the color space of each point in the data from RGB to Lab, converting the depth value d to a three-dimensional space coordinate (x, y, z), then taking each seed point as a clustering center, and iteratively calculating the distance between a non-seed point and the seed point by integrating the color distance and the three-dimensional space distance to obtain an initial voxel division result;
and 3, fusing the superpixels, constructing an undirected graph G (V, E) for the edges by using the initial superpixels as vertexes and the adjacency relation among the superpixels as edges, measuring the difference between the adjacent superpixels by using the Lab color space distance and the normal vector direction included angle of each initial superpixel, obtaining the variable-scale superpixels by combining the initial superpixels, and completing the segmentation of the variable-scale superpixels of the RGB-D image.
2. The RGB-D image scale-variable hyper-voxel segmentation method according to claim 1, wherein the seed point selection in step 1 specifically comprises the following steps:
step 1-1, randomly selecting a point P in a frame of RGB-D image data P0As an initial seed point, an active sampling point queue L1Initialized to null, p0Adding L1Queue the inactive sampling points L2Initialization is null;
step 1-2, judging an active sampling point queue L1Whether or not it is empty, if L is1If not, then from L1A point p of intermediate dequeueiWith piRandomly selecting candidate sampling points in a concentric circle region with the radius of R and 2R as the circle center, and adding the candidate sampling points into L if the distance between the candidate sampling points and the existing seed points is greater than R1(ii) a If K times of trying still no qualified candidate sampling point, p is addediFrom L1Is deleted and L is added2(ii) a Wherein R adopts a pixel coordinate unit, and K is a preset numerical value;
step 1-3, L2The point in (1) is the selected seed point.
3. The RGB-D image scale-variable voxel segmentation method according to claim 1 or 2, wherein the pre-segmentation of the voxels in step 2 comprises the following steps:
step 2-1, converting the depth value of the point in the RGB-D image into a three-dimensional space coordinate, and setting the point of the ith row and the jth column in P as PijDepth value of dijThe depth values are converted to three-dimensional space coordinates using equation (1):
p i j ( x , y , z ) T = d i j ( i - c x f , j - c y f , 1 ) T - - - ( 1 )
where f is the focal length of the camera and (c)x,cy) For the center coordinates of the image, the RGB colors of each point in P are converted to Lab colors using a standard color space conversion formula, and the points in P are all represented as 6-dimensional vectors [ l, a, b, x, y, z];
Step 2-2, taking the seed points obtained in the step 1 as initial clustering centers, carrying out area search clustering, calculating the distance between each point in the neighborhood range of 2R x 2R of each clustering center and the clustering center, and classifying each non-clustering center point intoAnd the clustering center with the minimum characteristic distance to complete the first clustering process, two points P in PiAnd pjThe feature distance metric between the following equations:
D ( p i , p j ) = d l a b + &lambda; d x y z R - - - ( 2 )
d l a b = ( l i - l j ) 2 + ( a i - a j ) 2 + ( b i - b j ) 2 - - - ( 3 )
d x y z = ( x i - x j ) 2 + ( y i - y j ) 2 + ( z i - z j ) 2 - - - ( 4 )
in the formula (2), dlabIs the color distance, dxyzIs the spatial distance, λ is the weight used to determine the color information and the spatial distance information, the larger λ the more important the spatial distance,the smaller λ is, the more important the color distance is;
step 2-3, an iterative clustering process, wherein the clustering center of each class is recalculated according to the result of the first clustering in the step 2-2; the new clustering center characteristic value is the average value of all the point characteristics of each class, then the point closest to the new clustering center characteristic value is searched in one class to be used as a new clustering center point, and the class of each non-clustering center point is recalculated according to the characteristic distance calculation and classification method in 2-2; ending the iteration for k times; after the iterative computation is finished, the points in each class form an initial hyper-voxel.
4. The method for estimating the normal vector of the cloud point of the three-dimensional point according to claim 1, wherein the hyper-voxel fusion in the step 3 specifically comprises the following steps:
step 3-1, setting the obtained super voxel set of the initial segmentation as C, and each super voxel as CiTo super voxel ciIs a vertex viEstablishing a set of vertices V, neighboring superpixels V with a common boundaryi,vjBetween establishes the edge eijForming an edge set E, and constructing an undirected graph G by using V and E; calculating the normal vector of each point by using a standard k neighbor point coordinate covariance matrix principal component decomposition method according to the three-dimensional space coordinate of the point P, and then assigning a weight w (v) to each edge of G by using an equation (5)i,vj)
w ( v i , v j ) = &alpha; | l &OverBar; i - l &OverBar; j | l m a x + ( 1 - &alpha; ) &theta; i j &pi; - - - ( 5 )
&theta; i j = cos - 1 ( n &RightArrow; i &CenterDot; n &RightArrow; j | n &RightArrow; i | | n &RightArrow; j | ) - - - ( 6 )
Wherein,lmaxare respectively hyper-voxel ciAverage and maximum luminance of all points in, thetaijIs a hyper-voxel ciAnd cjThe normal quantity of the hyper-voxel is taken as the normal vector mean value of all points in the hyper-voxel, and α is a weight factor;
step 3-2, merging the initial superpixels, and initializing each superpixel before merging into a region Ai,vkAnd vlIs a region AiDefine the internal dissimilarity as the maximum edge of the minimum spanning tree of the region, as shown in equation (7)
Int(Ai)=max w(vk,vl),vk,vl∈Ai,(vk,vl)∈E (7)
Two regions A are defined by the formula (8)iAnd AjOf (a) external dissimilarity of, wherein vmIs a region AiOf (5) supra voxel vnIs a region AjSuper voxel in (1)
Dif(Ai,Aj)=min w(vm,vn),vm∈Ai,vn∈Aj,(vm,vn)∈E (8)
MInt(Ai,Aj) Is AiAnd AjThe minimum internal dissimilarity of these two regions is then calculated using equation (9)
MInt(Ai,Aj)=min((Int(Ai)+τ(Ai)),(Int(Aj+τ(Aj))) (9)
Wherein τ (A)i)=e/|Ai|,|AiI is the region AiThe number of the included points, e is a set constant;
then comparing the external dissimilarity and the internal dissimilarity of the two regions, merging the two regions if the formula (10) is satisfied, otherwise not merging
MInt(Ai,Aj)<Dif(Ai,Aj) (10)
The region merging process is carried out until the regions in P can not be merged, and the variable-scale hyper-voxel is obtained.
CN201710168730.6A 2017-03-21 2017-03-21 A kind of super voxel dividing method of RGB D image mutative scales Pending CN106997591A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710168730.6A CN106997591A (en) 2017-03-21 2017-03-21 A kind of super voxel dividing method of RGB D image mutative scales

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710168730.6A CN106997591A (en) 2017-03-21 2017-03-21 A kind of super voxel dividing method of RGB D image mutative scales

Publications (1)

Publication Number Publication Date
CN106997591A true CN106997591A (en) 2017-08-01

Family

ID=59431712

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710168730.6A Pending CN106997591A (en) 2017-03-21 2017-03-21 A kind of super voxel dividing method of RGB D image mutative scales

Country Status (1)

Country Link
CN (1) CN106997591A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146894A (en) * 2018-08-07 2019-01-04 庄朝尹 A kind of model area dividing method of three-dimensional modeling
CN109257543A (en) * 2018-11-29 2019-01-22 维沃移动通信有限公司 Screening-mode control method and mobile terminal
CN109917419A (en) * 2019-04-12 2019-06-21 中山大学 A kind of depth fill-in congestion system and method based on laser radar and image
CN110610505A (en) * 2019-09-25 2019-12-24 中科新松有限公司 Image segmentation method fusing depth and color information
CN110852949A (en) * 2019-11-07 2020-02-28 上海眼控科技股份有限公司 Point cloud data completion method and device, computer equipment and storage medium
CN111488769A (en) * 2019-01-28 2020-08-04 北京工商大学 Unsupervised fusion point cloud superpixelization method based on light spot divergence size
CN112348816A (en) * 2021-01-07 2021-02-09 北京明略软件系统有限公司 Brain magnetic resonance image segmentation method, storage medium, and electronic device
CN112580641A (en) * 2020-11-23 2021-03-30 上海明略人工智能(集团)有限公司 Image feature extraction method and device, storage medium and electronic equipment

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101877128A (en) * 2009-12-23 2010-11-03 中国科学院自动化研究所 Method for segmenting different objects in three-dimensional scene
CN104616289A (en) * 2014-12-19 2015-05-13 西安华海盈泰医疗信息技术有限公司 Removal method and system for bone tissue in 3D CT (Three Dimensional Computed Tomography) image
CN105719295A (en) * 2016-01-21 2016-06-29 浙江大学 Intracranial hemorrhage area segmentation method based on three-dimensional super voxel and system thereof

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101877128A (en) * 2009-12-23 2010-11-03 中国科学院自动化研究所 Method for segmenting different objects in three-dimensional scene
CN104616289A (en) * 2014-12-19 2015-05-13 西安华海盈泰医疗信息技术有限公司 Removal method and system for bone tissue in 3D CT (Three Dimensional Computed Tomography) image
CN105719295A (en) * 2016-01-21 2016-06-29 浙江大学 Intracranial hemorrhage area segmentation method based on three-dimensional super voxel and system thereof

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
PENGXU等: "Scale adaptive supervoxel segmentation of RGB-D image", 《2016 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND BIOMIMETICS (ROBIO)》 *

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109146894A (en) * 2018-08-07 2019-01-04 庄朝尹 A kind of model area dividing method of three-dimensional modeling
CN109257543A (en) * 2018-11-29 2019-01-22 维沃移动通信有限公司 Screening-mode control method and mobile terminal
CN111488769A (en) * 2019-01-28 2020-08-04 北京工商大学 Unsupervised fusion point cloud superpixelization method based on light spot divergence size
CN109917419A (en) * 2019-04-12 2019-06-21 中山大学 A kind of depth fill-in congestion system and method based on laser radar and image
CN109917419B (en) * 2019-04-12 2021-04-13 中山大学 Depth filling dense system and method based on laser radar and image
CN110610505A (en) * 2019-09-25 2019-12-24 中科新松有限公司 Image segmentation method fusing depth and color information
CN110852949A (en) * 2019-11-07 2020-02-28 上海眼控科技股份有限公司 Point cloud data completion method and device, computer equipment and storage medium
CN112580641A (en) * 2020-11-23 2021-03-30 上海明略人工智能(集团)有限公司 Image feature extraction method and device, storage medium and electronic equipment
CN112580641B (en) * 2020-11-23 2024-06-04 上海明略人工智能(集团)有限公司 Image feature extraction method and device, storage medium and electronic equipment
CN112348816A (en) * 2021-01-07 2021-02-09 北京明略软件系统有限公司 Brain magnetic resonance image segmentation method, storage medium, and electronic device

Similar Documents

Publication Publication Date Title
CN106997591A (en) A kind of super voxel dividing method of RGB D image mutative scales
CN108573276B (en) Change detection method based on high-resolution remote sensing image
CN104331699B (en) A kind of method that three-dimensional point cloud planarization fast search compares
CN111462120B (en) Defect detection method, device, medium and equipment based on semantic segmentation model
CN103258203B (en) The center line of road extraction method of remote sensing image
CN102708370B (en) Method and device for extracting multi-view angle image foreground target
CN103413151B (en) Hyperspectral image classification method based on figure canonical low-rank representation Dimensionality Reduction
CN108230338B (en) Stereo image segmentation method based on convolutional neural network
CN104240244B (en) A kind of conspicuousness object detecting method based on communication mode and manifold ranking
CN107067405B (en) Remote sensing image segmentation method based on scale optimization
CN102982539B (en) Characteristic self-adaption image common segmentation method based on image complexity
CN112862792B (en) Wheat powdery mildew spore segmentation method for small sample image dataset
CN107369158B (en) Indoor scene layout estimation and target area extraction method based on RGB-D image
CN108229551B (en) Hyperspectral remote sensing image classification method based on compact dictionary sparse representation
CN110443809A (en) Structure sensitive property color images super-pixel method with boundary constraint
CN103310481A (en) Point cloud reduction method based on fuzzy entropy iteration
CN103093470A (en) Rapid multi-modal image synergy segmentation method with unrelated scale feature
CN107146219B (en) Image significance detection method based on manifold regularization support vector machine
CN103903275A (en) Method for improving image segmentation effects by using wavelet fusion algorithm
CN104408733A (en) Object random walk-based visual saliency detection method and system for remote sensing image
CN108388901B (en) Collaborative significant target detection method based on space-semantic channel
CN114299405A (en) Unmanned aerial vehicle image real-time target detection method
CN101383046B (en) Three-dimensional reconstruction method on basis of image
CN112330639A (en) Significance detection method for color-thermal infrared image
CN109345570A (en) A kind of multichannel three-dimensional colour point clouds method for registering based on geometry

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20170801

RJ01 Rejection of invention patent application after publication