WO2021082168A1 - Method for matching specific target object in scene image - Google Patents

Method for matching specific target object in scene image Download PDF

Info

Publication number
WO2021082168A1
WO2021082168A1 PCT/CN2019/122673 CN2019122673W WO2021082168A1 WO 2021082168 A1 WO2021082168 A1 WO 2021082168A1 CN 2019122673 W CN2019122673 W CN 2019122673W WO 2021082168 A1 WO2021082168 A1 WO 2021082168A1
Authority
WO
WIPO (PCT)
Prior art keywords
matrix
image
target object
clustering
similarity
Prior art date
Application number
PCT/CN2019/122673
Other languages
French (fr)
Chinese (zh)
Inventor
郑李明
于涛
崔兵兵
Original Assignee
南京原觉信息科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 南京原觉信息科技有限公司 filed Critical 南京原觉信息科技有限公司
Publication of WO2021082168A1 publication Critical patent/WO2021082168A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/23Clustering techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/56Extraction of image or video features relating to colour

Definitions

  • the invention belongs to the field of image processing, and relates to an application method based on an image clustering algorithm, in particular to an application method for specific target objects in scene images that can be used in the fields of visual navigation, target tracking and positioning, panoramic fusion, and three-dimensional simulation. Matching method.
  • Image clustering is the use of a computer to analyze the images in the image library, and classify each pixel or area in the image into one of several feature categories to replace the human visual judgment of the image.
  • the process of image clustering is essentially the process of image comprehension based on knowledge, and it is also the extension and development of humans' visual discrimination of images.
  • Image clustering technology is to search based on the semantic and perceptual characteristics of the image.
  • the specific implementation is to extract specific information clues or feature indicators from the image data, and then search based on these clues from a large number of images stored in the image database.
  • Image data with similar characteristics Image clustering technology first clusters images according to a certain similarity principle, and aggregates similar images into one category. The retrieval process is carried out within the category, thereby greatly reducing the scope of image retrieval and achieving rapid and accurate image retrieval. purpose.
  • Image clustering technology has broad application prospects in all walks of life.
  • image recognition technology has been widely used in the public security industry.
  • Video pictures have been obtained by means of camera capture and picture structuring, forming a dynamic resource library.
  • Machine vision analysis technology based on image clustering can provide strong support for public security prevention and control, criminal investigation and solving, anti-terrorism and riot prevention.
  • a visual automatic navigation system that uses the surrounding environment information to navigate through a camera installed on the vehicle body. The image information obtained by the camera can be analyzed and processed to obtain the position and posture information of the vehicle relative to the road, make corresponding path planning, and realize the automatic navigation of the vehicle.
  • the current conventional image clustering method is spectral clustering.
  • the main advantage of the spectral clustering method is that spectral clustering only needs the similarity matrix between data, so it is very effective for processing sparse data clustering, which is difficult for traditional clustering algorithms such as K-Means. And because of the use of dimensionality reduction, the complexity of processing high-dimensional data clustering is better than traditional clustering algorithms.
  • the main disadvantage of the spectral clustering method is that if the dimensionality of the final cluster is very high, the computational complexity of dimensionality reduction will be high, so the running speed of spectral clustering is slow and the final clustering effect is not ideal.
  • Feature matching is the matching of key target objects with the same or similar features in multiple images. It is a key link in image clustering and machine vision recognition technologies. It has important applications in the fields of panoramic fusion, monitoring, live broadcasting, and 3D simulation.
  • the visual features adopted by current image clustering methods lack autonomous learning capabilities, resulting in poor image expression, high computational complexity, low clustering efficiency, and difficulty in adapting to the current big data environment. Therefore, it is difficult to adapt to the specific target in the scene image.
  • the matching of objects is also very inefficient.
  • the present invention proposes a matching of a specific target object in a scene image.
  • the purpose of the method is to achieve high efficiency and convenience in matching specific target objects in the scene image.
  • the present invention provides the following technical solutions:
  • a method for matching a specific target object in a scene image includes the following steps:
  • the first step is to segment the scene image into super pixel blocks and extract the super pixel center attributes in each super pixel block, where the super pixel center attributes include a position center and a color center;
  • the second step is to obtain an adjacency matrix reflecting the adjacency relationship between each super pixel block
  • the third step is to obtain a similarity matrix reflecting the similarity between adjacent super-pixel tiles according to the adjacency matrix, where the similarity includes the similarity of the position adjacent relationship and the color;
  • the fourth step is to complete clustering of super pixel tiles according to the similarity matrix
  • the fifth step is to perform image selection and feature extraction on specific target objects in the clustered scene images
  • the sixth step is to search for color tiles that are similar to the feature value of the target object in the scene image.
  • the superpixel center attribute includes the following attributes: coordinate center(x,y) in the image, color_info(l,a,b), superpixel unique identifier id labels, and number of superpixels num_pixels.
  • the specific algorithm for calculating the adjacency matrix is implemented as follows:
  • i and j respectively represent the sequence numbers of the super-pixel tiles
  • the relationship between the super pixel block itself and itself is defined as adjacent.
  • the step of calculating the similarity matrix is to calculate the similarity of the two superpixels according to the adjacent relationship of the superpixel tiles in the adjacency matrix, when the similarity must be greater than a certain threshold, the corresponding element value is set to 1, otherwise it is set to 0, the specific algorithm is implemented as follows:
  • L th , ⁇ th , M th , L th0 , and ⁇ th0 are the threshold values of the three components in the L ⁇ M color space
  • M Cth is the threshold value for distinguishing color and black and white color spaces by modulus length component, usually the value is less than or equal to 2.
  • L i , L j , ⁇ i , ⁇ j , M i , M j are the mean values of super pixel tiles i and j in the L ⁇ M color space respectively; w(i,j) is expressed as two super pixel tiles The similarity of, where a value of 1 is similar, and a value of 0 is dissimilar.
  • the step of clustering is to generate a similarity matrix W using similarity w(i, j), and W is the clustering relationship graph.
  • the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes the step of converting the similarity matrix W into a triangular matrix
  • Triangular matrix set all the lower left corners to zero
  • the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes: the step of completing clustering,
  • a(i min ,n) a(i min ,n) ⁇ ... ⁇ a(n,n)
  • a(i min ,n-1) a(i min ,n-1) ⁇ ... ⁇ a(n,n-1)
  • a(i min ,i min ) a(i min ,i min ) ⁇ ... ⁇ a(n,i min )
  • a(i min ,j n ) a(i min ,j n ) ⁇ ... ⁇ a(n,j n )
  • a(i min ,j n-1 ) a(i min ,j n-1 ) ⁇ ... ⁇ a(n,j n-1 )
  • a(i min ,i min ) a(i min ,j min ) ⁇ ... ⁇ a(n,i min )
  • a(i min ,j n ) a(i min ,j n ) ⁇ ... ⁇ a(n,j n )
  • a(i min ,j n-1 ) a(i min ,j n-1 ) ⁇ ... ⁇ a(n,j n-1 )
  • a(i min ,i min ) a(i min ,j min ) ⁇ ... ⁇ a(n,i min )
  • each row of the triangular matrix is traversed once, and the following similar matrix will be obtained:
  • the image selection and feature extraction of the specific target object in the scene image refers to selecting the image block of the specific target object in the scene according to the image clustering result and extracting the values of L, ⁇ , and M in the corresponding block.
  • the specific method of searching for a color tile similar to the feature value of the target object tile is as follows:
  • L th , ⁇ th , and M th are the threshold values of the three components in the L ⁇ M color space
  • M Cth is the threshold value for distinguishing between color and black and white color spaces by modulus length components.
  • the value is less than or equal to 2
  • Li , L j , ⁇ i , ⁇ j ,M i ,M j are the mean values of super pixel tiles i and j in the L ⁇ M color space, respectively; where i represents the number of the selected target object in the sample image, and j represents the search image
  • the block number of; w(i,j) represents the similarity of two superpixel blocks, where a value of 1 is similar, and a value of 0 is dissimilar;
  • the value of the pixel in the search image block is set to a value that is not in the color space, such as -1, then the block will not participate in the subsequent calculations;
  • Triangular matrix set all the lower left corners to zero:
  • All non-zero row arrays in the matrix are the target object clustering tiles.
  • the image clustering method used in the present invention is a clustering method that simulates the process of human eye recognition of objects. First, the image is divided into super pixel blocks to extract the central attributes of the super pixels in each super pixel block, and then the reflection is calculated. The adjacency matrix of the adjacency between each super-pixel block, and then calculate the similarity matrix reflecting the similarity of the super-pixels between adjacent super-pixel blocks according to the adjacency matrix, and finally complete the super-pixel block according to the similarity matrix Clustering.
  • CPU model i5 4590 main frequency: 3.3GHz;
  • the number of CUDA cores of the GPU is 2880 and the main frequency is 705MHz.
  • the resolution of the calculated image is 1920 ⁇ 1080
  • Clustering method name Number of iterations Operation time (unit: second) Spectral clustering 5 180 Histogram 5 60 New type spectrum clustering (this patent) 1 0.05
  • the present invention proposes a matching method for a specific target object in a scene image on the basis of an image clustering method with a brand-new concept and based on the same concept. First, perform image selection and feature extraction on the specific target object in the clustered scene image; then search the scene image for color blocks that are similar to the feature value of the target object.
  • the calculation speed of the matching calculation of image clustering and specific target objects can be improved, thereby optimizing applications in the fields of target tracking and positioning, panoramic fusion, and three-dimensional simulation.
  • Figure 1 is the original image of the sample scene.
  • Figure 2 is an image after the sample cluster is segmented.
  • Figure 3 is the extraction of the corresponding tiles in the sample scene where the target objects are people and bags.
  • Figure 1 is the original image of the sample scene.
  • Step 1 Recalculate the cluster center Seed.
  • This step is a process of labeling each pixel in the image, which makes the pixels with the same label have a certain common visual characteristic.
  • the result of superpixel segmentation is a collection of sub-regions on the image. The entirety of these sub-regions covers the entire image, or a collection of contour lines extracted from the image, such as edge detection.
  • Each pixel in a super pixel block is similar under a certain characteristic measurement or calculated characteristic, such as color, brightness, and texture. Adjacent regions are very different under a certain characteristic measurement.
  • superpixels are widely used in the initial stage of image segmentation and understanding.
  • the use of superpixels can effectively reduce the redundancy of image local information and reduce the complexity of image processing.
  • Pixels are not the focus of human vision. Because humans obtain images from a region where many pixels are combined, a single pixel has no practical meaning, and only when combined together is meaningful to humans. So in this case there is the concept of "super pixels".
  • super pixel is a small area in the image composed of a series of adjacent pixels with similar characteristics such as color, brightness, texture, etc. Most of these small areas retain effective information for further image segmentation, and generally will not damage Boundary information of objects in the image. Therefore, substituting superpixels for the original pixels as nodes of the graph for image segmentation can greatly reduce the scale of image processing and bring computational advantages.
  • the code is implemented as follows:
  • Step 2 Calculate the adjacency matrix E.
  • This step of the present invention takes into account that since in the clustering of super pixel blocks, only the mutual clustering between adjacent super pixel blocks needs to be considered, and no calculation is required for non-adjacent super pixel blocks. So we first give the adjacency matrix E, which is used to calculate the adjacency matrix for the subsequent similarity clustering.
  • the present invention uses parallel computing:
  • Similarity measure is a function used to compare images.
  • the similarity between images or parts of images is a very important issue at the bottom of the computer vision field.
  • similarity plays a decisive and key role, and different similarity measures will lead to completely different clustering effects.
  • the algorithm idea of this step of the present invention is to calculate the similarity of two superpixels based on the adjacent relationship of the superpixel tiles in the adjacency matrix E.
  • the similarity must be greater than a certain threshold, the corresponding element value is set to 1, otherwise it is set to 0.
  • the specific algorithm is implemented as follows.
  • the first is the transformation of color space, CIE That is, convert from CIE Lab space to L ⁇ M space.
  • This step effectively simulates the conversion of human recognition methods based on the surface color and brightness of the object under different color saturation conditions, and realizes the effective clustering of objects with different color saturation in the scene image, which improves the clustering of the image. Effectiveness and anti-interference ability, the dimensionality reduction effect of image clustering and segmentation is obvious, which can effectively improve the efficiency and accuracy of image analysis.
  • the code is implemented as follows:
  • L th , ⁇ th , M th , L th0 , and ⁇ th0 are the threshold values of the three components in the L ⁇ M color space
  • M Cth is the threshold value for distinguishing color and black and white color spaces by the modulus component, usually the value is less than or equal to 2
  • L i , L j , ⁇ i , ⁇ j , M i , and M j are the mean values of super pixel tiles i and j in the L ⁇ M color space, respectively.
  • w(i,j) is expressed as the similarity of two superpixel tiles, where a value of 1 means similarity, and a value of 0 means dissimilar.
  • the algorithm of this step is to use the similarity w(i,j) to generate the similarity matrix W (W is the clustering relationship graph);
  • the similarity matrix W is transformed into a triangular matrix
  • a(i min ,n) a(i min ,n) ⁇ ... ⁇ a(n,n)
  • a(i min ,n-1) a(i min ,n-1) ⁇ ... ⁇ a(n,n-1)
  • a(i min ,i min ) a(i min ,i min ) ⁇ ... ⁇ a(n,i min )
  • a(i min ,j n ) a(i min ,j n ) ⁇ ... ⁇ a(n,j n )
  • a(i min ,j n-1 ) a(i min ,j n-1 ) ⁇ ... ⁇ a(n,j n-1 )
  • a(i min ,i min ) a(i min ,j min ) ⁇ ... ⁇ a(n,i min )
  • a(i min ,j n ) a(i min ,j n ) ⁇ ... ⁇ a(n,j n )
  • a(i min ,j n-1 ) a(i min ,j n-1 ) ⁇ ... ⁇ a(n,j n-1 )
  • a(i min ,i min ) a(i min ,j min ) ⁇ ... ⁇ a(n,i min )
  • each row of the triangular matrix is traversed once, and the following similar matrix will be obtained:
  • Step 5 Image selection and feature extraction of specific target objects in the scene
  • the image block of a specific target object is selected in the scene and the values of L, ⁇ , and M in the corresponding block are extracted.
  • the target objects in the scene shown in FIG. 3 are the L, ⁇ , and M feature values of people and bags and corresponding tiles. From this, the feature values of the target object block in the image are as follows:
  • Step 6 Search for the target object
  • L th , ⁇ th , and M th are the threshold values of the three components in the L ⁇ M color space
  • M Cth is the threshold value for distinguishing between color and black and white color spaces by modulus length components.
  • the value is less than or equal to 2
  • Li , L j , ⁇ i, ⁇ j, M i, M j respectively superpixel tile i, j Means L ⁇ M color space.
  • i represents the block number of the selected target object in the sample image
  • j represents the block number of the search image.
  • w(i,j) is expressed as the similarity of two superpixel tiles, where a value of 1 means similarity, and a value of 0 means dissimilar.
  • the value of the pixel in the search image block is set to a value that is not in the color space, such as -1, and the block will not participate in the subsequent calculations.
  • Triangular matrix set all the lower left corners to zero:
  • the clustering of the target object is also completed using the method of completing the clustering in step 4. So as to achieve the matching of target objects in different scene images.
  • All non-zero row arrays in the matrix are the target object clustering tiles.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Image Analysis (AREA)

Abstract

A method for matching a specific target object in a scene image, relating to the field of image processing. For the problems, in the existing image clustering and feature matching algorithms, of high calculation complexity, low efficiency, and being difficult to adapt to the current big-data environment, the following method is proposed. The method comprises: performing superpixel image block segmentation on a scene image, extracting a central attribute of superpixels, calculating an adjacent matrix reflecting an adjacent relationship between superpixel image blocks, calculating a similarity matrix reflecting a superpixel between adjacent superpixelimage blocks, clustering superpixel image blocks according to the similarity matrix, performing image selection and feature extraction on a specific target object in the scene image, and searching for image blocks having a color close to a feature value of the target object. Said method improves the speed of matching calculation for a specific target object in a scene image, achieves efficient, convenient, and fast matching, and optimizes use of the technology in the fields of visual navigation, target measurement, and target tracking and positioning.

Description

一种场景图像中特定目标对象的匹配方法A matching method for specific target objects in scene images 技术领域Technical field
本发明属于图像处理领域,涉及一种基于图像聚类算法的应用方法,尤其是一种可以运用于视觉导航、目标追踪与定位、全景融合、三维仿真等领域的对于场景图像中特定目标对象的匹配方法。The invention belongs to the field of image processing, and relates to an application method based on an image clustering algorithm, in particular to an application method for specific target objects in scene images that can be used in the fields of visual navigation, target tracking and positioning, panoramic fusion, and three-dimensional simulation. Matching method.
背景技术Background technique
图像聚类是利用计算机对图像库中的图像进行分析,把图像中的每个像元或区域划归为若干种特征类别中的一种,以代替人类对图像的视觉判别。图像聚类的过程实质上就是基于知识的图像理解过程,同时也是人类对图像的视觉判别的延伸和发展。Image clustering is the use of a computer to analyze the images in the image library, and classify each pixel or area in the image into one of several feature categories to replace the human visual judgment of the image. The process of image clustering is essentially the process of image comprehension based on knowledge, and it is also the extension and development of humans' visual discrimination of images.
图像聚类技术就是根据图像的语义和感知特征进行检索,具体实现就是从图像数据中提取出特定的信息线索或特征指标,然后根据这些线索从大量存储在图像数据库的图像中进行查找,检索出具有相似特征的图像数据。图像聚类技术是先对图像按照某种相似性原则进行聚类,把相似的图像聚合为一类,检索过程在类内进行,从而大大的缩小图像检索范围,就能够达到快速准确检索图像的目的。Image clustering technology is to search based on the semantic and perceptual characteristics of the image. The specific implementation is to extract specific information clues or feature indicators from the image data, and then search based on these clues from a large number of images stored in the image database. Image data with similar characteristics. Image clustering technology first clusters images according to a certain similarity principle, and aggregates similar images into one category. The retrieval process is carried out within the category, thereby greatly reducing the scope of image retrieval and achieving rapid and accurate image retrieval. purpose.
图像聚类技术在各行各业都有着广泛的应用前景。例如在公安行业,随着公安信息化的不断发展,图像识别技术已在公安行业广泛应用,通过摄像头抓拍、图片结构化等手段获取了视频图片,形成了动态资源库。基于图像聚类的机器视觉分析技术可以为公安治安防控、刑侦破案、反恐防暴等工作提供有力支撑。又例如在导航领域,目前有通过安装在车身上的摄像头,利用周围环境信息来导航的视觉自动导航系统。通过摄像头获取的图像信息,经过分析处理可以得到车辆相对于道路的位置与姿态信息,做出相应的路径规划,实现车辆 的自动导航。Image clustering technology has broad application prospects in all walks of life. For example, in the public security industry, with the continuous development of public security informatization, image recognition technology has been widely used in the public security industry. Video pictures have been obtained by means of camera capture and picture structuring, forming a dynamic resource library. Machine vision analysis technology based on image clustering can provide strong support for public security prevention and control, criminal investigation and solving, anti-terrorism and riot prevention. For another example, in the field of navigation, there is currently a visual automatic navigation system that uses the surrounding environment information to navigate through a camera installed on the vehicle body. The image information obtained by the camera can be analyzed and processed to obtain the position and posture information of the vehicle relative to the road, make corresponding path planning, and realize the automatic navigation of the vehicle.
目前常规的图像聚类方法是谱聚类法。谱聚类法的主要优点是谱聚类只需要数据之间的相似度矩阵,因此对于处理稀疏数据的聚类很有效,这点传统聚类算法比如K-Means很难做到。并且由于使用了降维,因此在处理高维数据聚类时的复杂度比传统聚类算法好。但是谱聚类法的主要缺点是如果最终聚类的维度非常高,则由于降维的运算复杂度会较高,因此谱聚类的运行速度较慢且最后的聚类效果不够理想。The current conventional image clustering method is spectral clustering. The main advantage of the spectral clustering method is that spectral clustering only needs the similarity matrix between data, so it is very effective for processing sparse data clustering, which is difficult for traditional clustering algorithms such as K-Means. And because of the use of dimensionality reduction, the complexity of processing high-dimensional data clustering is better than traditional clustering algorithms. However, the main disadvantage of the spectral clustering method is that if the dimensionality of the final cluster is very high, the computational complexity of dimensionality reduction will be high, so the running speed of spectral clustering is slow and the final clustering effect is not ideal.
特征匹配是对多幅图像中具有相同或者相似特征的关键目标对象进行匹配,是图像聚类和机器视觉识别等技术的关键环节,在全景融合、监控、直播及三维仿真等领域有着重要应用。基于当前图像聚类方法采用的视觉特征缺乏自主学习能力,导致图像表达能力不强,计算复杂度较高,聚类效率低,难以适应当前的大数据环境的问题,因此对场景图像中特定目标对象的匹配实现起来效率也十分低下。Feature matching is the matching of key target objects with the same or similar features in multiple images. It is a key link in image clustering and machine vision recognition technologies. It has important applications in the fields of panoramic fusion, monitoring, live broadcasting, and 3D simulation. The visual features adopted by current image clustering methods lack autonomous learning capabilities, resulting in poor image expression, high computational complexity, low clustering efficiency, and difficulty in adapting to the current big data environment. Therefore, it is difficult to adapt to the specific target in the scene image. The matching of objects is also very inefficient.
因此,寻找一种高效率且方便快捷的图像聚类方法,然后基于该图像聚类方法实现场景图像中特定目标对象的匹配,已经成为进行图像处理工作的重要基础和必不可少的重要环节。Therefore, finding an efficient and convenient image clustering method, and then realizing the matching of specific target objects in the scene image based on the image clustering method, has become an important foundation and an indispensable important link for image processing.
发明内容Summary of the invention
针对上述背景技术中所提到的目前的图像聚类以及特征匹配计算复杂度较高,效率低下,难以适应当前的大数据环境的问题,本发明提出了一种场景图像中特定目标对象的匹配方法,目的在于实现对于场景图像中特定目标对象匹配工作的高效率和方便快捷。In view of the current image clustering and feature matching mentioned in the background art, the computational complexity is relatively high, the efficiency is low, and it is difficult to adapt to the current big data environment. The present invention proposes a matching of a specific target object in a scene image. The purpose of the method is to achieve high efficiency and convenience in matching specific target objects in the scene image.
为了达到上述目的,本发明提供如下技术方案:In order to achieve the above objective, the present invention provides the following technical solutions:
一种场景图像中特定目标对象的匹配方法,包括以下步骤:A method for matching a specific target object in a scene image includes the following steps:
第一步、对场景图像进行超像素图块的分割并提取各个超像素图块中的超像素中心属性,所述超像素中心属性包含位置中心和色彩中心;The first step is to segment the scene image into super pixel blocks and extract the super pixel center attributes in each super pixel block, where the super pixel center attributes include a position center and a color center;
第二步、获得反映各个超像素图块之间相邻关系的邻接矩阵;The second step is to obtain an adjacency matrix reflecting the adjacency relationship between each super pixel block;
第三步、根据所述邻接矩阵获得反映相邻超像素图块之间的相似程度的相似度矩阵,所述相似度包含位置相邻关系和色彩的相似程度;The third step is to obtain a similarity matrix reflecting the similarity between adjacent super-pixel tiles according to the adjacency matrix, where the similarity includes the similarity of the position adjacent relationship and the color;
第四步、根据所述相似度矩阵对超像素图块完成聚类;The fourth step is to complete clustering of super pixel tiles according to the similarity matrix;
第五步、在聚类后的场景图像中对特定目标对象进行图像选择与特征提取;The fifth step is to perform image selection and feature extraction on specific target objects in the clustered scene images;
第六步、在场景图像中搜索与目标对象的特征值相近的颜色图块。The sixth step is to search for color tiles that are similar to the feature value of the target object in the scene image.
优选的,所述超像素中心属性包括如下属性:在图像中的坐标center(x,y),颜色color_info(l,a,b),超像素唯一标识id labels,超像素个数num_pixels。Preferably, the superpixel center attribute includes the following attributes: coordinate center(x,y) in the image, color_info(l,a,b), superpixel unique identifier id labels, and number of superpixels num_pixels.
优选的,所述计算邻接矩阵的具体的算法实现如下:Preferably, the specific algorithm for calculating the adjacency matrix is implemented as follows:
Figure PCTCN2019122673-appb-000001
Figure PCTCN2019122673-appb-000001
其中,i,j分别是代表超像素图块序号;Among them, i and j respectively represent the sequence numbers of the super-pixel tiles;
邻接矩阵E中每个元数e(i,j)满足如下函数关系:Each element e(i,j) in the adjacency matrix E satisfies the following functional relationship:
Figure PCTCN2019122673-appb-000002
Figure PCTCN2019122673-appb-000002
其中,超像素图块自身与自身之间的关系定义为相邻。Among them, the relationship between the super pixel block itself and itself is defined as adjacent.
优选的,所述计算相似度矩阵的步骤是根据邻接矩阵中超像素图块的相邻关系计算两个超像素的相似度,当相似度必须大于一定阈值时将相应元数值置1,否则置为0,具体的算法实现如下:Preferably, the step of calculating the similarity matrix is to calculate the similarity of the two superpixels according to the adjacent relationship of the superpixel tiles in the adjacency matrix, when the similarity must be greater than a certain threshold, the corresponding element value is set to 1, otherwise it is set to 0, the specific algorithm is implemented as follows:
(1)从CIE Lab色彩空间转换为LθM色彩空间(1) Convert from CIE Lab color space to LθM color space
θ′=atan2(B,A) θ′∈(-π,π]    (公式3-1)θ′=atan2(B,A) θ′∈(-π,π] (Formula 3-1)
Figure PCTCN2019122673-appb-000003
Figure PCTCN2019122673-appb-000003
Figure PCTCN2019122673-appb-000004
Figure PCTCN2019122673-appb-000004
l=L l∈[0,100]l=L l∈[0, 100]
(2)相似度计算(2) Similarity calculation
Figure PCTCN2019122673-appb-000005
Figure PCTCN2019122673-appb-000005
其中,L thth,M th,L th0th0分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值;w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似。 Among them, L th , θ th , M th , L th0 , and θ th0 are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing color and black and white color spaces by modulus length component, usually the value is less than or equal to 2. L i , L j , θ i , θ j , M i , M j are the mean values of super pixel tiles i and j in the LθM color space respectively; w(i,j) is expressed as two super pixel tiles The similarity of, where a value of 1 is similar, and a value of 0 is dissimilar.
优选的,所述聚类的步骤是利用相似度w(i,j)生成相似度矩阵W,W即为聚类关系图。Preferably, the step of clustering is to generate a similarity matrix W using similarity w(i, j), and W is the clustering relationship graph.
优选的,所述基于相似度矩阵W完成聚类的具体的算法实现包括:将相似度矩阵W转换为三角矩阵的步骤,Preferably, the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes the step of converting the similarity matrix W into a triangular matrix,
相似度矩阵Similarity matrix
Figure PCTCN2019122673-appb-000006
Figure PCTCN2019122673-appb-000006
三角矩阵,将左下角全部置零,Triangular matrix, set all the lower left corners to zero,
Figure PCTCN2019122673-appb-000007
Figure PCTCN2019122673-appb-000007
优选的,所述基于相似度矩阵W完成聚类的具体的算法实现包括:完成聚类的步骤,Preferably, the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes: the step of completing clustering,
对三角矩阵执行聚类算法Perform clustering algorithm on triangular matrix
Figure PCTCN2019122673-appb-000008
Figure PCTCN2019122673-appb-000008
第一步:first step:
从矩阵的第n行n列开始,搜索所有n列上为1的数组,如果第n列上为1的数组只有第n行,则a(n,n)=1,否则a(n,n)=0。Starting from the nth row and nth column of the matrix, search all the arrays with 1 on the nth column. If the array with 1 on the nth column has only the nth row, then a(n,n)=1, otherwise a(n,n )=0.
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000009
Figure PCTCN2019122673-appb-000009
如果in case
a(n,n)=0a(n,n)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给n列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number among the n columns [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero item or calculation algorithm of the column is as follows:
a(i min,n)=a(i min,n)∪...∪a(n,n) a(i min ,n)=a(i min ,n)∪...∪a(n,n)
a(i min,n-1)=a(i min,n-1)∪...∪a(n,n-1) a(i min ,n-1)=a(i min ,n-1)∪...∪a(n,n-1)
………………………………………………………………………………
a(i min,i min)=a(i min,i min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,i min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(n,n)=0a(n,n)=0
本次运算结束;The end of this calculation;
第二步:The second step:
从矩阵的第n-1行n-1列开始,搜索所有n-1列上为1的数组,如果第n-1列上为1的数组只有第n-1行,则a(n-1,n-1)=1否则a(n-1,n-1)=0Starting from the n-1th row and n-1 column of the matrix, search all the arrays with 1 in the n-1 column. If the array with 1 in the n-1th column has only the n-1th row, then a(n-1 , N-1)=1 otherwise a(n-1, n-1)=0
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000010
Figure PCTCN2019122673-appb-000010
如果in case
a(n-1,n-1)=0a(n-1, n-1)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n-1)进行逻辑或运算,并将结果赋值给n-1列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列的非零项或运算算法如下: Then these arrays are logically ORed on each column (1, 2, 3...n-1) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in the n-1 column [0 ,0,……a(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero items or calculation algorithms of the columns are as follows:
a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
………………………………………………………………………………
a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(n-1,n~n-1)=0a(n-1, n~n-1)=0
,本次运算结束;, This calculation is over;
第三步:third step:
以此类推,从矩阵的第i行i列开始,搜索所有i列上为1的数组,如果第i列上为1的数组只有第i行,则a(i,i)=1否则a(i,i)=0By analogy, starting from the i-th row and i-column of the matrix, search for all arrays with 1 on the i-th column. If the array with 1 on the i-th column has only the i-th row, then a(i, i) = 1 otherwise a( i, i) = 0
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000011
Figure PCTCN2019122673-appb-000011
如果in case
a(i,i)=0a(i,i)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给i列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列中的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in column i [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero term or calculation algorithm in the column is as follows:
a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
………………………………………………………………………………
a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(i,n~n-i)=0a(i, n~n-i)=0
本次运算结束;The end of this calculation;
第四步:the fourth step:
根据以上的算法将三角矩阵的每一行都遍历一遍,将得到如下类似矩阵:According to the above algorithm, each row of the triangular matrix is traversed once, and the following similar matrix will be obtained:
Figure PCTCN2019122673-appb-000012
Figure PCTCN2019122673-appb-000012
则矩阵中所有非零的行数组即为聚类图块的数组。Then all the non-zero row arrays in the matrix are the arrays of clustering tiles.
优选的,所述场景图像中特定目标对象的图像选择与特征提取是指根据图像聚类结果在场景中选择特定目标对象的图像块并提取出对应图块中L、θ、M的取值。Preferably, the image selection and feature extraction of the specific target object in the scene image refers to selecting the image block of the specific target object in the scene according to the image clustering result and extracting the values of L, θ, and M in the corresponding block.
优选的,所述搜索与目标对象图块的特征值相近的颜色图块的具体方法如下:Preferably, the specific method of searching for a color tile similar to the feature value of the target object tile is as follows:
Figure PCTCN2019122673-appb-000013
Figure PCTCN2019122673-appb-000013
其中L thth,M th分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值;其中i表示样本图像中已选定的目标对象的图块号,j表示搜索图像的图块号;w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似; Among them, L th , θ th , and M th are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing between color and black and white color spaces by modulus length components. Usually the value is less than or equal to 2, Li , L jij ,M i ,M j are the mean values of super pixel tiles i and j in the LθM color space, respectively; where i represents the number of the selected target object in the sample image, and j represents the search image The block number of; w(i,j) represents the similarity of two superpixel blocks, where a value of 1 is similar, and a value of 0 is dissimilar;
对搜索图像的图块进行如下操作:Perform the following operations on the tiles of the search image:
如果w(i,j)=1,则保留搜索图像图块中像素的原始数值不变;If w(i,j)=1, keep the original value of the pixel in the search image block unchanged;
如果w(i,j)=0,则将搜索图像图块中像素的值设置为不在色彩空间的数值如-1,则该图块将不参与后面的运算;If w(i,j)=0, then the value of the pixel in the search image block is set to a value that is not in the color space, such as -1, then the block will not participate in the subsequent calculations;
将搜索到的图块构建邻接矩阵:Construct an adjacency matrix with the searched tiles:
Figure PCTCN2019122673-appb-000014
Figure PCTCN2019122673-appb-000014
三角矩阵,将左下角全部置零:Triangular matrix, set all the lower left corners to zero:
Figure PCTCN2019122673-appb-000015
Figure PCTCN2019122673-appb-000015
对三角矩阵采用与上述同样的聚类的方法完成对目标对象的聚类;从而实现在不同场景图像中对目标对象的匹配:Use the same clustering method as the above to complete the clustering of the target object for the triangular matrix; thus achieve the matching of the target object in different scene images:
Figure PCTCN2019122673-appb-000016
Figure PCTCN2019122673-appb-000016
矩阵中所有非零的行数组即为目标对象聚类图块。All non-zero row arrays in the matrix are the target object clustering tiles.
由于采用上述方案,本发明的有益效果是:Due to the above scheme, the beneficial effects of the present invention are:
本发明采用的图像聚类方法是模拟人眼对物件识别的过程的聚类方法,首先通过对图像进行超像素图块的分割,提取各个超像素图块中的超像素中心属性,接着计算反映各个超像素图块之间相邻关系的邻接矩阵,然后根据邻接矩阵计算反映相邻超像素图块之间的超像素的相似程度的相似度矩阵,最后根据相似度矩阵对超像素图块完成聚类。The image clustering method used in the present invention is a clustering method that simulates the process of human eye recognition of objects. First, the image is divided into super pixel blocks to extract the central attributes of the super pixels in each super pixel block, and then the reflection is calculated. The adjacency matrix of the adjacency between each super-pixel block, and then calculate the similarity matrix reflecting the similarity of the super-pixels between adjacent super-pixel blocks according to the adjacency matrix, and finally complete the super-pixel block according to the similarity matrix Clustering.
本发明方法与传统的谱聚类、直方图聚类方法的运算性能的比较如下:The calculation performance of the method of the present invention and the traditional spectral clustering and histogram clustering methods are as follows:
计算机的配置:CPU+GPUComputer configuration: CPU+GPU
其中:CPU型号i5=4590主频:3.3GHz;Among them: CPU model i5=4590 main frequency: 3.3GHz;
GPU的CUDA核心数2880主频705MHz。The number of CUDA cores of the GPU is 2880 and the main frequency is 705MHz.
运算图像的分辨率为:1920×1080The resolution of the calculated image is 1920×1080
不同图像聚类方法的运算性能的比较:Comparison of computing performance of different image clustering methods:
聚类方法名称Clustering method name 迭次数Number of iterations 运算时间(单位:秒)Operation time (unit: second)
谱聚类Spectral clustering 55 180180
直方图Histogram 55 6060
新型类谱聚类(本专利)New type spectrum clustering (this patent) 11 0.050.05
由上述对比表格可知,本发明提出的聚类方法在运算性能上,明显优于传统的谱聚类、直方图聚类方法。It can be seen from the above comparison table that the clustering method proposed by the present invention is significantly better than the traditional spectral clustering and histogram clustering methods in terms of operational performance.
本发明在采用全新理念的图像聚类方法的基础上,基于同样的理念提出场景图像中特定目标对象的匹配方法。首先在聚类后的场景图像中对特定目标对象进行图像选择与特征提取;然后在场景图像中搜索与目标对象的特征值相近的颜色图块。The present invention proposes a matching method for a specific target object in a scene image on the basis of an image clustering method with a brand-new concept and based on the same concept. First, perform image selection and feature extraction on the specific target object in the clustered scene image; then search the scene image for color blocks that are similar to the feature value of the target object.
通过本发明的技术方案,可以提高对图像聚类和特定目标对象的匹配计算的运算速度,从而优化目标追踪与定位、全景融合及三维仿真等领域的应用。Through the technical scheme of the present invention, the calculation speed of the matching calculation of image clustering and specific target objects can be improved, thereby optimizing applications in the fields of target tracking and positioning, panoramic fusion, and three-dimensional simulation.
附图说明Description of the drawings
图1是样本场景原始图像。Figure 1 is the original image of the sample scene.
图2是样本聚类分割后的图像。Figure 2 is an image after the sample cluster is segmented.
图3是样本场景中的目标对象为人和包的对应图块的提取。Figure 3 is the extraction of the corresponding tiles in the sample scene where the target objects are people and bags.
具体实施方式Detailed ways
下面将结合示意图对本发明的具体实施方式进行更详细的描述。根据下列描述,本发明的优点和特征将更清楚。需说明的是,附图均采用非常简化的形式且均使用非精准的比例,仅用以方便、明晰地辅助说明本发明实施例的目的。The specific embodiments of the present invention will be described in more detail below in conjunction with the schematic diagrams. Based on the following description, the advantages and features of the present invention will become clearer. It should be noted that the drawings all adopt a very simplified form and all use imprecise proportions, which are only used to conveniently and clearly assist in explaining the purpose of the embodiments of the present invention.
以下通过一个实施例对本发明的计算过程以及聚类和匹配效果进行验证。In the following, an embodiment is used to verify the calculation process and the clustering and matching effects of the present invention.
请参见图1,为样本场景原始图像。Please refer to Figure 1, which is the original image of the sample scene.
本发明的算法具体步骤如下:The specific steps of the algorithm of the present invention are as follows:
步骤一、重算聚类中心Seed。Step 1. Recalculate the cluster center Seed.
这一步骤是对图像中的每个像素加标签的一个过程,该过程使得具有相同 标签的像素具有某种共同视觉特性。超像素分割的结果是图像上子区域的集合,这些子区域的全体覆盖了整个图像,或是从图像中提取的轮廓线的集合,例如边缘检测。一个超像素图块中的每个像素在某种特性的度量下或是由计算得出的特性都是相似的,例如颜色、亮度、纹理。邻接区域在某种特性的度量下有很大的不同。This step is a process of labeling each pixel in the image, which makes the pixels with the same label have a certain common visual characteristic. The result of superpixel segmentation is a collection of sub-regions on the image. The entirety of these sub-regions covers the entire image, or a collection of contour lines extracted from the image, such as edge detection. Each pixel in a super pixel block is similar under a certain characteristic measurement or calculated characteristic, such as color, brightness, and texture. Adjacent regions are very different under a certain characteristic measurement.
在计算机视觉领域,超像素被广泛应用于图像分割与理解的初始阶段,使用超像素可以有效减少图像局部信息的冗余,使图像处理复杂度降低。像素并不是人类视觉的着重点。因为人类获得图像是从许多的像素点的组合的一个区域而来的,单一的某个像素点并不什么实际意义,只有组合在一起对人类而言才有意义。因而在这种情形下有了“超像素”的概念。所谓超像素,即在图像中由一系列位置相邻且颜色、亮度、纹理等特征相似的像素点组成的小区域,这些小区域大多保留了进一步进行图像分割的有效信息,且一般不会破坏图像中物体的边界信息。所以,以超像素代替原来的像素点作为图的节点进行图像分割可以大大减小图像处理的规模,带来计算上的优势。In the field of computer vision, superpixels are widely used in the initial stage of image segmentation and understanding. The use of superpixels can effectively reduce the redundancy of image local information and reduce the complexity of image processing. Pixels are not the focus of human vision. Because humans obtain images from a region where many pixels are combined, a single pixel has no practical meaning, and only when combined together is meaningful to humans. So in this case there is the concept of "super pixels". The so-called super pixel is a small area in the image composed of a series of adjacent pixels with similar characteristics such as color, brightness, texture, etc. Most of these small areas retain effective information for further image segmentation, and generally will not damage Boundary information of objects in the image. Therefore, substituting superpixels for the original pixels as nodes of the graph for image segmentation can greatly reduce the scale of image processing and bring computational advantages.
本发明中定义超像素中心属性如下:In the present invention, the central attributes of superpixels are defined as follows:
Figure PCTCN2019122673-appb-000017
Figure PCTCN2019122673-appb-000017
代码实现如下:The code is implemented as follows:
Figure PCTCN2019122673-appb-000018
Figure PCTCN2019122673-appb-000018
Figure PCTCN2019122673-appb-000019
Figure PCTCN2019122673-appb-000019
上述代码仅作为参考。The above code is for reference only.
步骤二、计算邻接矩阵E。Step 2: Calculate the adjacency matrix E.
本发明的这一步骤考虑到,由于在对超像素图块的聚类中,只需考虑邻接的超像素图块之间相互聚类,而对于不相邻的超像素图块无需进行计算,所以我们首先给出邻接矩阵E,这一步骤计算邻接矩阵是为后续的相似度聚类服务的。This step of the present invention takes into account that since in the clustering of super pixel blocks, only the mutual clustering between adjacent super pixel blocks needs to be considered, and no calculation is required for non-adjacent super pixel blocks. So we first give the adjacency matrix E, which is used to calculate the adjacency matrix for the subsequent similarity clustering.
本发明采用并行计算:The present invention uses parallel computing:
Figure PCTCN2019122673-appb-000020
Figure PCTCN2019122673-appb-000020
(注:i,j分别是代表超像素图块序号)(Note: i and j respectively represent the serial numbers of super pixel tiles)
邻接矩阵E中每个元数e(i,j)满足如下函数关系:Each element e(i,j) in the adjacency matrix E satisfies the following functional relationship:
Figure PCTCN2019122673-appb-000021
Figure PCTCN2019122673-appb-000021
(注:超像素图块自身与自身之间的关系定义为相邻)(Note: The relationship between the super pixel block itself and itself is defined as adjacent)
步骤三、相似度矩阵WStep three, similarity matrix W
相似度量用于比较图像的一个函数。图像与图像之间或者图像的一部分之间的相似度是计算机视觉领域底层十分重要的问题。对于我们提出的图像聚类算法而言,相似度起着决定性的关键作用,不同的相似度量方式会导致截然不同的聚类效果。Similarity measure is a function used to compare images. The similarity between images or parts of images is a very important issue at the bottom of the computer vision field. For the image clustering algorithm we proposed, similarity plays a decisive and key role, and different similarity measures will lead to completely different clustering effects.
本发明这一步骤的算法思想是根据邻接矩阵E中超像素图块的相邻关系计算两个超像素的相似度,当相似度必须大于一定阈值时将相应元数值置1,否则置为0,具体的算法实现如下。The algorithm idea of this step of the present invention is to calculate the similarity of two superpixels based on the adjacent relationship of the superpixel tiles in the adjacency matrix E. When the similarity must be greater than a certain threshold, the corresponding element value is set to 1, otherwise it is set to 0. The specific algorithm is implemented as follows.
(注:算法可以根据场景的不同而改变为不同的参数和公式)(Note: The algorithm can be changed to different parameters and formulas according to different scenarios)
计算公式如下:Calculated as follows:
首先是色彩空间的变换,CIE
Figure PCTCN2019122673-appb-000022
即从CIE Lab空间转换为LθM空间。
The first is the transformation of color space, CIE
Figure PCTCN2019122673-appb-000022
That is, convert from CIE Lab space to LθM space.
这一步骤有效的模拟了人类对不同色彩饱和度条件下对基于物体表面颜色和亮度的识别方式的转换,实现的对场景图像中不同色彩饱和度物体的有效聚类,提高了图像的聚类效果和抗干扰能力,对图像聚类分割的降维效果明显,可有效提高图像分析的效率和准确度。This step effectively simulates the conversion of human recognition methods based on the surface color and brightness of the object under different color saturation conditions, and realizes the effective clustering of objects with different color saturation in the scene image, which improves the clustering of the image. Effectiveness and anti-interference ability, the dimensionality reduction effect of image clustering and segmentation is obvious, which can effectively improve the efficiency and accuracy of image analysis.
此色彩空间具体可参见申请人的公开号为CN104063707A、专利号为ZL201410334974.3的中国专利申请文件《基于人类视觉多尺度感知特性的彩色图像聚类分割方法》上所述的色彩空间。For details of this color space, please refer to the color space described in the Chinese patent application document "Color image clustering and segmentation method based on multi-scale perception characteristics of human vision" with the applicant's publication number CN104063707A and patent number ZL201410334974.3.
θ′=atan2(B,A) θ′∈(-π,π]    (公式3-1)θ′=atan2(B,A) θ′∈(-π,π] (Formula 3-1)
Figure PCTCN2019122673-appb-000023
Figure PCTCN2019122673-appb-000023
Figure PCTCN2019122673-appb-000024
Figure PCTCN2019122673-appb-000024
l=L l∈[0,100]l=L l∈[0, 100]
代码实现如下:The code is implemented as follows:
Figure PCTCN2019122673-appb-000025
Figure PCTCN2019122673-appb-000025
Figure PCTCN2019122673-appb-000026
Figure PCTCN2019122673-appb-000026
然后是,相似度计算Then, the similarity calculation
Figure PCTCN2019122673-appb-000027
Figure PCTCN2019122673-appb-000027
其中L thth,M th,L th0th0分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值。w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似。 Among them, L th , θ th , M th , L th0 , and θ th0 are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing color and black and white color spaces by the modulus component, usually the value is less than or equal to 2 , L i , L j , θ i , θ j , M i , and M j are the mean values of super pixel tiles i and j in the LθM color space, respectively. w(i,j) is expressed as the similarity of two superpixel tiles, where a value of 1 means similarity, and a value of 0 means dissimilar.
步骤四、聚类Step four, clustering
本步骤的算法是利用相似度w(i,j)生成相似度矩阵W(W即为聚类关系图);The algorithm of this step is to use the similarity w(i,j) to generate the similarity matrix W (W is the clustering relationship graph);
基于相似度矩阵W完成聚类的算法步骤如下:The algorithm steps to complete clustering based on the similarity matrix W are as follows:
首先,相似度矩阵W转换为三角矩阵First, the similarity matrix W is transformed into a triangular matrix
相似度矩阵Similarity matrix
Figure PCTCN2019122673-appb-000028
Figure PCTCN2019122673-appb-000028
三角矩阵(将左下角全部置零)Triangular matrix (set all the lower left corners to zero)
Figure PCTCN2019122673-appb-000029
Figure PCTCN2019122673-appb-000029
然后,完成聚类Then, complete the clustering
对三角矩阵执行聚类算法Perform clustering algorithm on triangular matrix
Figure PCTCN2019122673-appb-000030
Figure PCTCN2019122673-appb-000030
第一步:first step:
从矩阵的第n行n列开始,搜索所有n列上为1的数组,如果第n列上为1的数组只有第n行,则a(n,n)=1,否则a(n,n)=0。Starting from the nth row and nth column of the matrix, search all the arrays with 1 on the nth column. If the array with 1 on the nth column has only the nth row, then a(n,n)=1, otherwise a(n,n )=0.
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000031
Figure PCTCN2019122673-appb-000031
如果in case
a(n,n)=0a(n,n)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给n列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中。列的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number among the n columns [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]. The non-zero term OR calculation algorithm of the column is as follows:
a(i min,n)=a(i min,n)∪...∪a(n,n) a(i min ,n)=a(i min ,n)∪...∪a(n,n)
a(i min,n-1)=a(i min,n-1)∪...∪a(n,n-1) a(i min ,n-1)=a(i min ,n-1)∪...∪a(n,n-1)
………………………………………………………………………………
a(i min,i min)=a(i min,i min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,i min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(n,n)=0a(n,n)=0
本次运算结束。This calculation is over.
第二步:The second step:
从矩阵的第n-1行n-1列开始,搜索所有n-1列上为1的数组,如果第n-1列上为1的数组只有第n-1行,则a(n-1,n-1)=1否则a(n-1,n-1)=0Starting from the n-1th row and n-1 column of the matrix, search all the arrays with 1 in the n-1 column. If the array with 1 in the n-1th column has only the n-1th row, then a(n-1 , N-1)=1 otherwise a(n-1, n-1)=0
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000032
Figure PCTCN2019122673-appb-000032
如果in case
a(n-1,n-1)=0a(n-1, n-1)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n-1)进行逻辑或运算,并将结果赋值给n-1列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中。列的非零项或运算算法如下: Then these arrays are logically ORed on each column (1, 2, 3...n-1) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in the n-1 column [0 ,0,……a(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]. The non-zero term OR calculation algorithm of the column is as follows:
a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
………………………………………………………………………………
a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(n-1,n~n-1)=0a(n-1, n~n-1)=0
,本次运算结束。, This calculation is over.
第三步:third step:
以此类推,从矩阵的第i行i列开始,搜索所有i列上为1的数组,如果第i列上为1的数组只有第i行,则a(i,i)=1否则a(i,i)=0By analogy, starting from the i-th row and i-column of the matrix, search for all arrays with 1 on the i-th column. If the array with 1 on the i-th column has only the i-th row, then a(i, i) = 1 otherwise a( i, i) = 0
公式如下:The formula is as follows:
Figure PCTCN2019122673-appb-000033
Figure PCTCN2019122673-appb-000033
如果in case
a(i,i)=0a(i,i)=0
则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给i列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中。列中的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in column i [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]. The non-zero term OR calculation algorithm in the column is as follows:
a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
………………………………………………………………………………
a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
赋值运算:Assignment operation:
a(i,n~n-i)=0a(i, n~n-i)=0
,本次运算结束。, This calculation is over.
第四步:the fourth step:
根据以上的算法将三角矩阵的每一行都遍历一遍,将得到如下类似矩阵:According to the above algorithm, each row of the triangular matrix is traversed once, and the following similar matrix will be obtained:
Figure PCTCN2019122673-appb-000034
Figure PCTCN2019122673-appb-000034
则矩阵中所有非零的行数组即为聚类图块的数组。Then all the non-zero row arrays in the matrix are the arrays of clustering tiles.
请参见图2,为样本聚类分割后的图像。Please refer to Figure 2 for the image after the sample clustering segmentation.
步骤五、场景中特定目标对象的图像选择与特征提取Step 5. Image selection and feature extraction of specific target objects in the scene
第一步:first step:
根据图像聚类结果在场景中选择特定目标对象的图像块并提取出对应图块中L、θ、M的取值。According to the result of image clustering, the image block of a specific target object is selected in the scene and the values of L, θ, and M in the corresponding block are extracted.
例如图3所示的场景中的目标对象为人和包及对应图块的L、θ、M特征值。由此得到图像中的目标对象图块的特征值如下:For example, the target objects in the scene shown in FIG. 3 are the L, θ, and M feature values of people and bags and corresponding tiles. From this, the feature values of the target object block in the image are as follows:
人物的图块特征值Character's tile feature value
图块名称Tile name LL θθ MM
头部head 52.920852.9208 65.725665.7256 8.590288.59028
衣领collar 50.684750.6847 264.383264.383 6.218156.21815
上衣coat 56.484356.4843 256.406256.406 8.633198.63319
手臂1Arm 1 52.205352.2053 80.977880.9778 11.640511.6405
手臂2Arm 2 54.012754.0127 32.827432.8274 3.961283.96128
手臂3Arm 3 54.956154.9561 75.598975.5989 17.092317.0923
裤子1Pants 1 42.673142.6731 230.979230.979 3.419893.41989
裤子2Pants 2 37.002137.0021 178.001178.001 1.578441.57844
裤子3Pants 3 48.993248.9932 232.782232.782 4.127334.12733
裤子4Pants 4 47.472247.4722 187.426187.426 1.553891.55389
裤子5Pants 5 43.570543.5705 220.827220.827 2.292072.29207
裤子6Pants 6 48.94848.948 241.587241.587 3.68873.6887
裤子7Pants 7 48.803148.8031 200.716200.716 1.693211.69321
鞋子1Shoes 1 55.20755.207 210.237210.237 3.210313.21031
鞋子2Shoes 2 52.88552.885 181.793181.793 3.261883.26188
包的图块特征值Block eigenvalues of the package
图块名称Tile name LL θθ MM
包的黄色块Pack of yellow blocks 78.391878.3918 94.334994.3349 49.071549.0715
包的红色块1Pack of red blocks 1 83.08583.085 20.727120.7271 11.836711.8367
包的红色块2Pack of red blocks 2 58.962258.9622 27.18427.184 34.952334.9523
注:上述颜色是指的样本图像的实际颜色,本申请文件中附图的由于格式规定只提供黑白图样作为参考。Note: The above colors refer to the actual colors of the sample images. The drawings in this application file only provide black and white drawings for reference due to the format regulations.
步骤六、目标对象的搜索Step 6: Search for the target object
根据如前所述的步骤一至五的图像聚类方法,得到场景聚类分割图,并在场景中利用(公式3-4)搜索目标对象图块的特征值相近的颜色图块。具体方法法如下:According to the image clustering method of steps 1 to 5 as described above, a scene clustering segmentation map is obtained, and (Equation 3-4) is used to search for color blocks with similar feature values of the target object blocks in the scene. The specific method is as follows:
Figure PCTCN2019122673-appb-000035
Figure PCTCN2019122673-appb-000035
其中L thth,M th分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值。其中i表示样本图像中已选定的目标对象的图块号,j表示搜索图像的图块号。w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似。 Among them, L th , θ th , and M th are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing between color and black and white color spaces by modulus length components. Usually the value is less than or equal to 2, Li , L j , θ i, θ j, M i, M j respectively superpixel tile i, j Means LθM color space. Where i represents the block number of the selected target object in the sample image, and j represents the block number of the search image. w(i,j) is expressed as the similarity of two superpixel tiles, where a value of 1 means similarity, and a value of 0 means dissimilar.
对搜索图像的图块进行如下操作:Perform the following operations on the tiles of the search image:
如果w(i,j)=1,则保留搜索图像图块中像素的原始数值不变;If w(i,j)=1, keep the original value of the pixel in the search image block unchanged;
如果w(i,j)=0,则将搜索图像图块中像素的值设置为不在色彩空间的数值如-1,则该图块将不参与后面的运算。If w(i,j)=0, then the value of the pixel in the search image block is set to a value that is not in the color space, such as -1, and the block will not participate in the subsequent calculations.
然后,将搜索到的图块构建邻接矩阵:Then, construct the adjacency matrix of the searched tiles:
Figure PCTCN2019122673-appb-000036
Figure PCTCN2019122673-appb-000036
三角矩阵,将左下角全部置零:Triangular matrix, set all the lower left corners to zero:
Figure PCTCN2019122673-appb-000037
Figure PCTCN2019122673-appb-000037
同样采用步骤四中完成聚类的方法完成对目标对象的聚类。从而实现在不同场景图像中对目标对象的匹配。The clustering of the target object is also completed using the method of completing the clustering in step 4. So as to achieve the matching of target objects in different scene images.
Figure PCTCN2019122673-appb-000038
Figure PCTCN2019122673-appb-000038
矩阵中所有非零的行数组即为目标对象聚类图块。All non-zero row arrays in the matrix are the target object clustering tiles.
上述仅为本发明的优选实施例而已,并不对本发明起到任何限制作用。任何所属技术领域的技术人员,在不脱离本发明的技术方案的范围内,对本发明揭露的技术方案和技术内容做任何形式的等同替换或修改等变动,均属未脱离 本发明的技术方案的内容,仍属于本发明的保护范围之内。The foregoing are only preferred embodiments of the present invention, and do not play any restrictive effect on the present invention. Any person skilled in the art, without departing from the scope of the technical solution of the present invention, makes any form of equivalent replacement or modification or other changes to the technical solution and technical content disclosed by the present invention, which does not depart from the technical solution of the present invention. The content still falls within the protection scope of the present invention.

Claims (9)

  1. 一种场景图像中特定目标对象的匹配方法,其特征在于,包括以下步骤:A method for matching a specific target object in a scene image is characterized in that it comprises the following steps:
    第一步、对场景图像进行超像素图块的分割并提取各个超像素图块中的超像素中心属性,所述超像素中心属性包含位置中心和色彩中心;The first step is to segment the scene image into super pixel blocks and extract the super pixel center attributes in each super pixel block, where the super pixel center attributes include a position center and a color center;
    第二步、获得反映各个超像素图块之间相邻关系的邻接矩阵;The second step is to obtain an adjacency matrix reflecting the adjacency relationship between each super pixel block;
    第三步、根据所述邻接矩阵获得反映相邻超像素图块之间相似程度的相似度矩阵,所述相似度包含位置相邻关系和色彩的相似程度;The third step is to obtain a similarity matrix reflecting the degree of similarity between adjacent super-pixel tiles according to the adjacency matrix, where the similarity includes the degree of similarity of the position adjacent relationship and the color;
    第四步、根据所述相似度矩阵对超像素图块完成聚类;The fourth step is to complete clustering of super pixel tiles according to the similarity matrix;
    第五步、在聚类后的场景图像中对特定目标对象进行图块选择与特征值提取;The fifth step is to perform block selection and feature value extraction for specific target objects in the clustered scene image;
    第六步、在场景图像中搜索与目标对象的特征值相近的颜色图块。The sixth step is to search for color tiles that are similar to the feature value of the target object in the scene image.
  2. 根据权利要求1所述的场景图像中特定目标对象的匹配方法,其特征在于,所述超像素中心属性包括如下属性:在图像中心的坐标center(x,y),颜色均值color_info(l,a,b),超像素唯一标识id labels,超像素个数num_pixels。The method for matching a specific target object in a scene image according to claim 1, wherein the superpixel center attribute includes the following attributes: the coordinate center (x, y) at the center of the image, the color average color_info (l, a ,b), the unique identification id labels of super pixels, the number of super pixels num_pixels.
  3. 根据权利要求2所述的场景图像中特定目标对象的匹配方法,其特征在于,所述计算邻接矩阵的具体的算法实现如下:The method for matching a specific target object in a scene image according to claim 2, wherein the specific algorithm for calculating the adjacency matrix is implemented as follows:
    Figure PCTCN2019122673-appb-100001
    Figure PCTCN2019122673-appb-100001
    其中,i,j分别是代表超像素图块序号;Among them, i and j respectively represent the sequence numbers of the super-pixel tiles;
    邻接矩阵E中每个元数e(i,j)满足如下函数关系:Each element e(i,j) in the adjacency matrix E satisfies the following functional relationship:
    Figure PCTCN2019122673-appb-100002
    Figure PCTCN2019122673-appb-100002
    其中,超像素图块自身与自身之间的关系定义为相邻。Among them, the relationship between the super pixel block itself and itself is defined as adjacent.
  4. 根据权利要求3所述的场景图像中特定目标对象的匹配方法,其特征在于,所述计算相似度矩阵的步骤是根据邻接矩阵中超像素图块的相邻关系计算相邻两个超像素的相似度,当相似度必须大于一定阈值时将相应元数值置1,否则置为0,具体的算法实现如下:The method for matching a specific target object in a scene image according to claim 3, wherein the step of calculating the similarity matrix is to calculate the similarity of two adjacent super pixels according to the adjacent relationship of the super pixel blocks in the adjacency matrix. When the similarity must be greater than a certain threshold, the corresponding element value is set to 1, otherwise it is set to 0. The specific algorithm is implemented as follows:
    (1)从CIE Lab色彩空间转换为LθM色彩空间(1) Convert from CIE Lab color space to LθM color space
    θ′=atan2(B,A) θ′∈(-π,π](公式3-1)θ′=atan2(B,A) θ′∈(-π,π] (Formula 3-1)
    Figure PCTCN2019122673-appb-100003
    Figure PCTCN2019122673-appb-100003
    Figure PCTCN2019122673-appb-100004
    Figure PCTCN2019122673-appb-100004
    l=L  l∈[0,100]l=L l∈[0, 100]
    (2)相似度计算(2) Similarity calculation
    Figure PCTCN2019122673-appb-100005
    Figure PCTCN2019122673-appb-100005
    其中,L thth,M th,L th0th0分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值;w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似。 Among them, L th , θ th , M th , L th0 , and θ th0 are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing color and black and white color spaces by modulus length component, usually the value is less than or equal to 2. L i , L j , θ i , θ j , M i , M j are the mean values of super pixel tiles i and j in the LθM color space respectively; w(i,j) is expressed as two super pixel tiles The similarity of, where a value of 1 is similar, and a value of 0 is dissimilar.
  5. 根据权利要求4所述的场景图像中特定目标对象的匹配方法,其特征在于,所述聚类的步骤是利用相似度w(i,j)生成相似度矩阵W,W即为聚类关系图。The method for matching a specific target object in a scene image according to claim 4, wherein the step of clustering is to generate a similarity matrix W using similarity w(i, j), and W is the clustering relationship graph .
  6. 根据权利要求5所述的场景图像中特定目标对象的匹配方法,其特征在于,所述基于相似度矩阵W完成聚类的具体的算法实现包括:将相似度矩阵W转换为三角矩阵的步骤,The method for matching a specific target object in a scene image according to claim 5, wherein the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes the step of converting the similarity matrix W into a triangular matrix,
    相似度矩阵Similarity matrix
    Figure PCTCN2019122673-appb-100006
    Figure PCTCN2019122673-appb-100006
    三角矩阵,将左下角全部置零,Triangular matrix, set all the lower left corners to zero,
    Figure PCTCN2019122673-appb-100007
    Figure PCTCN2019122673-appb-100007
  7. 根据权利要求6所述的场景图像中特定目标对象的匹配方法,其特征在于,所述基于相似度矩阵W完成聚类的具体的算法实现包括:完成聚类的步骤,对三角矩阵执行聚类算法The method for matching a specific target object in a scene image according to claim 6, wherein the implementation of the specific algorithm for completing clustering based on the similarity matrix W includes: completing the step of clustering, performing clustering on the triangular matrix algorithm
    Figure PCTCN2019122673-appb-100008
    Figure PCTCN2019122673-appb-100008
    第一步:first step:
    从矩阵的第n行n列开始,搜索所有n列上为1的数组,如果第n列上为1的数组只有第n行,则a(n,n)=1,否则a(n,n)=0Starting from the nth row and nth column of the matrix, search all the arrays with 1 on the nth column. If the array with 1 on the nth column has only the nth row, then a(n,n)=1, otherwise a(n,n )=0
    公式如下:The formula is as follows:
    Figure PCTCN2019122673-appb-100009
    Figure PCTCN2019122673-appb-100009
    如果in case
    a(n,n)=0a(n,n)=0
    则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给n列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number among the n columns [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero item or calculation algorithm of the column is as follows:
    a(i min,n)=a(i min,n)∪...∪a(n,n) a(i min ,n)=a(i min ,n)∪...∪a(n,n)
    a(i min,n-1)=a(i min,n-1)∪...∪a(n,n-1) a(i min ,n-1)=a(i min ,n-1)∪...∪a(n,n-1)
    ………………………………………………………………………………
    a(i min,i min)=a(i min,i min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,i min )∪...∪a(n,i min )
    赋值运算:Assignment operation:
    a(n,n)=0a(n,n)=0
    本次运算结束;The end of this calculation;
    第二步:The second step:
    从矩阵的第n-1行n-1列开始,搜索所有n-1列上为1的数组,如果第n-1列上为1的数组只有第n-1行,则a(n-1,n-1)=1否则a(n-1,n-1)=0Starting from the n-1th row and n-1 column of the matrix, search all the arrays with 1 in the n-1 column. If the array with 1 in the n-1th column has only the n-1th row, then a(n-1 , N-1)=1 otherwise a(n-1, n-1)=0
    公式如下:The formula is as follows:
    Figure PCTCN2019122673-appb-100010
    Figure PCTCN2019122673-appb-100010
    如果in case
    a(n-1,n-1)=0a(n-1, n-1)=0
    则对这些数组按照“行降序”的顺序将其各列(1、2、3……n-1)进行逻辑或运算,并将结果赋值给n-1列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列的非零项或运算算法如下: Then these arrays are logically ORed on each column (1, 2, 3...n-1) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in the n-1 column [0 ,0,……a(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero items or calculation algorithms of the columns are as follows:
    a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
    a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
    ………………………………………………………………………………
    a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
    赋值运算:Assignment operation:
    a(n-1,n~n-1)=0a(n-1, n~n-1)=0
    本次运算结束;The end of this calculation;
    第三步:third step:
    以此类推,从矩阵的第i行i列开始,搜索所有i列上为1的数组,如果第i列上为1的数组只有第i行,则a(i,i)=1否则a(i,i)=0By analogy, starting from the i-th row and i-column of the matrix, search for all arrays with 1 on the i-th column. If the array with 1 on the i-th column has only the i-th row, then a(i, i) = 1 otherwise a( i, i) = 0
    公式如下:The formula is as follows:
    Figure PCTCN2019122673-appb-100011
    Figure PCTCN2019122673-appb-100011
    如果in case
    a(i,i)=0a(i,i)=0
    则对这些数组按照“行降序”的顺序将其各列(1、2、3……n)进行逻辑或运算,并将结果赋值给i列中行号最小的非零数组[0,0,……a(i min,i min),……,a(i min,n-1),a(i min,n)]中;列中的非零项或运算算法如下: Then these arrays are logically ORed for each column (1, 2, 3...n) in the "row descending order" order, and the result is assigned to the non-zero array with the smallest row number in column i [0,0,... …A(i min ,i min ),……,a(i min ,n-1),a(i min ,n)]; the non-zero term or calculation algorithm in the column is as follows:
    a(i min,j n)=a(i min,j n)∪...∪a(n,j n) a(i min ,j n )=a(i min ,j n )∪...∪a(n,j n )
    a(i min,j n-1)=a(i min,j n-1)∪...∪a(n,j n-1) a(i min ,j n-1 )=a(i min ,j n-1 )∪...∪a(n,j n-1 )
    ………………………………………………………………………………
    a(i min,i min)=a(i min,j min)∪...∪a(n,i min) a(i min ,i min )=a(i min ,j min )∪...∪a(n,i min )
    赋值运算:Assignment operation:
    a(i,n~n-i)=0a(i, n~n-i)=0
    ,本次运算结束;, This calculation is over;
    第四步:the fourth step:
    根据以上的算法将三角矩阵的每一行都遍历一遍,将得到如下类似矩阵:According to the above algorithm, each row of the triangular matrix is traversed once, and the following similar matrix will be obtained:
    Figure PCTCN2019122673-appb-100012
    Figure PCTCN2019122673-appb-100012
    则矩阵中所有非零的行数组即为聚类图块的数组。Then all the non-zero row arrays in the matrix are the arrays of clustering tiles.
  8. 根据权利要求7所述的场景图像中特定目标对象的匹配方法,其特征在于,所述场景图像中特定目标对象的图像选择与特征提取是指根据图像聚类结果在场景中选择特定目标对象的图像块并提取出对应图块中L、θ、M的取值。The method for matching a specific target object in a scene image according to claim 7, wherein the image selection and feature extraction of the specific target object in the scene image refers to the selection of the specific target object in the scene according to the result of image clustering. Image block and extract the values of L, θ, M in the corresponding block.
  9. 根据权利要求8所述的场景图像中特定目标对象的匹配方法,其特征在于,所述搜索与目标对象图块的特征值相近的颜色图块的具体方法如下:The method for matching a specific target object in a scene image according to claim 8, wherein the specific method for searching for a color tile similar to the feature value of the target object tile is as follows:
    Figure PCTCN2019122673-appb-100013
    Figure PCTCN2019122673-appb-100013
    其中L thth,M th分别为LθM色彩空间中三个分量的阈值,M Cth为以模长分量区分彩色和黑白颜色空间的阈值,通常取值为小于等于2,L i,L jij,M i,M j分别为超像素图块i,j在LθM色彩空间中的均值;其中i表示样本图像中已选定的目标对象的图块号,j表示搜索图像的图块号;w(i,j)表示为两个超像素图块的相似度,其中取值为1则为相似,取值为0则为不相似; Among them, L th , θ th , and M th are the threshold values of the three components in the LθM color space, and M Cth is the threshold value for distinguishing between color and black and white color spaces by modulus length components. Usually the value is less than or equal to 2, Li , L jij ,M i ,M j are the mean values of super pixel tiles i and j in the LθM color space, respectively; where i represents the number of the selected target object in the sample image, and j represents the search image The number of the block; w(i,j) represents the similarity of two superpixel blocks, where a value of 1 is similar, and a value of 0 is dissimilar;
    对搜索图像的图块进行如下操作:Perform the following operations on the tiles of the search image:
    如果w(i,j)=1,则保留搜索图像图块中像素的原始数值不变;If w(i,j)=1, keep the original value of the pixel in the search image block unchanged;
    如果w(i,j)=0,则将搜索图像图块中像素的值设置为不在色彩空间的数值如-1,则该图块将不参与后面的运算;If w(i,j)=0, then the value of the pixel in the search image block is set to a value that is not in the color space, such as -1, then the block will not participate in the subsequent calculations;
    将搜索到的图块构建邻接矩阵:Construct an adjacency matrix with the searched tiles:
    Figure PCTCN2019122673-appb-100014
    Figure PCTCN2019122673-appb-100014
    三角矩阵,将左下角全部置零:Triangular matrix, set all the lower left corners to zero:
    Figure PCTCN2019122673-appb-100015
    Figure PCTCN2019122673-appb-100015
    对三角矩阵采用与权利要求7同样的聚类的方法完成对目标对象的聚类;从而实现在不同场景图像中对目标对象的匹配:Using the same clustering method as in claim 7 for the triangular matrix to complete the clustering of the target object; thereby achieving the matching of the target object in different scene images:
    Figure PCTCN2019122673-appb-100016
    Figure PCTCN2019122673-appb-100016
    矩阵中所有非零的行数组即为目标对象聚类图块。All non-zero row arrays in the matrix are the target object clustering tiles.
PCT/CN2019/122673 2019-11-01 2019-12-03 Method for matching specific target object in scene image WO2021082168A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201911060675.4 2019-11-01
CN201911060675.4A CN112766291B (en) 2019-11-01 2019-11-01 Matching method for specific target object in scene image

Publications (1)

Publication Number Publication Date
WO2021082168A1 true WO2021082168A1 (en) 2021-05-06

Family

ID=75692124

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/122673 WO2021082168A1 (en) 2019-11-01 2019-12-03 Method for matching specific target object in scene image

Country Status (2)

Country Link
CN (1) CN112766291B (en)
WO (1) WO2021082168A1 (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113569513A (en) * 2021-06-30 2021-10-29 广东国腾量子科技有限公司 Waveguide mode-based on-chip multidimensional logic gate design method
CN113838061A (en) * 2021-07-28 2021-12-24 中科云谷科技有限公司 Method and device for image annotation and storage medium
CN114998792A (en) * 2022-05-30 2022-09-02 中用科技有限公司 Safety monitoring method with AI network camera
CN115457050A (en) * 2022-08-29 2022-12-09 武汉纺织大学 Super-pixel-based ancient mural image optimization segmentation method and system
CN116128916A (en) * 2023-04-13 2023-05-16 中国科学院国家空间科学中心 Infrared dim target enhancement method based on spatial energy flow contrast
CN116704249A (en) * 2023-06-07 2023-09-05 唯思电子商务(深圳)有限公司 Similar background image classification method based on visual large model

Families Citing this family (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN117173175B (en) * 2023-11-02 2024-02-09 湖南格尔智慧科技有限公司 Image similarity detection method based on super pixels
CN117746260B (en) * 2024-02-20 2024-05-31 山东省国土测绘院 Remote sensing data intelligent analysis method and system

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103413316A (en) * 2013-08-24 2013-11-27 西安电子科技大学 SAR image segmentation method based on superpixels and optimizing strategy
CN107341812A (en) * 2017-07-04 2017-11-10 太原理工大学 A kind of sequence Lung neoplasm image partition method based on super-pixel and Density Clustering
US20180005039A1 (en) * 2015-01-30 2018-01-04 Thomson Licensing Method and apparatus for generating an initial superpixel label map for an image
CN109522908A (en) * 2018-11-16 2019-03-26 董静 Image significance detection method based on area label fusion

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105118049A (en) * 2015-07-22 2015-12-02 东南大学 Image segmentation method based on super pixel clustering
CN110111338B (en) * 2019-04-24 2023-03-31 广东技术师范大学 Visual tracking method based on superpixel space-time saliency segmentation

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103413316A (en) * 2013-08-24 2013-11-27 西安电子科技大学 SAR image segmentation method based on superpixels and optimizing strategy
US20180005039A1 (en) * 2015-01-30 2018-01-04 Thomson Licensing Method and apparatus for generating an initial superpixel label map for an image
CN107341812A (en) * 2017-07-04 2017-11-10 太原理工大学 A kind of sequence Lung neoplasm image partition method based on super-pixel and Density Clustering
CN109522908A (en) * 2018-11-16 2019-03-26 董静 Image significance detection method based on area label fusion

Cited By (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113569513A (en) * 2021-06-30 2021-10-29 广东国腾量子科技有限公司 Waveguide mode-based on-chip multidimensional logic gate design method
CN113569513B (en) * 2021-06-30 2024-04-16 广东国腾量子科技有限公司 On-chip multidimensional logic gate design method based on waveguide mode
CN113838061A (en) * 2021-07-28 2021-12-24 中科云谷科技有限公司 Method and device for image annotation and storage medium
CN114998792A (en) * 2022-05-30 2022-09-02 中用科技有限公司 Safety monitoring method with AI network camera
CN114998792B (en) * 2022-05-30 2024-05-14 中用科技有限公司 Security monitoring method with AI network camera
CN115457050A (en) * 2022-08-29 2022-12-09 武汉纺织大学 Super-pixel-based ancient mural image optimization segmentation method and system
CN115457050B (en) * 2022-08-29 2023-06-27 武汉纺织大学 Method and system for optimizing and dividing ancient mural image based on superpixels
CN116128916A (en) * 2023-04-13 2023-05-16 中国科学院国家空间科学中心 Infrared dim target enhancement method based on spatial energy flow contrast
CN116704249A (en) * 2023-06-07 2023-09-05 唯思电子商务(深圳)有限公司 Similar background image classification method based on visual large model
CN116704249B (en) * 2023-06-07 2024-05-24 唯思电子商务(深圳)有限公司 Similar background image classification method based on visual large model

Also Published As

Publication number Publication date
CN112766291A (en) 2021-05-07
CN112766291B (en) 2024-03-22

Similar Documents

Publication Publication Date Title
WO2021082168A1 (en) Method for matching specific target object in scene image
Ge et al. Detecting masked faces in the wild with lle-cnns
WO2019134327A1 (en) Facial expression recognition feature extraction method employing edge detection and sift
WO2018107760A1 (en) Collaborative deep network model method for pedestrian detection
CN105528575B (en) Sky detection method based on Context Reasoning
CN102194108B (en) Smile face expression recognition method based on clustering linear discriminant analysis of feature selection
CN105528794A (en) Moving object detection method based on Gaussian mixture model and superpixel segmentation
CN107688830B (en) Generation method of vision information correlation layer for case serial-parallel
CN104850822B (en) Leaf identification method under simple background based on multi-feature fusion
Yadav et al. Survey on content-based image retrieval and texture analysis with applications
Zohourian et al. Superpixel-based Road Segmentation for Real-time Systems using CNN.
Ahn et al. Face and hair region labeling using semi-supervised spectral clustering-based multiple segmentations
CN106909884A (en) A kind of hand region detection method and device based on hierarchy and deformable part sub-model
Hirano et al. Industry and object recognition: Applications, applied research and challenges
CN103324753A (en) Image retrieval method based on symbiotic sparse histogram
Liao et al. Deep-learning-based object-level contour detection with CCG and CRF optimization
Gao et al. Occluded person re-identification based on feature fusion and sparse reconstruction
Hu et al. Fast face detection based on skin color segmentation using single chrominance Cr
Deshmukh et al. Real-time traffic sign recognition system based on colour image segmentation
AU2011265494A1 (en) Kernalized contextual feature
Li et al. Salient traffic sign recognition based on sparse representation of visual perception
Jin et al. A method for traffic signs detection, tracking and recognition
Li et al. Combining shape and appearance for automatic pedestrian segmentation
Sima et al. Texture superpixels merging by color-texture histograms for color image segmentation
Chen et al. Outdoor scene classification using labeled segments

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19950508

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19950508

Country of ref document: EP

Kind code of ref document: A1

122 Ep: pct application non-entry in european phase

Ref document number: 19950508

Country of ref document: EP

Kind code of ref document: A1

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205N DATED 19.05.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 19950508

Country of ref document: EP

Kind code of ref document: A1