CN103336835B - Image retrieval method based on weight color-sift characteristic dictionary - Google Patents

Image retrieval method based on weight color-sift characteristic dictionary Download PDF

Info

Publication number
CN103336835B
CN103336835B CN201310294385.2A CN201310294385A CN103336835B CN 103336835 B CN103336835 B CN 103336835B CN 201310294385 A CN201310294385 A CN 201310294385A CN 103336835 B CN103336835 B CN 103336835B
Authority
CN
China
Prior art keywords
image
color
retrieved
edge
weight
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related
Application number
CN201310294385.2A
Other languages
Chinese (zh)
Other versions
CN103336835A (en
Inventor
李平舟
刘燕
刘宪龙
杨国瑞
孙雪萍
赵楠
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN201310294385.2A priority Critical patent/CN103336835B/en
Publication of CN103336835A publication Critical patent/CN103336835A/en
Application granted granted Critical
Publication of CN103336835B publication Critical patent/CN103336835B/en
Expired - Fee Related legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Landscapes

  • Image Analysis (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)

Abstract

一种基于权值color‑sift特征字典的图像检索方法,其包括,在待检索图像中随机选取训练图像,提取所述训练图像的边缘,并提取所有训练图像边缘点的color‑sift特征,并以所述color‑sift特征构建特征字典;输入要检索的图像并提取检索图像和待检索图像边缘点的color‑sift特征,基于所述特征字典对检索图像和待检索图像提取权值直方图特征;把检索图像和数据库中待检索图像进行基于权值直方图特征的相似性匹配;检测是否遍历所有数据库中的全部待检索图像,若是,则按照相似性匹配结果,显示图像检索结果,若否,则重新进行相似性匹配。提高了应用于大型图像数据库检索时的准确率和回调率。该算法具有尺度、平移、旋转不变性。其具有局部性、殊性、多量性和高效性等特征。

An image retrieval method based on a weight color-sift feature dictionary, which includes randomly selecting a training image in an image to be retrieved, extracting the edge of the training image, and extracting the color-sift feature of all training image edge points, and Build a feature dictionary with the color-sift feature; input the image to be retrieved and extract the color-sift feature of the edge point of the search image and the image to be retrieved, and extract the weight histogram feature of the search image and the image to be retrieved based on the feature dictionary ; Perform similarity matching based on the weight histogram feature between the retrieved image and the image to be retrieved in the database; check whether to traverse all the images to be retrieved in all databases, if so, display the image retrieval result according to the similarity matching result, if not , re-do similarity matching. Improved accuracy and callback rates when applied to retrieval of large image databases. The algorithm is invariant to scale, translation and rotation. It has the characteristics of locality, particularity, quantity and efficiency.

Description

基于权值color-sift特征字典的图像检索方法Image retrieval method based on weighted color-sift feature dictionary

技术领域technical field

本发明属于图像检索技术领域,具体是一种基于权值color-sift特征字典的图像检索方法,立足于图像内容,基于图像特征提取以实现对图像进行分析和检索的过程,允许用户输入一张或多张图片,以查找具有相同或相似内容的其他图片。The invention belongs to the technical field of image retrieval, and specifically relates to an image retrieval method based on a weight color-sift feature dictionary, which is based on image content and image feature extraction to realize the process of image analysis and retrieval, and allows users to input a or multiple images to find other images with the same or similar content.

背景技术Background technique

图像是对客观对象的一种相似性的、生动性的描述或写真。或者说图像是客观对象的一种表示,它包含了被描述对象的有关信息。它是人们最主要的信息源。据统计,一个人获取的信息大约有75%来自视觉。俗话说“百闻不如一见”,“一目了然”,都反映了图像在信息传递中的独特效果。如何快速准确地提取图像内容是图像检索最关键的一步。SIFT是David G Lowe2004年总结不变量技术的特征检测方法,提出的一种对尺度空间、图像缩放、旋转甚至仿射不变的图像局部特征描述算子。所谓基于图像内容检索,即从图像库中查找含有特定目标的图像,也包括从连续的视频图像中检索含有特定目标的视频片段。它区别于传统的图像检索手段,本发明提出了一种基于权值color-sift特征字典的图像检索方法,融合了权值color-sift特征字典技术,从而可以提供更有效的检索手段。可应用于数字图书馆、医疗诊断、图像分类、WEB相关应用、公共安全和犯罪调查等等An image is a similar and vivid description or portrait of an objective object. In other words, an image is a representation of an objective object, which contains information about the object being described. It is people's primary source of information. According to statistics, about 75% of the information a person obtains comes from vision. As the saying goes, "seeing is better than hearing a hundred times" and "clear at a glance", both reflect the unique effect of images in information transmission. How to quickly and accurately extract image content is the most critical step in image retrieval. SIFT is a feature detection method that summarizes invariant technology in 2004 by David G Lowe, and proposes an image local feature description operator that is invariant to scale space, image scaling, rotation, and even affine. The so-called image content-based retrieval refers to searching for images containing specific objects from the image library, and also includes retrieving video clips containing specific objects from continuous video images. It is different from the traditional image retrieval method. The present invention proposes an image retrieval method based on the weight color-sift feature dictionary, which integrates the weight color-sift feature dictionary technology, thereby providing a more effective retrieval method. Can be applied to digital libraries, medical diagnosis, image classification, WEB related applications, public safety and criminal investigation, etc.

清华大学大学提出的专利申请“一种基于草图特征提取的图像检索方法”(专利申请号201110196051.2,公开号201110196051.2)公开了一种基于草图特征提取的图像检索方法,涉及图像检索领域。所述方法包括步骤:提取训练特征向量,得到特征词典;提取输入特征向量,得到输入特征向量集,对特征词典进行计数操作,得到输入特征频率向量,进而得到兴趣特征词和非兴趣特征词;提取检索特征向量,得到检索特征向量集,进而得到检索特征频率向量;进而得到兴趣检索特征频率向量、非兴趣检索特征频率向量、兴趣输入特征频率向量和非兴趣输入特征频率向量;进而计算输入草图与各个检索草图的相似度,输出检索结果。该方法具有良好的用户交互性提高了图像检索的效率和准确度,但是特征维数较大,考虑了兴趣和非兴趣两种特征,导致在应用于大型数据库时检索效率较低,速度较慢。The patent application "An Image Retrieval Method Based on Sketch Feature Extraction" (Patent Application No. 201110196051.2, Publication No. 201110196051.2) filed by Tsinghua University discloses an image retrieval method based on sketch feature extraction, which involves the field of image retrieval. The method comprises the steps of: extracting a training feature vector to obtain a feature dictionary; extracting an input feature vector to obtain an input feature vector set, performing a counting operation on the feature dictionary to obtain an input feature frequency vector, and then obtaining an interesting feature word and a non-interest feature word; Extract the retrieval feature vector, obtain the retrieval feature vector set, and then obtain the retrieval feature frequency vector; then obtain the interesting retrieval feature frequency vector, the non-interest retrieval feature frequency vector, the interesting input feature frequency vector and the non-interest input feature frequency vector; and then calculate the input sketch The similarity with each retrieval sketch is outputted as a retrieval result. This method has good user interaction and improves the efficiency and accuracy of image retrieval, but the feature dimension is large, and two features of interest and non-interest are considered, resulting in low retrieval efficiency and slow speed when applied to large databases .

中国传媒大学提出的专利申请“一种结合用户评价与标注的交互式图像检索方法”(专利申请号201310128036.3,公开号CN103164539A)公开了一种结合用户评价和标注的交互式图像检索方法,属于多媒体信息检索领域。该方法利用了基于图像的物理特征和文本相结合的综合检索方法,在检索过程中,允许用户对查询图像进行文本信息描述,或者选择系统提供的关键字,通过对检索结果进行“满意”或“不满意”的相关评价,图像检索系统自动对用户标记的相关满意图像进行文本标记,形成高层语义信息;随着用户的不断使用,该系统会生成丰富的语义信息数据库。考虑到不同用户对同一图片,同一用户不同时间对同一图片文本标注的差异,本发明在生成语义信息数据库的过程中结合了用户的可信度。进行检索时,对存在语义信息的查询图像采用基于特征和文本相结合的综合检索方式进行检索,提高了检索结果的准确度。该方法虽然结合了用户评价和标注的交互式图像检索方式,获取高层语义信息,提高了实时检索的准确率,但是在应用于大型图像数据库时,由于各类图像的相似度量繁杂,处理人工繁杂的标注提高了计算复杂度,降低了图像的检索效率,导致返回检索结果集的回调率和检索率不高。The patent application "An interactive image retrieval method combining user evaluation and annotation" (patent application number 201310128036.3, publication number CN103164539A) filed by Communication University of China discloses an interactive image retrieval method combining user evaluation and annotation, which belongs to multimedia the field of information retrieval. This method uses a comprehensive retrieval method based on the combination of physical characteristics of images and text. During the retrieval process, users are allowed to describe the query image in text, or select keywords provided by the system. For the relevant evaluation of "unsatisfied", the image retrieval system automatically marks the relevant satisfactory images marked by the user to form high-level semantic information; with the continuous use of the user, the system will generate a rich semantic information database. Considering the differences in text annotations of the same picture by different users and by the same user at different times, the present invention combines the user's credibility in the process of generating the semantic information database. When searching, the query image with semantic information is retrieved using a comprehensive retrieval method based on the combination of features and text, which improves the accuracy of the retrieval results. Although this method combines the interactive image retrieval method of user evaluation and annotation to obtain high-level semantic information and improve the accuracy of real-time retrieval, when it is applied to a large image database, due to the complicated similarity metrics of various images, the processing is complicated. The labeling increases the computational complexity and reduces the retrieval efficiency of the image, resulting in a low callback rate and retrieval rate of the returned retrieval result set.

发明内容Contents of the invention

本发明针对上述现有技术的不足,提出一种基于权值color-sift特征字典的图像检索方法,提高了应用大型数据库时检索的效率、速度和回调率。The present invention aims at the deficiencies of the above-mentioned prior art, and proposes an image retrieval method based on a weighted color-sift feature dictionary, which improves retrieval efficiency, speed and callback rate when large databases are used.

一种基于权值color-sift特征字典的图像检索方法,其包括,An image retrieval method based on weight color-sift feature dictionary, which includes,

在待检索图像中随机选取训练图像,提取所述训练图像的边缘,并提取所有训练图像边缘点的color-sift特征,并以所述color-sift特征构建特征字典;Randomly select a training image in the image to be retrieved, extract the edge of the training image, and extract the color-sift feature of all training image edge points, and construct a feature dictionary with the color-sift feature;

输入要检索的图像并提取检索图像和待检索图像边缘点的color-sift特征,基于所述特征字典对检索图像和待检索图像提取权值直方图特征;把检索图像和数据库中待检索图像进行基于权值直方图特征的相似性匹配;Input the image to be retrieved and extract the color-sift feature of the image to be retrieved and the edge point of the image to be retrieved, based on the feature dictionary to extract the weight histogram feature of the image to be retrieved and the image to be retrieved; Similarity matching based on weight histogram features;

检测是否遍历所有数据库中的全部待检索图像,若是,则按照相似性匹配结果,显示图像检索结果,若否,则重新进行相似性匹配。Detect whether to traverse all the images to be retrieved in all databases, if yes, display the image retrieval results according to the similarity matching results, if not, perform similarity matching again.

在上述技术方案的基础上,在待检索图像中随机选取训练图像包括在待检索图像数据库中对于每类图像随机选取l张图像组成训练图像数据库。On the basis of the above technical solution, randomly selecting training images from the images to be retrieved includes randomly selecting one image for each type of image in the image database to be retrieved to form the training image database.

在上述技术方案的基础上,提取所述训练图像的边缘包括:提取所述训练图像的边缘包括在训练图像数据库中随机选取一张训练图像做灰度变换,并通过方向可调滤波器进行处理,选取二维高斯函数为滤波器核函数,得到每个像素点2L个方向上的能量函数Wσ(x,y,θ),并经过阈值判断提取图像的边缘显著像素点,其中L表示方向的个数,x和y表示像素点的坐标值,θ为方向的值,范围是0~2π,间隔为π/L。On the basis of the above technical solution, extracting the edge of the training image includes: extracting the edge of the training image includes randomly selecting a training image in the training image database for gray scale transformation, and processing it through a direction adjustable filter , select the two-dimensional Gaussian function as the filter kernel function, get the energy function W σ (x,y,θ) of each pixel point in 2L directions, and extract the significant pixel points of the edge of the image through threshold judgment, where L represents the direction The number of , x and y represent the coordinate value of the pixel point, θ is the value of the direction, the range is 0~2π, and the interval is π/L.

在上述技术方案的基础上,提取所有训练图像边缘点的color-sift特征包括:根据所选取训练图像的边缘显著像素点对原训练彩色图像分别在红色-R、绿色-G和蓝色-B三个通道提取color-sift特征,得到该图像每一个边缘显著像素点的红色-R、绿色-G和蓝色-B三个通道的color-sift特征decr(e)、decg(e)和decb(e),e表示该图像的第e个边缘点,e=1,2,...,E,E为该图像所有边缘显著像素点的总数。On the basis of the above technical solution, extracting the color-sift feature of all training image edge points includes: according to the edge salient pixels of the selected training image, the original training color image is respectively divided into red-R, green-G and blue-B The color-sift features are extracted by three channels, and the color-sift features dec r (e), dec g (e) of the red-R, green-G and blue-B three channels of the red-R, green-G and blue-B channels of each prominent pixel point of the image are obtained and dec b (e), e represents the e-th edge point of the image, e=1, 2,..., E, E is the total number of all edge salient pixels of the image.

在上述技术方案的基础上,所述提取所有训练图像边缘点的color-sift特征包括:对训练图像数据库中每一幅训练图像所有边缘显著像素点的color-sift特征提取,遍历训练图像数据库中的所有图像,其特征在三个颜色通道内依次为decr,m(em)、decg,m(em)和decb,m(em),m=1,2,...,M,M为训练图像数据库大小,em表示第m张训练图像的第em个边缘像素点,em=1,2,...,Em,Em为第m张训练图像所有边缘显著像素点的总数。On the basis of the above technical solution, the extraction of the color-sift feature of all training image edge points includes: extracting the color-sift feature of all edge salient pixels of each training image in the training image database, traversing the training image database All images of , whose features are sequentially dec r,m (e m ), dec g,m (e m ) and dec b,m (e m ) in the three color channels, m=1,2,... , M, M is the size of the training image database, em represents the em edge pixel of the mth training image, em = 1,2,...,E m , E m is all the mth training images The total number of edge salient pixels.

在上述技术方案的基础上,构建特征字典步骤包括对全部训练图像的所有边缘显著像素点的红色-R通道绿色-G通道和蓝色-B通道的color-sift特征decr,m(em),decg,m(em)和decb,m(em),通过K-means聚类计算,取K个聚类中心得到红色-R通道、绿色-G通道和蓝色-B通道的w行、K列的二维特征字典codr、codg和codb,其中K为K-means聚类中心的个数,即字典的大小。On the basis of the above-mentioned technical scheme, the step of constructing a feature dictionary includes the color-sift feature dec r,m (e m ), dec g,m (e m ) and dec b,m (e m ), through K-means clustering calculation, take K cluster centers to get red-R channel, green-G channel and blue-B channel The two-dimensional feature dictionaries cod r , cod g and cod b of w rows and K columns, where K is the number of K-means cluster centers, that is, the size of the dictionary.

在上述技术方案的基础上,所述权值直方图特征X包含了红色-R、绿色-G和蓝色-B三通道的权值直方图特征,把检索图像和待检索图像进行基于权值直方图特征的相似性匹配,计算2范数相似性距离得到Di(X,X′i)。On the basis of the above technical solution, the weight histogram feature X includes the weight histogram features of the three channels of red-R, green-G and blue-B, and the retrieval image and the image to be retrieved are calculated based on the weight Similarity matching of histogram features, calculate 2-norm similarity distance to get D i (X,X′ i ).

在上述技术方案的基础上,对检索图像编码计算得到检索图像的基于color-sift特征字典的权值直方图特征,包括以下步骤:On the basis of the above-mentioned technical solution, the weight histogram feature based on the color-sift feature dictionary of the retrieved image is obtained by encoding and calculating the retrieved image, including the following steps:

1)检索图像在红色通道R上所有边缘显著像素点的color-sift特征secr(o),计算secr(o)对应于红色通道二维特征字典codr的K个聚类中心的欧式距离,选择距离值最小时的聚类中心作为该边缘点的所属中心,对所有边缘显著点进行一阶距统计计算得到检索图像在红色通道对应二维特征字典codr的频率直方图hisr1) Retrieve the color-sift feature sec r (o) of all edge significant pixels on the red channel R of the image, and calculate sec r (o) corresponding to the Euclidean distance of the K cluster centers of the red channel two-dimensional feature dictionary cod r , select the clustering center with the smallest distance value as the center of the edge point, and perform first-order distance statistical calculation on all edge significant points to obtain the frequency histogram histogram corresponding to the two-dimensional feature dictionary cod r of the retrieved image in the red channel ;

2)假设落在第k个聚类中心qk个边缘显著点,对聚类到第k个聚类中心的qk个边缘显著点计算该显著点对于该聚类中心的离心权值li(k),计算该聚类中心所有显著点最大的离心权值得到该聚类中心的权值向量α(k),对检索图像的频率直方图hisr和对应权值向量α(k)做矩阵点乘运算得到检索图像的权值向量hstr,即对应元素相乘,hstr为一K维列向量,k代表第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小;2) Assuming that q k edge significant points fall in the k -th cluster center, calculate the centrifugal weight li( k), calculate the maximum centrifugal weight of all significant points in the cluster center to obtain the weight vector α(k) of the cluster center, and make a matrix for the frequency histogram his r of the retrieved image and the corresponding weight vector α(k) The dot multiplication operation obtains the weight vector hst r of the retrieved image, that is, the multiplication of corresponding elements, hst r is a K-dimensional column vector, k represents the k-th cluster center, and the value is 1, 2,...,K, K represents the number of cluster centers, that is, the size of the dictionary;

3)在绿色通道G、蓝色通道B做同样于红色通道R的计算,最终得到图像绿色通道G的权值向量hstg和蓝色通道B的权值向量hstb,对三个通道内的权值向量进行整合计算得到检索图像的基于color-sift特征字典的权值直方图特征X。3) Do the same calculation as the red channel R in the green channel G and blue channel B, and finally get the weight vector hst g of the image green channel G and the weight vector hst b of the blue channel B, for the three channels The weight vector is integrated and calculated to obtain the weight histogram feature X based on the color-sift feature dictionary of the retrieved image.

在上述技术方案的基础上,所述的对聚类到第k个聚类中心的qk个边缘显著点计算该显著点对于该聚类中心的离心权值li(k),计算该聚类中心所有显著点最大的离心权值得到该聚类中心的权值向量α(k),采用如下公式计算:On the basis of the above technical solution, the q k edge salient points clustered to the kth cluster center are calculated to calculate the centrifugal weight li(k) of the salient point for the cluster center, and the cluster The weight vector α(k) of the cluster center is obtained by the largest centrifugal weight of all the significant points in the center, which is calculated by the following formula:

lili (( uu ,, kk )) == 11 ΣΣ vv == 1,21,2 ,, .. .. .. ,, KK || || secsec rr (( uu )) -- codcod (( kk )) || || 22 22 || || secsec rr (( uu )) -- codcod (( vv )) || || 22 22 ,,

α(k)=max(li(u,k)),α(k)=max(li(u,k)),

其中,u表示落在第k个聚类中心第u个边缘显著点,取值为1,2,...,qk,k表示第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小,qk表示落在第k个聚类中心的边缘显著点的总数。Among them, u represents the u-th edge significant point falling in the k-th cluster center, and the value is 1, 2,...,q k , k represents the k-th cluster center, and the value is 1, 2,. .., K, K represents the number of cluster centers, that is, the size of the dictionary, and q k represents the total number of edge significant points falling on the kth cluster center.

相对于现有技术,本发明使用方向可调滤波器进行边缘提取,可以有效地判断出每个像素点边缘主方向的指向,然后通过阈值判定就可以快速准确的提取出图像的边缘像素点信息,通过提取得到的图像边缘像素点信息可以快速准确的进行下一步的特征提取,提高了应用于实时人机交互和大型图像数据库时检索的速度和准确性。采用了提取边缘方向像素点color-sift特征和编码字典相结合的检索策略,提取边缘显著像素点的color-sift特征并通过基于权值color-sift特征构建的编码字典计算权值直方图特征,对图像的表示更加具有典型性,可以更加有效的表示图像的特征差异性,在应用到检索过程中时,提高了应用于大型图像数据库检索时的准确率和回调率。采用了基于RGB三个通道的color-sift特征字典的图像检索方法,它是一种多尺度的图像检索算法,将一幅图像转化为多个特征的集合,再通过计算两幅图像特征向量间的欧氏距离进行比较得出结果进而实现图像检索功能.实验结果说明该算法具有尺度、平移、旋转不变性,可以进行良好应用。它同时是一种对尺度空间,图像缩放,旋转图像局部特征描述算子。其具有局部性、殊性、多量性和高效性等特征。sift特征提取算法能够处理两幅图像之间发生平移、旋转、仿射变换情况下的匹配问题,提高了图像检索的准确率和回调率。Compared with the prior art, the present invention uses a direction-adjustable filter for edge extraction, which can effectively determine the direction of the main direction of each pixel edge, and then quickly and accurately extract the edge pixel information of the image through threshold determination , the image edge pixel information obtained by extraction can quickly and accurately perform feature extraction in the next step, which improves the speed and accuracy of retrieval when applied to real-time human-computer interaction and large image databases. A retrieval strategy combining the color-sift feature of pixel points in the edge direction and the coding dictionary is adopted to extract the color-sift feature of the significant pixel points on the edge and calculate the weight histogram feature through the coding dictionary constructed based on the weight color-sift feature. The representation of images is more typical, and can more effectively represent the feature differences of images. When applied to the retrieval process, it improves the accuracy and callback rate when applied to large image database retrieval. The image retrieval method based on the color-sift feature dictionary of RGB three channels is adopted. It is a multi-scale image retrieval algorithm, which converts an image into a collection of multiple features, and then calculates the distance between the two image feature vectors. The Euclidean distance is compared to get the results and then realize the image retrieval function. The experimental results show that the algorithm has scale, translation and rotation invariance, and can be well applied. It is also a local feature description operator for scale space, image scaling, and rotating images. It has the characteristics of locality, particularity, quantity and efficiency. The sift feature extraction algorithm can deal with the matching problem in the case of translation, rotation and affine transformation between two images, which improves the accuracy and callback rate of image retrieval.

附图说明Description of drawings

图1为本发明的流程图。Fig. 1 is a flowchart of the present invention.

具体实施措施Specific implementation measures

下面结合附图对发明做进一步描述。The invention will be further described below in conjunction with the accompanying drawings.

实施例1Example 1

本发明的基于权值color-sift特征字典的图像检索方法的实现参照图1,给出如下具体实施例:The realization of the image retrieval method based on the weight color-sift feature dictionary of the present invention is with reference to Fig. 1, provides following specific embodiment:

步骤1:在待检索图像数据库中对于每类图像随机选取l张图像组成训练图像数据库,本实例使用Corel-1000图像数据库,需要在Corel-1000图像数据库中检索出同类型的图像,图像库包括10类图像,每一类包括100张图像,本例中l取值为10,一共10类,总共选取100张训练图像。Step 1: Randomly select one image for each type of image in the image database to be retrieved to form the training image database. This example uses the Corel-1000 image database, and it is necessary to retrieve the same type of image in the Corel-1000 image database. The image database includes There are 10 categories of images, and each category includes 100 images. In this example, the value of l is 10, a total of 10 categories, and a total of 100 training images are selected.

步骤2:在训练图像数据库中随机选取一张训练图像做灰度变换,通过方向可调滤波器进行处理,选取二维高斯函数为滤波器核函数,选取合适的滤波器滑动窗口大小,得到每个像素点2L个方向上的能量函数Wσ(x,y,θ),经过阈值判断提取图像的边缘显著像素点,L表示方向的个数,L取值为6,x和y表示像素点的坐标值,σ为滤波器尺度参数,σ取值为1,θ为方向的值,范围是0~2π,间隔为π/L,本例中θ取为0,π/6,...,11π/6,2π。Step 2: Randomly select a training image in the training image database for grayscale transformation, process it through a direction-tunable filter, select a two-dimensional Gaussian function as the filter kernel function, and select an appropriate filter sliding window size to obtain each The energy function W σ (x, y, θ) of each pixel point in 2L directions, through the threshold judgment to extract the significant pixel points of the edge of the image, L represents the number of directions, and the value of L is 6, x and y represent the pixel points σ is the filter scale parameter, σ is 1, θ is the value of the direction, the range is 0~2π, and the interval is π/L. In this example, θ is 0, π/6,... ,11π/6,2π.

步骤3:根据所选取训练图像的边缘显著像素点对原训练彩色图像分别在红色-R、绿色-G和蓝色-B三个通道提取color-sift特征,得到该图像每一个边缘显著像素点的红色-R、绿色-G和蓝色-B三个通道的color-sift特征decr(e)、decg(e)和decb(e),e表示该图像的第e个边缘点,e=1,2,...,E,E为该训练图像所有边缘显著像素点的总数。Step 3: According to the edge salient pixels of the selected training image, extract the color-sift features of the original training color image in the three channels of red-R, green-G and blue-B respectively, and obtain each edge salient pixel of the image The color-sift features dec r (e), dec g (e) and dec b (e) of the three channels of red-R, green-G and blue-B, e represents the eth edge point of the image, e=1,2,...,E, where E is the total number of all edge salient pixels of the training image.

步骤4:对训练图像数据库中每一幅训练图像执行步骤2—步骤3进行每一张图像所有边缘显著像素点的color-sift特征提取,遍历训练图像数据库中的所有图像,其特征在三个颜色通道内依次为decr,m(em)、decg,m(em)和decb,m(em),m=1,2,...,M,M为训练图像数据库大小,em表示第m张训练图像的第em个边缘像素点,em=1,2,...,Em,Em为第m张训练图像所有边缘显著像素点的总数。Step 4: Perform step 2-step 3 for each training image in the training image database to extract the color-sift feature of all edge salient pixels of each image, traverse all images in the training image database, and its features are in three The color channels are dec r,m (e m ), dec g,m (e m ) and dec b,m (e m ), m=1,2,...,M, M is the size of the training image database , em represents the em-th edge pixel of the m -th training image, em = 1, 2,..., E m , and Em is the total number of all edge salient pixels of the m -th training image.

步骤5:对全部训练图像的所有边缘显著像素点的红色-R通道的color-sift特征decr,m(em),通过K-means聚类计算,取K个聚类中心得到红色-R通道的w行、K列的二维特征字典codr,K为K-means聚类中心的个数,即字典的大小,同样的方法在绿色-G通道和蓝色-B通道内,分别对decg,m(em)和decb,m(em)执行红色-R通道同样的计算分别得到绿色-G通道w行、K列的二维特征字典codg和蓝色-B通道的w行、K列的二维特征字典codb,本例中w为sift特征维数大小为128,K取值为500,即字典大小为500。Step 5: For the color-sift feature dec r,m (e m ) of the red-R channel of all edge significant pixels of all training images, through K-means clustering calculation, take K cluster centers to get red-R The two-dimensional feature dictionary cod r of row w and column K of the channel, K is the number of K-means clustering centers, that is, the size of the dictionary, the same method is used in the green-G channel and the blue-B channel, respectively. dec g,m (e m ) and dec b,m (e m ) perform the same calculation of the red-R channel to obtain the two-dimensional feature dictionary cod g of the green-G channel w row and K column, and the blue-B channel A two-dimensional feature dictionary cod b with w rows and K columns. In this example, w is a sift feature dimension with a size of 128, and K takes a value of 500, that is, the dictionary size is 500.

步骤6:输入检索图像,为公交车类图像中的一张,对于检索图像执行步骤2—步骤3同样提取图像边缘显著点红色-R、绿色-G和蓝色-B三个通道的color-sift特征secr(e)、secg(e)和secb(e),通过步骤5得到的特征字典codr、codg和codb进行编码计算得到检索图像的基于color-sift特征字典的权值直方图特征X,权值直方图特征X包含了红色-R、绿色-G和蓝色-B三通道的权值直方图特征;Step 6: Input the search image, which is one of the bus images. For the search image, perform steps 2-3 and also extract the red-R, green-G and blue-B color- The sift features sec r (e), sec g (e) and sec b (e), through the feature dictionary cod r , cod g and cod b obtained in step 5, are encoded and calculated to obtain the weight of the retrieved image based on the color-sift feature dictionary Value histogram feature X, weight histogram feature X contains the weight histogram features of red-R, green-G and blue-B three channels;

6a)检索图像在红色通道R上所有边缘显著像素点的color-sift特征secr(o),计算secr(o)对应于红色通道二维特征字典codr的K个聚类中心的欧式距离,选择距离值最小时的聚类中心作为该边缘点的所属中心,对所有边缘显著点进行一阶距统计计算得到检索图像在红色通道对应二维特征字典codr的频率直方图hisr6a) Retrieve the color-sift feature sec r (o) of all edge salient pixels of the image on the red channel R, and calculate the Euclidean distance of sec r (o) corresponding to the K cluster centers of the red channel two-dimensional feature dictionary cod r , select the clustering center with the smallest distance value as the center of the edge point, and perform first-order distance statistical calculation on all edge significant points to obtain the frequency histogram histogram corresponding to the two-dimensional feature dictionary cod r of the retrieved image in the red channel ;

6b)假设落在第k个聚类中心qk个边缘显著点,对聚类到第k个聚类中心的qk个边缘显著点计算该显著点对于该聚类中心的离心权值li(k),计算该聚类中心所有显著点最大的离心权值得到该聚类中心的权值向量α(k),采用如下公式计算:6b) Assuming that q k edge significant points fall in the k -th cluster center, calculate the centrifugal weight li( k), calculate the maximum centrifugal weight of all the significant points of the cluster center to obtain the weight vector α(k) of the cluster center, and use the following formula to calculate:

lili (( uu ,, kk )) == 11 ΣΣ vv == 1,21,2 ,, .. .. .. ,, KK || || secsec rr (( uu )) -- codcod (( kk )) || || 22 22 || || secsec rr (( uu )) -- codcod (( vv )) || || 22 22 ,,

α(k)=max(li(u,k)),α(k)=max(li(u,k)),

其中,u表示落在第k个聚类中心第u个边缘显著点,取值为1,2,...,qk,k表示第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小,qk表示落在第k个聚类中心的边缘显著点的总数。对检索图像的频率直方图hisr和对应权值向量α(k)做矩阵点乘运算得到检索图像的权值向量hstr,即对应元素相乘,hstr为一K维列向量,k代表第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小;Among them, u represents the u-th edge significant point falling in the k-th cluster center, and the value is 1, 2,...,q k , k represents the k-th cluster center, and the value is 1, 2,. .., K, K represents the number of cluster centers, that is, the size of the dictionary, and q k represents the total number of edge significant points falling on the kth cluster center. Perform matrix dot multiplication on the frequency histogram his r of the retrieved image and the corresponding weight vector α(k) to obtain the weight vector hst r of the retrieved image, that is, multiply the corresponding elements. hst r is a K-dimensional column vector, and k represents The kth cluster center, the value is 1, 2,..., K, K represents the number of cluster centers, that is, the size of the dictionary;

6c)在绿色通道G、蓝色通道B做同样于红色通道R的计算,最终得到图像绿色通道G的权值向量hstg和蓝色通道B的权值向量hstb,对三个通道内的权值向量进行整合计算得到检索图像的基于color-sift特征字典的权值直方图特征X。6c) Do the same calculation as the red channel R in the green channel G and blue channel B, and finally get the weight vector hst g of the image green channel G and the weight vector hst b of the blue channel B, for the three channels The weight vector is integrated and calculated to obtain the weight histogram feature X based on the color-sift feature dictionary of the retrieved image.

步骤7:从图像总数大小为S的待检索图像数据库中提取一幅待检索图像执行步骤2—步骤4进行每一张图像所有边缘显著像素点的color-sift特征提取,然后执行步骤6得到每一张待检索图像基于权值color-sift特征字典的权值直方图特征X′i,遍历图像数据库中的所有图像,i=1,2,...,S,S为待检索图像总数,本例中使用的数据库为Corel-1000,包括10类,每一类包括100张图像,S的值即为1000。。Step 7: Extract an image to be retrieved from the database of images to be retrieved with a total size of S. Execute steps 2-4 to extract the color-sift feature of all salient pixels on the edge of each image, and then perform step 6 to obtain each An image to be retrieved is based on the weight histogram feature X′ i of the weight color-sift feature dictionary, traversing all the images in the image database, i=1,2,...,S, S is the total number of images to be retrieved, The database used in this example is Corel-1000, including 10 categories, each category includes 100 images, and the value of S is 1000. .

步骤8:把检索图像和待检索图像进行基于权值直方图特征的相似性匹配,计算2范数相似性距离得到Di(X,X′i)。Step 8: Perform similarity matching based on the weight histogram feature between the retrieved image and the image to be retrieved, and calculate the 2-norm similarity distance to obtain D i (X, X′ i ).

步骤9:对于每幅待检索图像按照其Di(X,X′i)的值进行从小到大的顺序排列,显示其中前n张图像即为检索的结果,i=1,2,...,S,S为待检索图像总数,n为返回检索图像数目,取值为人为自主确定的正整数。本例中n取值为20。本发明成功的从Corel-1000图像数据中准确地检索出来20幅和公交车图像相关的公交车类图像,但就此项,检索准确率为100%。Step 9: Arrange each image to be retrieved according to its D i (X,X′ i ) value in ascending order, and display the first n images among them as the retrieval result, i=1,2,... ., S, S is the total number of images to be retrieved, n is the number of returned retrieved images, and the value is a positive integer determined by humans. In this example, the value of n is 20. The present invention successfully retrieves 20 bus images related to the bus image from the Corel-1000 image data, but with respect to this item, the retrieval accuracy rate is 100%.

对图像的内容进行准确快速的描述一直都是图像检索技术中研究的重点和难点.传统的图像特征提取方法,基本上是围绕图像的颜色,纹理,形状和空间关系来展开的。本发明首先在训练图像数据库中随机选取训练图像做灰度变换,通过方向可调滤波器进行处理,根据全部训练图像的边缘显著像素点对原训练彩色图像分别在红色-R、绿色-G和蓝色-B三个通道提取color-sift特征,得到训练图像每一个边缘显著像素点的红色-R、绿色-G和蓝色-B三个通道的color-sift特征。对检索图像和待检索图像基于特征字典进行编码计算得到基于color-sift特征字典的权值直方图,把检索图像和待检索图像进行基于权值直方图特征的相似性匹配,得到检索结果,提高了检索过程的效率、速度和回调率。Accurate and fast description of image content has always been the focus and difficulty of image retrieval technology. Traditional image feature extraction methods are basically developed around the image's color, texture, shape and spatial relationship. The present invention first randomly selects the training image in the training image database for grayscale transformation, processes it through a direction-adjustable filter, and performs red-R, green-G and The color-sift features are extracted from the three blue-B channels, and the color-sift features of the red-R, green-G, and blue-B channels of each edge salient pixel point of the training image are obtained. The retrieved image and the image to be retrieved are encoded and calculated based on the feature dictionary to obtain the weight histogram based on the color-sift feature dictionary, and the retrieved image and the image to be retrieved are matched based on the similarity of the weight histogram feature to obtain the retrieval result and improve The efficiency, speed and callback rate of the retrieval process are improved.

实施例2基于权值color-sift特征字典的图像检索方法同实施例1Embodiment 2 The image retrieval method based on the weight color-sift feature dictionary is the same as that in Embodiment 1

本实例同样选取Corel-1000图像数据库,图像数据库中包括10类图像,每一类包括100张图像,对数据库中的每一张图像执行实施例1同样的检索过程,计算当返回检索图像数目n为20时的全部10类中每一类的平均检索准确率和全部10类1000张图像的平均检索准确率,对检索结果统计并列表,并和本领域现有技术中几种熟知的检索方法如Jhanwar、Hung所提出的方法及基于color-texture-shape的方法、基于SIFT-BOF的方法和基于SIFT-SPM的方法进行了对比,对比结果如表1所示。从表1可见,本发明的在返回检索图像数目n为20时全部10类1000张图像的平均检索准确率明显高于上述每一种用于对比的检索方法,且在全部10类中每一类100张图像的平均检索准确率上高于大部分用于对比的检索方法。因此,本发明在应用于不同类别图像进行检索时,均可以取的较高的平均检索准确率,适用于图像种类较多的大型图像数据的图像检索,且对于每一类均可得到较稳定、较优的平均检索准确率。This example also selects the Corel-1000 image database. The image database includes 10 categories of images, and each category includes 100 images. The same retrieval process as in Embodiment 1 is performed for each image in the database, and the number of retrieved images n is calculated when returning For the average retrieval accuracy rate of each of the 10 categories at 20 o'clock and the average retrieval accuracy rate of 1000 images in all 10 categories, the retrieval results are counted and tabulated, and compared with several well-known retrieval methods in the prior art in the art For example, the method proposed by Jhanwar and Hung was compared with the method based on color-texture-shape, the method based on SIFT-BOF and the method based on SIFT-SPM. The comparison results are shown in Table 1. As can be seen from Table 1, the average retrieval accuracy rate of all 10 categories of 1000 images of the present invention when the number of returned retrieval images n is 20 is significantly higher than each of the above-mentioned retrieval methods for comparison, and each of the 10 categories The average retrieval accuracy of 100 images is higher than most of the retrieval methods used for comparison. Therefore, when the present invention is applied to different types of images for retrieval, it can obtain a higher average retrieval accuracy rate, and is suitable for image retrieval of large-scale image data with many types of images, and can obtain relatively stable results for each type. , Better average retrieval accuracy.

表1Table 1

以上是本发明的两个实例,并不构成对本发明的任何限制,仿真实验表明,本发明不仅能在应用与大型图像数据库时提高了速率,也能实现对于检索结果的拥有较高的准确率和回调率。The above are two examples of the present invention, which do not constitute any limitation to the present invention. The simulation experiment shows that the present invention can not only improve the speed when applied to large image databases, but also achieve higher accuracy for retrieval results. and callback rate.

综上,本发明的基于权值color-sift特征字典的图像检索方法,主要致力于现有技术应用于大型图像数据库时速度、准确率和回调率的提高。其方法步骤为:在待检索图像中随机选取训练图像,对图像做灰度变换;对训练图像通过方向可调滤波器处理;结合方向可调滤波器的结果提取图像的边缘;提取所有训练图像的color-sift特征;通过对所有训练图像的边缘像素点的color-sift特征进行K-means聚类构建特征字典;输入要检索的图像并对检索图像和待检索图像执行同于训练图像的步骤提取color-sift特征;对检索图像和待检索图像基于color-sift特征字典提取权值直方图特征;把检索图像和数据库中待检索图像进行基于权值直方图特征的相似性匹配;按照相似性匹配结果显示图像检索结果。本发明尤其对于大型图像数据库检索本发明具有检索速度快、准确率和回调率较高的优势,可应用于实时人机交互和大型图像数据库的图像检索。In summary, the image retrieval method based on the weighted color-sift feature dictionary of the present invention is mainly dedicated to improving the speed, accuracy and callback rate when the prior art is applied to a large image database. The steps of the method are as follows: randomly select a training image in the image to be retrieved, and perform grayscale transformation on the image; process the training image through a direction-stable filter; combine the results of the direction-stable filter to extract the edge of the image; extract all the training images The color-sift features of all training images are constructed by K-means clustering on the color-sift features of the edge pixels of all training images; input the image to be retrieved and perform the same steps as the training image for the retrieved image and the image to be retrieved Extract the color-sift feature; extract the weight histogram feature based on the color-sift feature dictionary for the retrieved image and the image to be retrieved; perform similarity matching based on the weight histogram feature between the retrieved image and the image to be retrieved in the database; Matching Results displays the image retrieval results. The present invention has the advantages of fast retrieval speed, high accuracy and callback rate especially for large-scale image database retrieval, and can be applied to real-time human-computer interaction and image retrieval of large-scale image databases.

Claims (8)

1.一种基于权值color-sift特征字典的图像检索方法,其特征在于:其包括,1. An image retrieval method based on weight color-sift feature dictionary, is characterized in that: it comprises, 在待检索图像中随机选取训练图像,提取所述训练图像的边缘,并提取所有训练图像边缘点的color-sift特征,并以所述color-sift特征构建特征字典;Randomly select a training image in the image to be retrieved, extract the edge of the training image, and extract the color-sift feature of all training image edge points, and construct a feature dictionary with the color-sift feature; 输入要检索的图像并提取检索图像和待检索图像边缘点的color-sift特征,基于所述特征字典对检索图像和待检索图像提取权值直方图特征;把检索图像和数据库中待检索图像进行基于权值直方图特征的相似性匹配;Input the image to be retrieved and extract the color-sift feature of the image to be retrieved and the edge point of the image to be retrieved, based on the feature dictionary to extract the weight histogram feature of the image to be retrieved and the image to be retrieved; Similarity matching based on weight histogram features; 检测是否遍历所有数据库中的全部待检索图像,若是,则按照相似性匹配结果,显示图像检索结果,若否,则重新进行相似性匹配;Detect whether to traverse all the images to be retrieved in all databases, if yes, display the image retrieval results according to the similarity matching results, if not, perform similarity matching again; 对检索图像编码计算得到检索图像的基于color-sift特征字典的权值直方图特征,包括以下步骤:The retrieval image encoding is calculated to obtain the weight histogram feature based on the color-sift feature dictionary of the retrieval image, including the following steps: 1)检索图像在红色通道R上所有边缘显著像素点的color-sift特征secr(o),计算secr(o)对应于红色通道二维特征字典codr的K个聚类中心的欧式距离,选择距离值最小时的聚类中心作为该边缘点的所属中心,对所有边缘显著点进行一阶距统计计算得到检索图像在红色通道对应二维特征字典codr的频率直方图hisr1) Retrieve the color-sift feature sec r (o) of all edge salient pixels of the image on the red channel R, and calculate sec r (o) corresponding to the Euclidean distance of the K cluster centers of the red channel two-dimensional feature dictionary cod r , select the clustering center with the smallest distance value as the center of the edge point, and perform first-order distance statistical calculation on all edge significant points to obtain the frequency histogram histogram corresponding to the two-dimensional feature dictionary cod r of the retrieved image in the red channel ; 2)假设落在第k个聚类中心qk个边缘显著点,对聚类到第k个聚类中心的qk个边缘显著点计算该显著点对于该聚类中心的离心权值li(k),计算该聚类中心所有显著点最大的离心权值得到该聚类中心的权值向量α(k),对检索图像的频率直方图hisr和对应权值向量α(k)做矩阵点乘运算得到检索图像的权值向量hstr,即对应元素相乘,hstr为一K维列向量,k代表第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小;2) Assuming that there are q k edge significant points falling in the k -th cluster center, calculate the centrifugal weight li( k), calculate the maximum centrifugal weight of all significant points in the cluster center to obtain the weight vector α(k) of the cluster center, and make a matrix for the frequency histogram his r of the retrieved image and the corresponding weight vector α(k) The dot multiplication operation obtains the weight vector hst r of the retrieved image, that is, the multiplication of corresponding elements, hst r is a K-dimensional column vector, k represents the k-th cluster center, and the value is 1, 2,...,K, K represents the number of cluster centers, that is, the size of the dictionary; 3)在绿色通道G、蓝色通道B做同样于红色通道R的计算,最终得到图像绿色通道G的权值向量hstg和蓝色通道B的权值向量hstb,对三个通道内的权值向量进行整合计算得到检索图像的基于color-sift特征字典的权值直方图特征X。3) Do the same calculation as the red channel R in the green channel G and blue channel B, and finally get the weight vector hst g of the image green channel G and the weight vector hst b of the blue channel B, for the three channels The weight vector is integrated and calculated to obtain the weight histogram feature X based on the color-sift feature dictionary of the retrieved image. 2.如权利要求1所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于:在待检索图像中随机选取训练图像包括在待检索图像数据库中对于每类图像随机选取l张图像组成训练图像数据库。2. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 1, it is characterized in that: randomly selecting training images in the image to be retrieved comprises randomly selecting for each type of image in the image database to be retrieved l images constitute the training image database. 3.如权利要求2所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于,提取所述训练图像的边缘包括:提取所述训练图像的边缘包括在训练图像数据库中随机选取一张训练图像做灰度变换,并通过方向可调滤波器进行处理,选取二维高斯函数为滤波器核函数,得到每个像素点2L个方向上的能量函数Wσ(x,y,θ),并经过阈值判断提取图像的边缘显著像素点,其中L表示方向的个数,x和y表示像素点的坐标值,θ为方向的值,范围是0~2π,间隔为π/L。3. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 2, is characterized in that, extracting the edge of described training image comprises: extracting the edge of described training image is included in training image database Randomly select a training image for grayscale transformation, and process it through a direction-tunable filter, select a two-dimensional Gaussian function as the filter kernel function, and obtain the energy function W σ (x,y , θ), and through the threshold judgment to extract the edge significant pixels of the image, where L represents the number of directions, x and y represent the coordinate values of the pixels, θ is the value of the direction, the range is 0 to 2π, and the interval is π/ L. 4.如权利要求3所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于,提取所有训练图像边缘点的color-sift特征包括:根据所选取训练图像的边缘显著像素点对原训练彩色图像分别在红色-R、绿色-G和蓝色-B三个通道提取color-sift特征,得到该图像每一个边缘显著像素点的红色-R、绿色-G和蓝色-B三个通道的color-sift特征decr(e)、decg(e)和decb(e),e表示该图像的第e个边缘点,e=1,2,...,E,E为该图像所有边缘显著像素点的总数。4. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 3, is characterized in that, extracting the color-sift feature of all training image edge points comprises: according to the edge salient pixel of selected training image Point pair the original training color image to extract color-sift features in the three channels of red-R, green-G and blue-B respectively, and obtain the red-R, green-G and blue- The color-sift features dec r (e), dec g (e) and dec b (e) of the three channels of B, e represents the e-th edge point of the image, e=1,2,...,E, E is the total number of all edge salient pixels of the image. 5.如权利要求4所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于,所述提取所有训练图像边缘点的color-sift特征包括:对训练图像数据库中每一幅训练图像所有边缘显著像素点的color-sift特征提取,遍历训练图像数据库中的所有图像,其特征在三个颜色通道内依次为decr,m(em)、decg,m(em)和decb,m(em),m=1,2,...,M,M为训练图像数据库大小,em表示第m张训练图像的第em个边缘像素点,em=1,2,...,Em,Em为第m张训练图像所有边缘显著像素点的总数。5. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 4, is characterized in that, the color-sift feature of described extracting all training image edge points comprises: to each training image database The color-sift feature extraction of all salient pixels on the edge of a training image, traverse all images in the training image database, and its features in the three color channels are dec r,m (e m ), dec g,m (e m ) and dec b, m (e m ), m=1, 2,..., M, M is the size of the training image database, e m represents the em edge pixel of the mth training image, e m = 1,2,...,E m , where E m is the total number of all edge salient pixels of the mth training image. 6.如权利要求5所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于,构建特征字典步骤包括对全部训练图像的所有边缘显著像素点的红色-R通道绿色-G通道和蓝色-B通道的color-sift特征decr,m(em),decg,m(em)和decb,m(em),通过K-means聚类计算,取K个聚类中心得到红色-R通道、绿色-G通道和蓝色-B通道的w行、K列的二维特征字典codr、codg和codb,其中K为K-means聚类中心的个数,即字典的大小。6. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 5, is characterized in that, builds feature dictionary step to comprise red-R channel green- The color-sift features dec r,m (e m ), dec g,m (e m ) and dec b,m (e m ) of the G channel and the blue-B channel are calculated by K-means clustering, and K is taken The two-dimensional feature dictionaries cod r , cod g , and cod b of the red-R channel, green-G channel, and blue-B channel of the w row and K column of the cluster center are obtained, where K is the K-means cluster center number, which is the size of the dictionary. 7.如权利要求1所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于:所述权值直方图特征X包含了红色-R、绿色-G和蓝色-B三通道的权值直方图特征,把检索图像和待检索图像进行基于权值直方图特征的相似性匹配,计算2范数相似性距离得到Di(X,Xi')。7. a kind of image retrieval method based on weight value color-sift feature dictionary as claimed in claim 1, is characterized in that: described weight value histogram feature X comprises red-R, green-G and blue-B The three-channel weight histogram feature, the retrieval image and the image to be retrieved are matched based on the weight histogram feature, and the 2-norm similarity distance is calculated to obtain D i (X,X i '). 8.根据权利要求1所述的一种基于权值color-sift特征字典的图像检索方法,其特征在于:所述的对检索图像编码计算得到检索图像的基于color-sift特征字典的权值直方图特征,其中步骤2)所述的对聚类到第k个聚类中心的qk个边缘显著点计算该显著点对于该聚类中心的离心权值li(k),计算该聚类中心所有显著点最大的离心权值得到该聚类中心的权值向量α(k),采用如下公式计算:8. A kind of image retrieval method based on weight color-sift feature dictionary according to claim 1, it is characterized in that: described retrieval image coding calculation obtains the weight value histogram based on color-sift feature dictionary of retrieval image Graph feature, wherein step 2) described to clustering to the q k edge salient points of the kth cluster center calculates the centrifugal weight li(k) of the salient point for the cluster center, and calculates the cluster center The weight vector α(k) of the cluster center is obtained by the largest centrifugal weight of all significant points, which is calculated by the following formula: ll ii (( uu ,, kk )) == 11 ΣΣ vv == 11 ,, 22 ,, ...... ,, KK || || secsec rr (( uu )) -- cc oo dd (( kk )) || || 22 22 || || secsec rr (( uu )) -- cc oo dd (( vv )) || || 22 22 ,, α(k)=max(li(u,k)),,α(k)=max(li(u,k)), 其中,u表示落在第k个聚类中心第u个边缘显著点,取值为1,2,...,qk,k表示第k个聚类中心,取值为1,2,...,K,K表示聚类中心数,即字典大小,qk表示落在第k个聚类中心的边缘显著点的总数。Among them, u represents the u-th edge significant point falling in the k-th cluster center, and the value is 1, 2,...,q k , k represents the k-th cluster center, and the value is 1, 2,. .., K, K represents the number of cluster centers, that is, the size of the dictionary, and q k represents the total number of edge significant points falling on the kth cluster center.
CN201310294385.2A 2013-07-12 2013-07-12 Image retrieval method based on weight color-sift characteristic dictionary Expired - Fee Related CN103336835B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201310294385.2A CN103336835B (en) 2013-07-12 2013-07-12 Image retrieval method based on weight color-sift characteristic dictionary

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201310294385.2A CN103336835B (en) 2013-07-12 2013-07-12 Image retrieval method based on weight color-sift characteristic dictionary

Publications (2)

Publication Number Publication Date
CN103336835A CN103336835A (en) 2013-10-02
CN103336835B true CN103336835B (en) 2017-02-08

Family

ID=49245000

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201310294385.2A Expired - Fee Related CN103336835B (en) 2013-07-12 2013-07-12 Image retrieval method based on weight color-sift characteristic dictionary

Country Status (1)

Country Link
CN (1) CN103336835B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103699436B (en) * 2013-12-30 2017-02-22 西北工业大学 Image coding method based on local linear constraint and global structural information
CN105138672B (en) * 2015-09-07 2018-08-21 北京工业大学 A kind of image search method of multiple features fusion
CN106503143B (en) * 2016-10-21 2020-02-07 广东工业大学 Image retrieval method and device
CN106570183B (en) * 2016-11-14 2019-11-15 宜宾学院 A Color Image Retrieval and Classification Method
CN107103325B (en) * 2017-04-20 2021-01-26 湘潭大学 Tissue pathology image classification method
CN109711441B (en) * 2018-12-13 2021-02-12 泰康保险集团股份有限公司 Image classification method and device, storage medium and electronic equipment
CN111104936A (en) * 2019-11-19 2020-05-05 泰康保险集团股份有限公司 Text image recognition method, device, equipment and storage medium
CN111428122B (en) * 2020-03-20 2023-09-01 南京中孚信息技术有限公司 Picture retrieval method and device and electronic equipment

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101546307A (en) * 2008-09-17 2009-09-30 厦门市美亚柏科资讯科技有限公司 Image retrieval system and image retrieval method based on edge gravity center characteristics
CN102236717A (en) * 2011-07-13 2011-11-09 清华大学 Image retrieval method based on sketch feature extraction
CN102629328A (en) * 2012-03-12 2012-08-08 北京工业大学 Probabilistic latent semantic model object image recognition method with fusion of significant characteristic of color
CN102693421A (en) * 2012-05-31 2012-09-26 东南大学 Bull eye iris image identifying method based on SIFT feature packs

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2011257979A (en) * 2010-06-09 2011-12-22 Olympus Imaging Corp Image retrieval device, image retrieval method, and camera

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101546307A (en) * 2008-09-17 2009-09-30 厦门市美亚柏科资讯科技有限公司 Image retrieval system and image retrieval method based on edge gravity center characteristics
CN102236717A (en) * 2011-07-13 2011-11-09 清华大学 Image retrieval method based on sketch feature extraction
CN102629328A (en) * 2012-03-12 2012-08-08 北京工业大学 Probabilistic latent semantic model object image recognition method with fusion of significant characteristic of color
CN102693421A (en) * 2012-05-31 2012-09-26 东南大学 Bull eye iris image identifying method based on SIFT feature packs

Also Published As

Publication number Publication date
CN103336835A (en) 2013-10-02

Similar Documents

Publication Publication Date Title
CN103336835B (en) Image retrieval method based on weight color-sift characteristic dictionary
CN111581405B (en) Cross-modal generalization zero sample retrieval method for generating confrontation network based on dual learning
Biswas et al. Linear support tensor machine with LSK channels: Pedestrian detection in thermal infrared images
CN103559196B (en) Video retrieval method based on multi-core canonical correlation analysis
CN103049446B (en) A kind of image search method and device
CN103793721B (en) Pedestrian repeat recognition method and system based on area related feedback
Lin et al. Representing and recognizing objects with massive local image patches
CN103226585B (en) Towards the self-adaptation Hash rearrangement method of image retrieval
CN103854016B (en) Jointly there is human body behavior classifying identification method and the system of feature based on directivity
CN103258037A (en) Trademark identification searching method for multiple combined contents
CN108304573A (en) Target retrieval method based on convolutional neural networks and supervision core Hash
Yasmin et al. Content based image retrieval by shape, color and relevance feedback
Zhang et al. 3D object retrieval with multi-feature collaboration and bipartite graph matching
CN103810252A (en) Image retrieval method based on group sparse feature selection
CN105868706A (en) Method for identifying 3D model based on sparse coding
CN103399863B (en) Image search method based on the poor characteristic bag of edge direction
CN109934258A (en) Image retrieval method based on feature weighting and region integration
CN106649665A (en) Object-level depth feature aggregation method for image retrieval
CN107480471B (en) Method of Sequence Similarity Analysis Based on Wavelet Transform
CN110083724B (en) Similar image retrieval method, device and system
CN110852292A (en) A sketch face recognition method based on cross-modal multi-task deep metric learning
CN104965928B (en) A Chinese character image retrieval method based on shape matching
Sun et al. Search by detection: Object-level feature for image retrieval
Xi et al. Finger vein recognition based on the hyperinformation feature
Gilbert et al. Image and video mining through online learning

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
C14 Grant of patent or utility model
GR01 Patent grant
CF01 Termination of patent right due to non-payment of annual fee
CF01 Termination of patent right due to non-payment of annual fee

Granted publication date: 20170208