CN106548180A - A kind of method for obtaining the Feature Descriptor for obscuring constant image - Google Patents

A kind of method for obtaining the Feature Descriptor for obscuring constant image Download PDF

Info

Publication number
CN106548180A
CN106548180A CN201610922185.0A CN201610922185A CN106548180A CN 106548180 A CN106548180 A CN 106548180A CN 201610922185 A CN201610922185 A CN 201610922185A CN 106548180 A CN106548180 A CN 106548180A
Authority
CN
China
Prior art keywords
image
feature
lpq
local
fuzzy
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201610922185.0A
Other languages
Chinese (zh)
Other versions
CN106548180B (en
Inventor
肖阳
冯晨
曹治国
陈俊
张骁迪
赵富荣
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Huazhong University of Science and Technology
Original Assignee
Huazhong University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Huazhong University of Science and Technology filed Critical Huazhong University of Science and Technology
Priority to CN201610922185.0A priority Critical patent/CN106548180B/en
Publication of CN106548180A publication Critical patent/CN106548180A/en
Application granted granted Critical
Publication of CN106548180B publication Critical patent/CN106548180B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/462Salient features, e.g. scale invariant feature transforms [SIFT]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/469Contour-based spatial representations, e.g. vector-coding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • G06V10/46Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
    • G06V10/469Contour-based spatial representations, e.g. vector-coding
    • G06V10/473Contour-based spatial representations, e.g. vector-coding using gradient analysis

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Physics & Mathematics (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)

Abstract

本发明公开了一种获取模糊不变图像的特征描述子的方法,先将图像划分为多个局部图块,对局部图块细分进行密集采样得到LPQ+;将每个局部图块对应的多个LPQ+合并获得局部图块的特征描述,然后对所有的局部图块的特征描述子进行FV(Fisher Vector)编码并进行相应正则化处理来获得更高维度的模糊图像特征描述。本发明提供的这种获取模糊不变图像的特征描述子的方法,针对模糊图像识别基于LPQ提出的一种更高效、更准确的模糊图像识别特征描述子LPQ+,具有判别性高和特征维度较低的优势;FV编码提高了传统描述子的综合性能;而FV编码与LPQ+的结合,相较于FV编码与传统描述子的结合,在识别精度和识别效率上均具有更好效果。

The invention discloses a method for obtaining a feature descriptor of a fuzzy invariant image. Firstly, the image is divided into a plurality of local blocks, and the subdivision of the local blocks is densely sampled to obtain LPQ+; LPQ+ merge to obtain the feature description of the local block, and then perform FV (Fisher Vector) encoding on all the feature descriptors of the local block and perform corresponding regularization processing to obtain a higher-dimensional fuzzy image feature description. The method for obtaining feature descriptors of fuzzy invariant images provided by the present invention is a more efficient and accurate fuzzy image recognition feature descriptor LPQ+ proposed based on LPQ for fuzzy image recognition, which has high discriminative and relatively small feature dimensions. Compared with the combination of FV coding and traditional descriptors, the combination of FV coding and LPQ+ has better results in recognition accuracy and recognition efficiency.

Description

一种获取模糊不变图像的特征描述子的方法A Method of Obtaining Feature Descriptors of Fuzzy Invariant Images

技术领域technical field

本发明属于数字图像识别技术领域,更具体地,涉及一种获取模糊不变图像的特征描述子的方法。The invention belongs to the technical field of digital image recognition, and more specifically relates to a method for obtaining a feature descriptor of a fuzzy invariant image.

背景技术Background technique

图像采集过程中,环境的干扰、相机本身的虚晃或者失焦容易造成图像模糊,这给诸如人脸识别的、纹理分类和目标检测等图像识别任务带来了现实性的挑战。In the process of image acquisition, the interference of the environment, the swaying of the camera itself or the loss of focus can easily cause image blurring, which brings practical challenges to image recognition tasks such as face recognition, texture classification and object detection.

针对模糊图像识别的问题,现有技术包括以下有两种解决方案;一种是先对图像进行去模糊,然后利用传统的描述子(如LBP、SIFT和HOG等)来进行目标识别,这种方法适用于已知模糊函数的情况,利用非盲反卷积的方式进行图像的去模糊;另一种是直接对图像抽取具有模糊不变性的特征进行识别,而LPQ(Local Phase Quantization)则是其中的代表,利用图像的相位信息,对其进行去相关和量化后投射到一个八维空间,从而在整个图像上构建一个直方图,抽取一个对于中心对称模糊具有不变性的特征描述子。For the problem of fuzzy image recognition, the existing technology includes the following two solutions; one is to first deblur the image, and then use traditional descriptors (such as LBP, SIFT and HOG, etc.) to carry out target recognition. The method is suitable for the situation where the fuzzy function is known, and the non-blind deconvolution method is used to deblur the image; the other is to directly identify the fuzzy invariant features of the image extraction, and LPQ (Local Phase Quantization) is The representative one uses the phase information of the image, decorrelates and quantizes it and projects it into an eight-dimensional space, so as to construct a histogram on the entire image and extract a feature descriptor that is invariant to the central symmetric blur.

上述方法存在以下缺陷,对于第一种方法,通常情况下模糊函数是未知的,因此去模糊处理的时间开销很大,对应此方法下PSF(Point Spread Function)估计的最优化问题的难点是图像模糊类型的选择;第二种方法中的LPQ在面临复杂的视觉识别任务时不能很好的平衡效率与精度;现有技术通过将LPQ与FV(Fisher Vector)结合,将LPQ特征进行编码形成一个更高维度的描述子来表征图像来提升LPQ的综合性能,但维度的提高不仅增加了实验开销,为了防止模型过拟合还增加了对训练样本的需求。The above method has the following defects. For the first method, the blur function is usually unknown, so the time overhead of deblurring processing is very large. The difficulty of the optimization problem of PSF (Point Spread Function) estimation under this method is that the image The choice of fuzzy type; LPQ in the second method cannot balance efficiency and accuracy well when faced with complex visual recognition tasks; the existing technology combines LPQ with FV (Fisher Vector) to encode LPQ features to form a Higher-dimensional descriptors are used to represent images to improve the comprehensive performance of LPQ, but the increase in dimension not only increases the experimental overhead, but also increases the demand for training samples to prevent model overfitting.

发明内容Contents of the invention

针对现有技术的以上缺陷或改进需求,本发明提供了一种获取模糊不变图像的特征描述子的方法,其目的在于通过对局部图块细分进行密集采样后得到的LPQ+进行FV编码来获得更高维度的模糊图像特征描述,以提高模糊图像识别精度和识别效率。Aiming at the above defects or improvement needs of the prior art, the present invention provides a method for obtaining feature descriptors of fuzzy invariant images. Obtain a higher-dimensional fuzzy image feature description to improve the recognition accuracy and efficiency of fuzzy images.

为实现上述目的,按照本发明的一个方面,提供了一种获取模糊不变图像的特征描述子的方法,包括以下步骤:In order to achieve the above object, according to one aspect of the present invention, a method for obtaining a feature descriptor of a blur-invariant image is provided, comprising the following steps:

(1)对源图像数据集里的样本进行模糊处理,并对模糊处理后的图像进行灰度化处理;(1) Blur the samples in the source image data set, and grayscale the blurred image;

(2)采用滑动窗口对灰度化处理获得的图像进行密集采样,获得多个局部图像块,并将各局部图像块划分为多个单元图块;(2) Using a sliding window to densely sample the image obtained by grayscale processing to obtain a plurality of local image blocks, and divide each local image block into a plurality of unit blocks;

(3)对各单元图块抽取LPQ+,然后将抽取到的多组LPQ+进行合并获得局部图块的特征描述子;(3) extracting LPQ+ for each unit block, and then merging the extracted multiple groups of LPQ+ to obtain the feature descriptor of the local block;

其中,LPQ+是基于LPQ的特征描述子,通过直接针对相位而非LPQ中描述的实部虚部(实质为STFT系数)进行量化获得LPQ+;Among them, LPQ+ is a feature descriptor based on LPQ, and LPQ+ is obtained by directly quantizing the phase rather than the real part and imaginary part (essentially STFT coefficient) described in LPQ;

(4)对所有的局部图块的特征描述子接续进行Fisher Vector编码处理、幂正则化和正则化处理,获得模糊图像的特征描述子。(4) Perform Fisher Vector encoding, power regularization and Regularization processing to obtain feature descriptors of blurred images.

优选地,上述获取模糊不变图像的特征描述子的方法,其步骤(4)之后还包括识别的步骤:Preferably, the above-mentioned method for obtaining the feature descriptor of the fuzzy invariant image also includes the step of identification after the step (4):

(5)根据训练数据集和对应特征描述子,采用一对一的多类分类策略训练线性的支撑向量机分类器;分类器的输出结果即为识别结果。(5) According to the training data set and the corresponding feature descriptor, a one-to-one multi-class classification strategy is used to train a linear support vector machine classifier; the output result of the classifier is the recognition result.

优选地,上述获取模糊不变图像的特征描述子的方法,其步骤(2)包括如下子步骤:Preferably, the above-mentioned method for obtaining the feature descriptor of the fuzzy invariant image, its step (2) includes the following sub-steps:

(2-1)采用固定尺寸、步长的滑动窗口对灰度化处理获得的图像进行密集采样,获得K组局部图像块;(2-1) Using a sliding window with a fixed size and a step size to densely sample the image obtained by grayscale processing to obtain K groups of local image blocks;

(2-2)将局部图像块划分为多个小单元,用于保存局部图块的空间结构信息。(2-2) Divide the local image block into multiple small units for storing the spatial structure information of the local block.

优选地,上述获取模糊不变图像的特征描述子的方法,其步骤(3)包括如下子步骤:Preferably, the above-mentioned method for obtaining the feature descriptor of the blur-invariant image, its step (3) includes the following sub-steps:

(3-1)采用STFT(Short Term Fourier Transform),以中心像素周围M×M大小的邻域为计算范围,对单元图块中的每个像素提取相位信息;(3-1) Using STFT (Short Term Fourier Transform), the phase information is extracted for each pixel in the unit block with the M×M size neighborhood around the central pixel as the calculation range;

其中,(x,y)、(u,v)分别指模糊图像和对应傅立叶变换域中的坐标;g(x,y)是指模糊图像,G(u,v)为其对应的傅立叶变换形式;Nx和Ny指像素点(x,y)的邻域范围。Among them, (x, y), (u, v) refer to the coordinates in the fuzzy image and the corresponding Fourier transform domain respectively; g(x, y) refers to the fuzzy image, and G(u, v) is its corresponding Fourier transform form ; N x and N y refer to the neighborhood range of the pixel point (x, y).

采取四个低频点,u1=(a,0),u2=(0,a),u3=(a,a),u4=(a,-a),并对其进行STFT,获得四个相位值∠G(u1),∠G(u2),∠G(u3),∠G(u4);其中,a为常数;Take four low-frequency points, u 1 =(a,0), u 2 =(0,a), u 3 =(a,a), u 4 =(a,-a), and perform STFT on them to obtain Four phase values ∠G(u 1 ), ∠G(u 2 ), ∠G(u 3 ), ∠G(u 4 ); where a is a constant;

(3-2)按对相位值进行划分,获得I个不同的角度区域,将上述四个相位值归入对应角度区域并按照下述投票值进行量化,(3-2) Press The phase value is divided to obtain 1 different angle areas, and the above four phase values are classified into the corresponding angle area and quantified according to the following voting values,

其中,α是可调参数,是指∠G(uj)在对应的角度区域i下获得的投票值;其中,j=1,2,3,4;Among them, α is an adjustable parameter, Refers to the voting value obtained by ∠G(u j ) under the corresponding angle area i; where, j=1,2,3,4;

(3-3)将四个相位值的量化值的分布表示成四个直方图并将直方图合并得到LPQ+;(3-3) Express the distribution of quantized values of the four phase values as four histograms and merge the histograms to obtain LPQ+;

(3-4)将同一图块的不同单元块的LPQ+合并,获得局部图块的特征描述子。(3-4) Merge the LPQ+ of different unit blocks of the same block to obtain the feature descriptor of the local block.

优选地,上述获取模糊不变图像的特征描述子的方法,其步骤(4)包括如下子步骤:Preferably, the above-mentioned method for obtaining the feature descriptor of the fuzzy invariant image, its step (4) includes the following sub-steps:

(4-1)将抽取的K个局部图块特征描述子记为X={xk,k=1,2,…,K};(4-1) Denote the extracted K local block feature descriptors as X={x k ,k=1,2,...,K};

(4-2)用N元的混合高斯模型uλ(x)来模拟X的生成过程,该模型记为(4-2) Use N-ary mixed Gaussian model u λ (x) to simulate the generation process of X, which is denoted as

其中,参数λ={ωlll,l=1,…,N},ωl,μl和σl分别是指高斯函数ul的混合权重、平均向量和标准差;参数λ依据EM准则(期望最大准则)在由大量训练集生成的局部特征描述子上进行估计;Among them, the parameter λ={ω lll ,l=1,…,N}, ω l , μ l and σ l refer to the mixing weight, mean vector and standard deviation of the Gaussian function u l respectively; the parameter λ is estimated on local feature descriptors generated from a large number of training sets according to the EM criterion (expectation maximum criterion);

(4-3)将高斯模型uλ(x)分别对μl和σl求偏导,获得两个梯度向量:(4-3) Calculate partial derivatives of Gaussian model u λ (x) for μ l and σ l respectively, and obtain two gradient vectors:

其中,γk(l)表示特征xk由第l个高斯函数生成的概率;Among them, γ k (l) represents the probability that the feature x k is generated by the l-th Gaussian function;

(4-4)将所有的N个高斯函数对应梯度向量进行合并,得到Fisher Vector编码 (4-4) Merge all N Gaussian function corresponding gradient vectors to obtain Fisher Vector code

(4-5)对的每一个维度m进行如下幂正则化(power normalization,处理:(4-5) right Each dimension m of is subjected to the following power normalization (power normalization, processing:

(4-6)对上述幂正则化后的进行正则化处理,得到模糊图像特征的描述子: (4-6) After the above power regularization conduct Regularization processing, get the descriptor of fuzzy image features:

总体而言,通过本发明所构思的以上技术方案与现有技术相比,能够取得下列有益效果:Generally speaking, compared with the prior art, the above technical solutions conceived by the present invention can achieve the following beneficial effects:

(1)本发明提供的获取模糊不变图像的特征描述子的方法,基于LPQ,但直接对相位而非LPQ中描述的实部和虚部(实质上为STFT系数)进行量化获得LPQ+;由于LPQ实质上只采用了STFT系数的符号信息而忽略了具体数值,而LPQ+为了弥补对具体的相位值进行量化,可更好地发挥LPQ的判别性;同时由于LPQ+特征维数比LPQ更低,因此具有提高处理效率的效果;(1) The method for obtaining the feature descriptor of the fuzzy invariant image provided by the present invention is based on LPQ, but directly quantizes the real part and the imaginary part (essentially STFT coefficients) described in the phase instead of LPQ to obtain LPQ+; LPQ essentially only uses the symbol information of the STFT coefficients and ignores the specific values, while LPQ+ can make better use of the discriminative properties of LPQ in order to compensate for the quantization of specific phase values; at the same time, because the LPQ+ feature dimension is lower than that of LPQ, Therefore, it has the effect of improving processing efficiency;

(2)本发明提供的获取模糊不变图像的特征描述子的方法,对局部特征描述采用Fisher Vector编码,将低维的特征描述投影至高维特征空间,结合线性支撑向量机分类器极大的提高了描述子的描述性能和模糊图像的识别精度;(2) The method for obtaining the feature descriptors of fuzzy invariant images provided by the present invention uses Fisher Vector encoding for local feature descriptions, and projects low-dimensional feature descriptions to high-dimensional feature spaces, combined with the extremely large linear support vector machine classifier Improve the description performance of the descriptor and the recognition accuracy of blurred images;

由此,本发明提供了一种识别精度、速度更高的获取模糊不变图像的特征描述子的方法。Therefore, the present invention provides a method for obtaining feature descriptors of blur-invariant images with higher recognition accuracy and higher speed.

附图说明Description of drawings

图1是本发明实施例提供的获取模糊不变图像的特征描述子的方法的流程示意图;FIG. 1 is a schematic flowchart of a method for obtaining a feature descriptor of a blur-invariant image provided by an embodiment of the present invention;

图2为本发明实施例中抽取特征描述子LPQ+的流程示意图;Fig. 2 is a schematic flow chart of extracting feature descriptor LPQ+ in an embodiment of the present invention;

图3为本发明实施例中密集抽取LPQ+得到局部特征描述子的流程示意图;FIG. 3 is a schematic flow diagram of obtaining a local feature descriptor by densely extracting LPQ+ in an embodiment of the present invention;

图4为本发明实施例中采用Fisher Vector对局部图块描述子进行编码形成图像特征描述子的流程示意图。Fig. 4 is a schematic flow diagram of encoding local block descriptors to form image feature descriptors by using Fisher Vector in an embodiment of the present invention.

具体实施方式detailed description

为了使本发明的目的、技术方案及优点更加清楚明白,以下结合附图及实施例,对本发明进行进一步详细说明。应当理解,此处所描述的具体实施例仅仅用以解释本发明,并不用于限定本发明。此外,下面所描述的本发明各个实施方式中所涉及到的技术特征只要彼此之间未构成冲突就可以相互组合。In order to make the object, technical solution and advantages of the present invention clearer, the present invention will be further described in detail below in conjunction with the accompanying drawings and embodiments. It should be understood that the specific embodiments described here are only used to explain the present invention, not to limit the present invention. In addition, the technical features involved in the various embodiments of the present invention described below can be combined with each other as long as they do not constitute a conflict with each other.

本发明提供了一种模糊不变的图像特征描述,其流程如图1所示,包括模糊图像的获取,对图像进行密集采样,抽取LPQ+,采用Fisher Vector进行编码,训练支撑向量机线性分类器,获得识别结果;以下结合实施例来具体阐述本发明提供的获取模糊不变图像的特征描述子的方法,具体步骤如下:The present invention provides a fuzzy invariant image feature description, the process of which is shown in Figure 1, including the acquisition of fuzzy images, intensive sampling of images, extraction of LPQ+, encoding with Fisher Vector, and training of support vector machine linear classifiers , to obtain the recognition result; the method for obtaining the feature descriptor of the fuzzy invariant image provided by the present invention is specifically described below in conjunction with the embodiments, and the specific steps are as follows:

步骤1,获取模糊图像样本,包括如下子步骤:Step 1, obtaining blurred image samples, including the following sub-steps:

(1-1)获取源图像数据集;在本实施例中采用五组不同种类的数据集:15类别的Yale人脸数据集,10类别的KTH纹理数据集,21类别的土地使用场景数据集,6类别的HUST云数据集以及17类的Oxford Flower数据集;(1-1) Acquire the source image data set; in the present embodiment, five groups of different types of data sets are adopted: the Yale face data set of 15 categories, the KTH texture data set of 10 categories, and the land use scene data set of 21 categories , 6 categories of HUST cloud datasets and 17 categories of Oxford Flower datasets;

(1-2)获取模糊图像;在实施例中采用三种图像模糊方式:高斯模糊、运动模糊和圆形模糊;具体地,针对高斯模糊,邻域窗口的尺寸设置为3×3,标准差分别设置为1、2、3;针对运动模糊,相机的线性运动参数分别设置为8、9;针对圆形模糊,半径分别设置为2、3;(1-2) Acquire blurred images; in the embodiment, three image blurring methods are adopted: Gaussian blur, motion blur and circular blur; specifically, for Gaussian blur, the size of the neighborhood window is set to 3×3, and the standard deviation Set to 1, 2, 3 respectively; for motion blur, the linear motion parameters of the camera are set to 8, 9 respectively; for circular blur, the radius is set to 2, 3 respectively;

(1-3)将所有的彩色图像根据下式进行灰度化:(1-3) Grayscale all color images according to the following formula:

K=0.2989×R+0.5870×G+0.1140×B;K=0.2989×R+0.5870×G+0.1140×B;

其中,K是灰度图像,R,G,B为彩色图像的三个通道。Among them, K is the grayscale image, and R, G, and B are the three channels of the color image.

步骤2,对图像进行密集采样,包括如下子步骤:Step 2, densely sample the image, including the following sub-steps:

(2-1)采用尺寸为16×16,横向步长和纵向步长均为8的滑动窗口对灰度化处理获得的图像进行密集采样,获得局部图像块;(2-1) Using a sliding window with a size of 16×16 and a horizontal step size and a vertical step size of 8 to densely sample the image obtained by grayscale processing to obtain a local image block;

(2-2)将获取的局部图像块划分为4个单元图块,用于保存局部图块的空间结构信息。(2-2) Divide the acquired local image block into 4 unit blocks, which are used to store the spatial structure information of the local block.

步骤3,对不同单元的图块抽取LPQ+,然后进行合并为局部图块的特征描述子;其中,局部图块特征描述子的抽取以及密集采样的流程图分别如图2和图3所示;具体步骤如下:Step 3, extracting LPQ+ from blocks of different units, and then merging them into feature descriptors of local blocks; wherein, the flow charts of extraction of feature descriptors of local blocks and dense sampling are shown in Figure 2 and Figure 3 respectively; Specific steps are as follows:

(3-1)采用STFT(Short Term Fourier Transform)对单元中的每个像素提取相位信息;STFT的计算范围为中心像素周围13×13大小的邻域:(3-1) Use STFT (Short Term Fourier Transform) to extract phase information for each pixel in the unit; the calculation range of STFT is a 13×13 neighborhood around the central pixel:

其中,(x,y)、(u,v)分别指模糊图像和对应傅立叶变换域中的坐标;g(x,y)是指模糊图像,G(u,v)为其对应的傅立叶变换形式;Nx和Ny指像素点(x,y)的邻域范围。Among them, (x, y), (u, v) refer to the coordinates in the fuzzy image and the corresponding Fourier transform domain respectively; g(x, y) refers to the fuzzy image, and G(u, v) is its corresponding Fourier transform form ; N x and N y refer to the neighborhood range of the pixel point (x, y).

采取四个低频点,u1=(a,0),u2=(0,a),u3=(a,a),u4=(a,-a),其中参数并对其进行STFT处理,得到四个相位值∠G(u1),∠G(u2),∠G(u3),∠G(u4);Taking four low-frequency points, u 1 =(a,0), u 2 =(0,a), u 3 =(a,a), u 4 =(a,-a), where the parameter And perform STFT processing on it to get four phase values ∠G(u 1 ), ∠G(u 2 ), ∠G(u 3 ), ∠G(u 4 );

(3-2)按照将相位值划分8个不同的角度区域;将上述四个相位值归入对应角度区域并按照下述投票值进行量化,(3-2) According to Divide the phase value into 8 different angle areas; classify the above four phase values into the corresponding angle area and quantify according to the following voting values,

是指∠G(uj)在对应的角度区域i下获得的投票值;其中,j=1,2,3,4; Refers to the voting value obtained by ∠G(u j ) under the corresponding angle area i; where, j=1,2,3,4;

(3-3)将四个相位的量化值的分布表示成四个直方图并将直方图合并得到LPQ+;(3-3) Expressing the distribution of the quantized values of the four phases as four histograms and merging the histograms to obtain LPQ+;

步骤4,对局部图块的特征描述子进行Fisher Vector编码处理、幂正则化和正则化处理,获得模糊图像的特征描述子;实施例中,Fisher Vector的实现基于VLFeat,其流程如图4,包括如下子步骤:Step 4, perform Fisher Vector encoding processing, power regularization and Regularization processing obtains the feature descriptor of the fuzzy image; in the embodiment, the realization of Fisher Vector is based on VLFeat, and its flow process is shown in Figure 4, including the following sub-steps:

(4-1)将抽取的K个局部图块特征描述子记为X={xk,k=1,2,…,K};(4-1) Denote the extracted K local block feature descriptors as X={x k ,k=1,2,...,K};

(4-2)用50元的混合高斯模型uλ(x)来模拟X的生成过程,(4-2) Use a 50-element mixed Gaussian model u λ (x) to simulate the generation process of X,

其中,参数λ={ωlll,l=1,…,50},ωl,μl和σl分别表示为高斯函数ul的混合权重、平均向量和标准差;参数λ依据EM准则,在由大量训练集生成的局部特征描述子上进行估计;Among them, the parameter λ={ω lll ,l=1,…,50}, ω l , μ l and σ l represent the mixture weight, mean vector and standard deviation of the Gaussian function u l respectively; the parameter λ is estimated on local feature descriptors generated from a large number of training sets according to the EM criterion;

(4-3)将高斯模型uλ(x)分别对μl和σl求偏导,得到两个梯度向量:(4-3) Calculate the partial derivative of the Gaussian model u λ (x) for μ l and σ l respectively, and obtain two gradient vectors:

其中,γk(l)表示特征xk由第l个高斯函数生成的概率;Among them, γ k (l) represents the probability that the feature x k is generated by the l-th Gaussian function;

(4-4)将所有的50个高斯函数对应梯度向量进行合并,得到Fisher Vector编码 (4-4) Merge all 50 Gaussian function corresponding gradient vectors to get Fisher Vector code

(4-5)对上述的每一个维度m进行如下幂正则化处理,其中,幂指数0.5:(4-5) For the above Each dimension m of is subjected to power regularization as follows, where the power index is 0.5:

(4-5)对幂正则化后的进行正则化处理,得到模糊图像特征的描述子 (4-5) After power regularization conduct Regularization processing to obtain descriptors of blurred image features

步骤5,根据训练数据集和对应特征描述子,采用一对一的多类分类策略训练线性的支撑向量机分类器;在本实施例中采取开源的LIBSVM来训练和建立支撑向量机分类器;分类器的输出结果即最后的识别结果。Step 5, according to the training data set and the corresponding feature descriptor, adopt a one-to-one multiclass classification strategy to train a linear support vector machine classifier; in this embodiment, open source LIBSVM is adopted to train and establish a support vector machine classifier; The output result of the classifier is the final recognition result.

本领域的技术人员容易理解,以上所述仅为本发明的较佳实施例而已,并不用以限制本发明,凡在本发明的精神和原则之内所作的任何修改、等同替换和改进等,均应包含在本发明的保护范围之内。It is easy for those skilled in the art to understand that the above descriptions are only preferred embodiments of the present invention, and are not intended to limit the present invention. Any modifications, equivalent replacements and improvements made within the spirit and principles of the present invention, All should be included within the protection scope of the present invention.

Claims (5)

1.一种获取模糊不变图像的特征描述子的方法,其特征在于,包括以下步骤:1. A method for obtaining a feature descriptor of a fuzzy invariant image, comprising the following steps: (1)对源图像里的样本进行模糊处理,并对模糊处理后的图像进行灰度化处理;(1) Blur the samples in the source image, and grayscale the blurred image; (2)采用滑动窗口对灰度化处理获得的图像进行密集采样,获得多个局部图像块,并将各局部图像块划分为多个单元图块;(2) Using a sliding window to densely sample the image obtained by grayscale processing to obtain a plurality of local image blocks, and divide each local image block into a plurality of unit blocks; (3)对各单元图块抽取LPQ+,然后将抽取到的多组LPQ+进行合并获得局部图块的特征描述子;(3) extracting LPQ+ for each unit block, and then merging the extracted multiple groups of LPQ+ to obtain the feature descriptor of the local block; (4)对所有的局部图块的特征描述子接续进行Fisher Vector编码处理、幂正则化和正则化处理,获得模糊图像的特征描述子。(4) Perform Fisher Vector encoding, power regularization and Regularization processing to obtain feature descriptors of blurred images. 2.如权利要求1所述的获取模糊不变图像的特征描述子的方法,其特征在于,所述步骤(4)之后还包括图像识别的步骤:2. the method for obtaining the feature descriptor of fuzzy invariant image as claimed in claim 1, is characterized in that, also comprises the step of image recognition after described step (4): (5)根据训练数据集和对应特征描述子,采用一对一的多类分类策略训练线性的支撑向量机分类器;分类器的输出结果即为图像识别结果。(5) According to the training data set and the corresponding feature descriptor, a one-to-one multi-class classification strategy is used to train a linear support vector machine classifier; the output result of the classifier is the image recognition result. 3.如权利要求1或2所述的获取模糊不变图像的特征描述子的方法,其特征在于,所述步骤(2)包括如下子步骤:3. the method for obtaining the feature descriptor of fuzzy invariant image as claimed in claim 1 or 2, is characterized in that, described step (2) comprises following substep: (2-1)采用固定尺寸、步长的滑动窗口对灰度化处理获得的图像进行密集采样,获得多个局部图像块;(2-1) Using a fixed-size, step-size sliding window to densely sample the image obtained by grayscale processing to obtain multiple local image blocks; (2-2)将各局部图像块划分为多个小单元,用于保存局部图块的空间结构信息。(2-2) Divide each local image block into a plurality of small units for storing the spatial structure information of the local block. 4.如权利要求1或2所述的获取模糊不变图像的特征描述子的方法,其特征在于,所述步骤(3)包括如下子步骤:4. the method for obtaining the feature descriptor of fuzzy invariant image as claimed in claim 1 or 2, is characterized in that, described step (3) comprises following substep: (3-1)采用STFT以中心像素周围M×M大小的邻域为计算范围,提取单元图块中每个像素的相位信息:(3-1) Using STFT to extract the phase information of each pixel in the unit block with the M×M neighborhood around the central pixel as the calculation range: GG (( uu ,, vv )) == ΣΣ xx ∈∈ NN xx ΣΣ ythe y ∈∈ NN ythe y gg (( xx ,, ythe y )) ee -- jj 22 ππ (( uu xx ++ vv ythe y )) Mm 其中,(x,y)、(u,v)分别指模糊图像与对应傅立叶变换域中的坐标;g(x,y)是指模糊图像,Nx和Ny指像素点(x,y)的邻域范围;Among them, (x, y), (u, v) refer to the coordinates in the blurred image and the corresponding Fourier transform domain; g(x, y) refers to the blurred image, N x and N y refer to the pixel point (x, y) the neighborhood range; 根据四个低频点u1=(a,0),u2=(0,a),u3=(a,a),u4=(a,-a),获得四个相位值∠G(u1),∠G(u2),∠G(u3),∠G(u4);其中,a为常数;According to the four low-frequency points u 1 =(a,0), u 2 =(0,a), u 3 =(a,a), u 4 =(a,-a), four phase values ∠G( u 1 ), ∠G(u 2 ), ∠G(u 3 ), ∠G(u 4 ); where a is a constant; (3-2)按i=1,2,…,I,对相位值进行划分,获得I个不同的角度区域,将上述四个相位值归入对应角度区域并按照下述投票值进行量化,(3-2) Press i=1,2,...,I, divide the phase value to obtain I different angle areas, classify the above four phase values into the corresponding angle areas and quantify according to the following voting values, vv jj ii == (( cc oo sthe s (( ∠∠ GG (( uu jj )) -- θθ ‾‾ ii )) )) αα 其中,α是可调参数,是指∠G(uj)在对应的角度区域i下获得的投票值;其中,j=1,2,3,4;Among them, α is an adjustable parameter, Refers to the voting value obtained by ∠G(u j ) under the corresponding angle area i; where, j=1,2,3,4; (3-3)根据四个相位值的量化值获得四个直方图并将直方图合并得到LPQ+;(3-3) Obtain four histograms according to the quantized values of the four phase values and merge the histograms to obtain LPQ+; (3-4)将同一局部图像块的各单元块的LPQ+合并,获得局部图块的特征描述子。(3-4) Merge the LPQ+ of each unit block of the same local image block to obtain the feature descriptor of the local block. 5.如权利要求1或2所述的获取模糊不变图像的特征描述子的方法,其特征在于,所述步骤(4)包括如下子步骤:5. the method for obtaining the feature descriptor of fuzzy invariant image as claimed in claim 1 or 2, is characterized in that, described step (4) comprises following substep: (4-1)将抽取的K个局部图块特征描述子记为X={xk,k=1,2,…,K};(4-1) Denote the extracted K local block feature descriptors as X={x k ,k=1,2,...,K}; (4-2)用N元的混合高斯模型uλ(x)来模拟X的生成过程,该模型记为(4-2) Use N-ary mixed Gaussian model u λ (x) to simulate the generation process of X, which is denoted as uu λλ (( xx )) == ΣΣ ll == 11 NN ωω ll uu ll (( xx )) 其中,参数λ={ωlll,l=1,…,N},ωl,μl和σl分别是指高斯函数ul的混合权重、平均向量和标准差;参数λ依据EM准则在由大量训练集生成的局部特征描述子上进行估计;Among them, the parameter λ={ω lll ,l=1,…,N}, ω l , μ l and σ l refer to the mixing weight, mean vector and standard deviation of the Gaussian function u l respectively; the parameter λ is estimated on local feature descriptors generated from a large training set according to the EM criterion; (4-3)将高斯模型uλ(x)分别对μl和σl求偏导,获得两个梯度向量:(4-3) Calculate partial derivatives of Gaussian model u λ (x) for μ l and σ l respectively, and obtain two gradient vectors: 其中,γk(l)表示特征xk由第l个高斯函数生成的概率;Among them, γ k (l) represents the probability that the feature x k is generated by the l-th Gaussian function; (4-4)将所有的N个高斯函数对应梯度向量进行合并,得到Fisher Vector编码 (4-4) Merge all N Gaussian function corresponding gradient vectors to obtain Fisher Vector code (4-5)对的每一个维度m进行如下幂正则化处理:(4-5) right Each dimension m of is subjected to power regularization as follows: (4-6)对所述幂正则化处理后的进行正则化处理,得到模糊图像特征的描述子: (4-6) After the power regularization process conduct Regularization processing, get the descriptor of fuzzy image features:
CN201610922185.0A 2016-10-21 2016-10-21 A method of obtaining the Feature Descriptor for obscuring constant image Active CN106548180B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201610922185.0A CN106548180B (en) 2016-10-21 2016-10-21 A method of obtaining the Feature Descriptor for obscuring constant image

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201610922185.0A CN106548180B (en) 2016-10-21 2016-10-21 A method of obtaining the Feature Descriptor for obscuring constant image

Publications (2)

Publication Number Publication Date
CN106548180A true CN106548180A (en) 2017-03-29
CN106548180B CN106548180B (en) 2019-04-12

Family

ID=58392232

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201610922185.0A Active CN106548180B (en) 2016-10-21 2016-10-21 A method of obtaining the Feature Descriptor for obscuring constant image

Country Status (1)

Country Link
CN (1) CN106548180B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110807464A (en) * 2019-10-21 2020-02-18 华中科技大学 Method and system for obtaining image fuzzy invariant texture feature descriptor
CN111553893A (en) * 2020-04-24 2020-08-18 成都飞机工业(集团)有限责任公司 Method for identifying automatic wiring and cutting identifier of airplane wire harness

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104123560A (en) * 2014-07-03 2014-10-29 中山大学 Phase encoding characteristic and multi-metric learning based vague facial image verification method
CN104517104A (en) * 2015-01-09 2015-04-15 苏州科达科技股份有限公司 Face recognition method and face recognition system based on monitoring scene
CN104537381A (en) * 2014-12-30 2015-04-22 华中科技大学 Blurred image identification method based on blurred invariant feature
CN105893916A (en) * 2014-12-11 2016-08-24 深圳市阿图姆科技有限公司 New method for detection of face pretreatment, feature extraction and dimensionality reduction description

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104123560A (en) * 2014-07-03 2014-10-29 中山大学 Phase encoding characteristic and multi-metric learning based vague facial image verification method
CN105893916A (en) * 2014-12-11 2016-08-24 深圳市阿图姆科技有限公司 New method for detection of face pretreatment, feature extraction and dimensionality reduction description
CN104537381A (en) * 2014-12-30 2015-04-22 华中科技大学 Blurred image identification method based on blurred invariant feature
CN104517104A (en) * 2015-01-09 2015-04-15 苏州科达科技股份有限公司 Face recognition method and face recognition system based on monitoring scene

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
MENGYU ZHU ET AL: "《BEYOND LOCAL PHASE QUANTIZATION: MID-LEVEL BLURRED IMAGE REPRESENTATION USING FISHER VECTOR》", 《IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING》 *
储久良等: "《基于LPQ和Fisherfaces的模糊人脸识别》", 《河南理工大学学报(自然科学版)》 *
朱梦宇: "《图像模糊不变特征提取与识别技术研究》", 《万方数据》 *

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110807464A (en) * 2019-10-21 2020-02-18 华中科技大学 Method and system for obtaining image fuzzy invariant texture feature descriptor
CN110807464B (en) * 2019-10-21 2022-09-20 华中科技大学 Method and system for obtaining image fuzzy invariant texture feature descriptor
CN111553893A (en) * 2020-04-24 2020-08-18 成都飞机工业(集团)有限责任公司 Method for identifying automatic wiring and cutting identifier of airplane wire harness

Also Published As

Publication number Publication date
CN106548180B (en) 2019-04-12

Similar Documents

Publication Publication Date Title
CN104966085B (en) A kind of remote sensing images region of interest area detecting method based on the fusion of more notable features
CN107341786B (en) The infrared and visible light image fusion method that wavelet transformation and joint sparse indicate
CN105809198A (en) SAR image target recognition method based on deep belief network
CN102122386B (en) SAR (stop and reveres) image segmentation method based on dictionary migration clustering
CN107563433A (en) A kind of infrared small target detection method based on convolutional neural networks
CN108154133B (en) Face portrait-photo recognition method based on asymmetric joint learning
CN105528794A (en) Moving object detection method based on Gaussian mixture model and superpixel segmentation
CN106203448B (en) A scene classification method based on nonlinear scale space
CN111242061B (en) A synthetic aperture radar ship target detection method based on attention mechanism
CN103489004A (en) Method for achieving large category image identification of deep study network
AU2011207120A1 (en) Identifying matching images
CN106874879A (en) Handwritten Digit Recognition method based on multiple features fusion and deep learning network extraction
CN105138951B (en) Human face portrait-photo array the method represented based on graph model
CN111127360A (en) Gray level image transfer learning method based on automatic encoder
CN106650765A (en) Hyperspectral data classification method through converting hyperspectral data to gray image based on convolutional neural network
CN104021567B (en) Based on the fuzzy altering detecting method of image Gauss of first numeral law
CN111639589A (en) Video false face detection method based on counterstudy and similar color space
CN109003247B (en) A Method of Removing Mixed Noise in Color Image
CN115272306B (en) Solar cell panel grid line enhancement method utilizing gradient operation
CN111242891B (en) Rail surface defect identification and classification method
CN104574352A (en) Crowd density grade classification method based on foreground image
CN102831621B (en) Video significance processing method based on spectral analysis
CN106157240A (en) Remote sensing image super resolution method based on dictionary learning
CN106548180A (en) A kind of method for obtaining the Feature Descriptor for obscuring constant image
CN114581789A (en) Hyperspectral image classification method and system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant