CN110188646B - Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern - Google Patents
Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern Download PDFInfo
- Publication number
- CN110188646B CN110188646B CN201910433620.7A CN201910433620A CN110188646B CN 110188646 B CN110188646 B CN 110188646B CN 201910433620 A CN201910433620 A CN 201910433620A CN 110188646 B CN110188646 B CN 110188646B
- Authority
- CN
- China
- Prior art keywords
- human ear
- feature
- value
- image
- pixel
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 230000004927 fusion Effects 0.000 title claims abstract description 12
- 239000013598 vector Substances 0.000 claims description 25
- 230000009467 reduction Effects 0.000 claims description 11
- 238000012847 principal component analysis method Methods 0.000 claims description 10
- 239000011159 matrix material Substances 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 6
- 238000012545 processing Methods 0.000 claims description 4
- 230000008859 change Effects 0.000 claims description 3
- 238000006243 chemical reaction Methods 0.000 claims description 2
- 239000000284 extract Substances 0.000 abstract 2
- 238000000513 principal component analysis Methods 0.000 abstract 1
- 238000005516 engineering process Methods 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000013527 convolutional neural network Methods 0.000 description 2
- 238000002474 experimental method Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000012549 training Methods 0.000 description 2
- 230000009471 action Effects 0.000 description 1
- 238000013473 artificial intelligence Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 210000000883 ear external Anatomy 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000014509 gene expression Effects 0.000 description 1
- 238000005286 illumination Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 238000010606 normalization Methods 0.000 description 1
- 239000013589 supplement Substances 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2135—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on approximation criteria, e.g. principal component analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/44—Local feature extraction by analysis of parts of the pattern, e.g. by detecting edges, contours, loops, corners, strokes or intersections; Connectivity analysis, e.g. of connected components
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/50—Extraction of image or video features by performing operations within image blocks; by using histograms, e.g. histogram of oriented gradients [HoG]; by summing image-intensity values; Projection analysis
- G06V10/507—Summing image-intensity values; Histogram projection analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
- G06V40/171—Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/40—Extraction of image or video features
- G06V10/46—Descriptors for shape, contour or point-related descriptors, e.g. scale invariant feature transform [SIFT] or bags of words [BoW]; Salient regional features
- G06V10/467—Encoded features or binary features, e.g. local binary patterns [LBP]
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- General Health & Medical Sciences (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Human Computer Interaction (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
本发明公开一种基于梯度方向直方图与局部二值模式融合的人耳识别方法,解决了从人耳图像中进行人耳识别识别率不高的问题。本发明首先提取出图像的梯度方向直方图,并用主成分分析法进行降维,再用局部二值模式提取出图像的纹理特征,然后将两个特征融合,最后使用最小距离分类器进行分类。本发明通过多特征融合,提高了人耳识别的识别率,具有良好的实施性与实效性。
The invention discloses a human ear recognition method based on the fusion of gradient direction histogram and local binary pattern, which solves the problem of low recognition rate of human ear recognition from human ear images. The invention first extracts the gradient direction histogram of the image, reduces the dimension by principal component analysis, then extracts the texture feature of the image by using the local binary mode, then fuses the two features, and finally uses the minimum distance classifier for classification. The present invention improves the recognition rate of human ear recognition through multi-feature fusion, and has good practicability and effectiveness.
Description
技术领域technical field
本发明涉及一种基于梯度方向直方图与局部二值模式融合的人耳识别方法,属于生物特征识别、深度学习、人工智能等交叉技术领域。The invention relates to a human ear recognition method based on the fusion of a gradient direction histogram and a local binary pattern, and belongs to the cross technical fields of biological feature recognition, deep learning, artificial intelligence and the like.
背景技术Background technique
人耳识别作为一种新的生物特征识别技术,其理论与应用研究在近两年受到了国内外学者更多的关注,具有重要的理论意义与实际应用价值。As a new biometric identification technology, human ear recognition has received more attention from scholars at home and abroad in the past two years, and has important theoretical significance and practical application value.
人耳识别是以人耳图像作为研究对象进行特征识别,既可作为其他生物识别技术的有益补充,也可以单独应用于个体身份识别的场合。在基于生物特征的身份识别技术中,人耳识别具有众多优点,如人耳图像尺寸小、计算量小,外耳的颜色分布一致,在转换为灰度图像时信息丢失少,人耳不受表情变化的影响,可以实现非打扰式识别等。Human ear recognition is based on human ear images as the research object for feature recognition, which can be used as a beneficial supplement to other biometric technologies, and can also be used alone in the occasion of individual identification. In biometric-based identification technology, human ear recognition has many advantages, such as the small size of the human ear image, the small amount of calculation, the consistent color distribution of the outer ear, less information loss when converting to a grayscale image, and the human ear is not affected by expressions. The impact of changes, non-intrusive recognition can be achieved, etc.
目前人耳识别的方法按照所提取的特征来划分可以归纳为两大类:一类是基于几何特征的方法这类方法通过寻找人耳轮廓和内部结构的关键点,构建几何特征。这类方法易受光照、成像角度影响,鲁棒性较差。一类是基于代数特征的方法,例如主元分析法、不变矩Ⅲ方法、小波变换方法等。这些方法在人耳姿态变化不大、图像质量较好的情况下取得了满意的结果。然而当人耳有旋转角度变化时,其二维图像会带来较大的变形,这时传统方法的识别率会急剧下降,因此更具成本效益且更准确的人耳识别方法,还需要进行大量的研究工作。The current methods of human ear recognition can be divided into two categories according to the extracted features: one is the method based on geometric features. Such methods are easily affected by illumination and imaging angle, and have poor robustness. One is the methods based on algebraic features, such as principal component analysis method, invariant moment III method, wavelet transform method and so on. These methods achieve satisfactory results under the condition that the human ear pose changes little and the image quality is good. However, when the rotation angle of the human ear changes, the two-dimensional image will bring about a large deformation. At this time, the recognition rate of the traditional method will drop sharply. Therefore, a more cost-effective and accurate human ear recognition method needs to be carried out A lot of research work.
发明内容SUMMARY OF THE INVENTION
技术问题:本发明所要解决的技术问题是如何对输入的人耳图像使用最小距离分类器进行人耳识别,以提高对人耳识别的训练速度和准确度。Technical problem: The technical problem to be solved by the present invention is how to use the minimum distance classifier to perform human ear recognition on the input human ear image, so as to improve the training speed and accuracy of human ear recognition.
技术方案:本发明的一种基于梯度方向直方图与局部二值模式融合的人耳识别方法,包括以下步骤:Technical solution: A human ear recognition method based on the fusion of gradient direction histogram and local binary pattern of the present invention includes the following steps:
步骤1)从人耳图像库中获取人耳的图像;Step 1) obtaining the image of the human ear from the human ear image library;
步骤2)计算人耳图像中每个像素上的特征值,分块并标准化后得到人耳图像的梯度方向直方图特征;Step 2) Calculate the feature value on each pixel in the human ear image, obtain the gradient direction histogram feature of the human ear image after dividing and standardizing;
步骤3)用主成分分析法对人耳图像的梯度方向直方图进行空间变换,使原来的坐标投影到一个新的维度较低的并且相互正交的空间上,实现对人耳图像的梯度方向直方图特征的降维;Step 3) Use the principal component analysis method to spatially transform the gradient direction histogram of the human ear image, so that the original coordinates are projected to a new space with lower dimensions and mutually orthogonal, so as to realize the gradient direction of the human ear image. Dimensionality reduction of histogram features;
步骤4)计算人耳图像上每个像素的局部二值模式值,得到人耳图像的局部二值模式特征;Step 4) calculating the local binary pattern value of each pixel on the human ear image to obtain the local binary pattern feature of the human ear image;
步骤5)将梯度方向直方图特征和局部二值模式特征的特征向量进行级联,得到新的特征向量,实现特征融合;Step 5) cascading the eigenvectors of the gradient direction histogram feature and the local binary pattern feature to obtain a new eigenvector to realize feature fusion;
步骤6)输入到最小距离分类器进行分类、识别。Step 6) Input to the minimum distance classifier for classification and identification.
其中,in,
所述步骤2)具体如下:Described step 2) is as follows:
步骤21)对步骤1)中获取的人耳图像进行颜色标准化处理,统一将图像转换成灰度图像,其转换公式为:H(x,y)=0.3*R(x,y)+0.59*G(x,y)+0.11*B(x,y),其中R(x,y)、G(x,y)、B(x,y)分别是图像中每个像素点的红、绿、蓝的色彩值,H(x,y)表示每个像素点的灰度值。Step 21) Perform color standardization on the human ear image obtained in step 1), and uniformly convert the image into a grayscale image. The conversion formula is: H(x,y)=0.3*R(x,y)+0.59* G(x,y)+0.11*B(x,y), where R(x,y), G(x,y), B(x,y) are the red, green, The color value of blue, H(x,y) represents the gray value of each pixel.
步骤22)使用Sobel算子计算图像各个像素梯度的模值和方向角:Step 22) use Sobel operator to calculate the modulus value and direction angle of each pixel gradient of the image:
其中,G(x,y)表示像素点的梯度幅值,α(x,y)表示像素点的梯度方向,H(x,y)表示像素点的灰度值。Among them, G(x, y) represents the gradient magnitude of the pixel point, α(x, y) represents the gradient direction of the pixel point, and H(x, y) represents the gray value of the pixel point.
步骤23)空间和方向单元格带权投票。首先计算方向权值:x(i)=cos(θ),y(i)=sin(θ),θ=θ+π/(Ndirection+1),其中:i为方向序号;θ为角度,初值为0;x(i)为x轴差分在i方向的权值;y(i)为y轴差分在i方向的权值,Ndirection为总的方向数,一般设置为9;Step 23) Space and orientation cells vote with weight. First calculate the direction weights: x(i)=cos(θ), y(i)=sin(θ), θ=θ+π/(N direction +1), where: i is the direction number; θ is the angle, The initial value is 0; x(i) is the weight of the x-axis difference in the i direction; y(i) is the weight of the y-axis difference in the i direction, and N direction is the total number of directions, generally set to 9;
步骤24)计算幅度和方向,幅度是x或y轴的图像差分的均方值,方向值取为各个方向上的带权最大值;Step 24) calculate the amplitude and direction, the amplitude is the mean square value of the image difference of the x or y axis, and the direction value is taken as the weighted maximum value in each direction;
步骤25)构建块并标准化,将单元格中的特征汇总组合成块,计算方法为: 其中B(x)、B(y)分别表示块x轴总数值和y轴总数值,C(x)、C(y)分别代表单元格x轴总数值和y轴总数值,B(size)为块的大小,B(step)是块变化的步长;Step 25) Building blocks and standardizing, summarizing and combining the features in the cells into blocks, the calculation method is: Among them, B(x) and B(y) represent the total value of the x-axis and the total value of the y-axis of the block respectively, C(x) and C(y) represent the total value of the x-axis and the total value of the y-axis of the cell, and B(size) is the size of the block, and B(step) is the step size of the block change;
步骤26)将不同方向和块上的特征值汇总,构建图像的方向梯度直方图。Step 26) Summarize the feature values in different directions and blocks, and construct the direction gradient histogram of the image.
所述步骤3)具体如下:Described step 3) is as follows:
步骤31)计算所有人耳图像中对应像素点的梯度方向直方图特征的均值 Step 31) Calculate the mean value of the gradient direction histogram feature of the corresponding pixel in the human ear image
步骤32)根据计算协方差矩阵,其中xi为需要降维的特征,UT为协方差矩阵。Step 32) According to Calculate the covariance matrix, where x i is the feature that needs to be reduced in dimension, and U T is the covariance matrix.
步骤33)取协方差矩阵前p个主成分,对人耳图像中每一个梯度方向直方图特征值进行特征降维,得到人耳图像的主成分分析法降维的梯度方向直方图特征,向量维数为p维。特征降维方法为其中y表示主成分特征。p根据实际情况通过实验获得,过大速度会慢,过小会影响准确率。Step 33) Take the first p principal components of the covariance matrix, and perform feature dimension reduction on the eigenvalues of each gradient direction histogram in the human ear image to obtain the gradient direction histogram feature of the dimension reduction by the principal component analysis method of the human ear image, the vector The dimension is p-dimension. The feature dimensionality reduction method is where y represents the principal component feature. p is obtained through experiments according to the actual situation. If it is too large, the speed will be slow, and if it is too small, the accuracy will be affected.
所述步骤4)具体如下:Described step 4) is as follows:
步骤41)定义3×3窗口的中心像素为阈值,将剩余8个像素的灰度值依次与阈值比较,大于中心像素点的标记为1,否则标记为0,组成8位二进制数即为此窗口的局部二值模式值的二进制表示。其中(xc,yc)代表3×3邻域的中心元素,它的像素值为ic,ip表示邻域内其他像素的值,s(x)是符号函数,当x≥0时为1,否则为0,LBP(xc,yc)是中心像素的局部二值模式值的二进制表示。Step 41) Define the central pixel of the 3×3 window as the threshold value, compare the gray values of the remaining 8 pixels with the threshold value in turn, and mark the pixel larger than the central pixel as 1, otherwise mark as 0, and form an 8-bit binary number. The binary representation of the local binary mode value of the window. where (x c , y c ) represents the central element of the 3×3 neighborhood, its pixel value is ic , ip represents the value of other pixels in the neighborhood, s(x) is the sign function, and when x≥0, it is 1, otherwise 0, LBP(x c , y c ) is the binary representation of the local binary pattern value of the center pixel.
步骤42)将每个像素的LBP(xc,yc)转换成十进制数得到最终的局部二值模式值,汇总后得到人耳图像的局部二值模式特征。Step 42) Convert the LBP(x c , y c ) of each pixel into decimal numbers to obtain the final local binary pattern value, and after summarizing the local binary pattern features of the human ear image.
所述步骤6)具体如下:Described step 6) is as follows:
将待识别人耳图像与若干个已知分类的人耳图像分别经过步骤2至5处理后得到的特征向量输入最小距离分类器中,分别计算待识别人耳图像的特征向量与各个已知分类的人耳图像的特征向量之间的距离,最小距离对应的分类即为待识别人耳图像的分类。距离计算公式如下:Input the eigenvectors obtained after the processing of steps 2 to 5 into the minimum distance classifier, respectively, to calculate the eigenvectors of the to-be-recognized human ear image and each known classification The distance between the feature vectors of the human ear images, and the classification corresponding to the smallest distance is the classification of the human ear image to be recognized. The distance calculation formula is as follows:
其中d(i)为待识别人耳图像与第i个已知分类的人耳图像之间的距离,t1(i)为第i个已知分类的人耳图像的第一特征向量,p1为待识别人耳图像的第一特征向量,tn(i)为第i个已知分类的图像的第n个特征向量,pn为待识别图像的第n个特征向量。where d(i) is the distance between the human ear image to be recognized and the i-th known-classified human ear image, t 1 (i) is the first feature vector of the i-th known-classified human ear image, p 1 is the first feature vector of the human ear image to be recognized, t n (i) is the n-th feature vector of the i-th known classified image, and p n is the n-th feature vector of the image to be recognized.
有益效果:本发明采用以上技术方案与现有技术相比,具有以下技术效果:Beneficial effect: the present invention adopts the above technical scheme compared with the prior art, has the following technical effects:
本发明使用了主成分分析法对梯度方向直方图特征进行降维,过滤掉了大量的冗余信息,大大提高了人耳识别的准确性,同时降低了特征向量的维数,提高了人耳识别的速度;本发明使用局部二值模式特征和梯度方向直方图特征进行融合,克服了一些噪声的干扰,日高了特征向量的鲁棒性,提高了人耳识别算法的稳定性,相比于使用单一特征进行识别,本发明识别率更高;本发明采用最小距离分类器,计算复杂度更低,速度更快。通过这些方法的应用,提高了人耳识别的准确性和稳定性,同时降低了计算复杂度,使系统具有较高的成本效益,具体来说:The invention uses the principal component analysis method to reduce the dimension of the gradient direction histogram feature, filters out a large amount of redundant information, greatly improves the accuracy of human ear recognition, reduces the dimension of the feature vector, and improves the human ear recognition. The speed of recognition; the invention uses the local binary pattern feature and the gradient direction histogram feature to fuse, overcomes some noise interference, increases the robustness of the feature vector, and improves the stability of the human ear recognition algorithm. The present invention has a higher recognition rate for using a single feature for identification; the present invention adopts a minimum distance classifier, which has lower computational complexity and faster speed. Through the application of these methods, the accuracy and stability of human ear recognition are improved, while the computational complexity is reduced, making the system highly cost-effective, specifically:
(1)本发明采用多特征融合进行分类识别,相比单一特征,具有更高的准确性。(1) The present invention adopts multi-feature fusion for classification and identification, which has higher accuracy than a single feature.
(2)本发明使用了主成分分析法对梯度方向直方图特征进行降维,过滤掉了大量的冗余信息,大大提高了人耳识别的准确性。(2) The present invention uses the principal component analysis method to reduce the dimension of the gradient direction histogram feature, filters out a large amount of redundant information, and greatly improves the accuracy of human ear recognition.
(3)本发明使用局部二值模式特征和梯度方向直方图特征,克服了一些噪声的干扰,日高了特征向量的鲁棒性,提高了人耳识别算法的稳定性。(3) The present invention uses the local binary pattern feature and the gradient direction histogram feature, overcomes some noise interference, increases the robustness of the feature vector, and improves the stability of the human ear recognition algorithm.
(4)本发明使用了主成分分析法对梯度方向直方图特征进行降维,相比传统的梯度方向直方图特征,降低了特征向量的维数,提高了整个人耳识别的速度。(4) The present invention uses the principal component analysis method to reduce the dimension of the gradient direction histogram feature. Compared with the traditional gradient direction histogram feature, the dimension of the feature vector is reduced, and the speed of the entire human ear recognition is improved.
(5)本发明采用的基于欧氏距离的最小距离分类器,相比其他分类器,计算复杂度更低,提高了人耳识别的速度。(5) Compared with other classifiers, the minimum distance classifier based on Euclidean distance adopted in the present invention has lower computational complexity and improves the speed of human ear recognition.
附图说明Description of drawings
图1是基于卷积神经网络的人体行为识别方法流程。Figure 1 is the flow of the human action recognition method based on convolutional neural network.
具体实施方式Detailed ways
在具体实施中,图1是基于卷积神经网络的人体行为识别方法流程。In a specific implementation, FIG. 1 is a flowchart of a method for recognizing human behavior based on a convolutional neural network.
本实例使用北京科技大学人耳实验库为实验对象,包含77人的人耳图像。人耳库中每个人耳有四张人耳图像,分别是:正常情况下人耳的正面图像、人耳+30度和-30度旋转的图像,人耳在光照变暗条件下的正面图像。This example uses the human ear experiment library of University of Science and Technology Beijing as the experimental object, which contains 77 human ear images. There are four human ear images for each human ear in the human ear library, namely: the frontal image of the human ear under normal conditions, the image rotated by +30 degrees and -30 degrees of the human ear, and the frontal image of the human ear under the condition of dim light. .
在具体实施中,每人有4幅人耳图像,其中3幅图像用于训练,1幅图像用于测试。In the specific implementation, each person has 4 human ear images, of which 3 images are used for training and 1 image is used for testing.
首先,将每人的3幅人耳图像输入到系统,进行颜色标准化处理,使用H(x,y)=0.3*R(x,y)+0.59*G(x,y)+0.11*B(x,y)将图像转换成灰度图像,其中R(x,y)、G(x,y)、B(x,y)分别是图像中每个像素点的红、绿、蓝的色彩值,H(x,y)表示每个像素点的灰度值;使用Sobel算子计算图像各个像素梯度的模值和方向角;计算方向权值、幅度和方向,幅度是x或y轴的图像差分的均方值,方向值取为各个方向上的带权最大值;构建块并标准化,将单元格中的特征汇总组合成块,计算方法为: 其中B(x)、B(y)分别表示块x轴总数值和y轴总数值,C(x)、C(y)分别代表单元格x轴总数值和y轴总数值,B(size)为块的大小,B(step)是块变化的步长;将不同方向和块上的特征值汇总,构建图像的方向梯度直方图。First, input 3 human ear images of each person into the system, and perform color normalization processing, using H(x,y)=0.3*R(x,y)+0.59*G(x,y)+0.11*B( x,y) converts the image into a grayscale image, where R(x,y), G(x,y), B(x,y) are the red, green, and blue color values of each pixel in the image, respectively , H(x,y) represents the gray value of each pixel; use the Sobel operator to calculate the modulus value and direction angle of the gradient of each pixel of the image; calculate the direction weight, magnitude and direction, the magnitude is the image of the x or y axis The mean square value of the difference, the direction value is taken as the weighted maximum value in each direction; the block is constructed and standardized, and the features in the cells are aggregated and combined into blocks. The calculation method is: Among them, B(x) and B(y) represent the total value of the x-axis and the total value of the y-axis of the block respectively, C(x) and C(y) represent the total value of the x-axis and the total value of the y-axis of the cell, and B(size) is the size of the block, and B(step) is the step size of the block change; the feature values in different directions and blocks are summarized to construct the directional gradient histogram of the image.
接着,用主成分分析法对人耳图像的梯度方向直方图进行空间变换,使原来的坐标投影到一个新的维度较低的并且相互正交的空间上,实现对人耳图像的梯度方向直方图特征的降维。具体方法是根据计算协方差矩阵,其中xi为需要降维的特征,为人耳图像梯度方向直方图特征的均值,UT为协方差矩阵。取协方差矩阵前p个主成分,对人耳图像中每一个梯度方向直方图特征值,用进行特征降维,得到人耳图像的主成分分析法降维的梯度方向直方图特征,其中向量维数为p维,y表示主成分特征。Then, use the principal component analysis method to spatially transform the gradient direction histogram of the human ear image, so that the original coordinates are projected to a new space with lower dimensions and orthogonal to each other, so as to realize the gradient direction histogram of the human ear image. Dimensionality reduction of graph features. The specific method is based on Calculate the covariance matrix, where x i is the feature that needs to be reduced in dimension, is the mean value of the histogram feature of the gradient direction of the human ear image, and U T is the covariance matrix. Take the first p principal components of the covariance matrix, and for each eigenvalue of the gradient direction histogram in the human ear image, use The feature dimension reduction is performed to obtain the gradient direction histogram feature of the principal component analysis method of the human ear image, wherein the vector dimension is p dimension, and y represents the principal component feature.
然后,定义3×3窗口的中心像素为阈值,将剩余8个像素的灰度值依次与阈值比较,大于中心像素点的标记为1,否则标记为0,组成8位二进制数即为此窗口的局部二值模式值的二进制表示。其中(xc,yc)代表3×3邻域的中心元素,它的像素值为ic,ip表示邻域内其他像素的值,s(x)是符号函数,当x≥0时为1,否则为0,LBP(xc,yc)是中心像素的局部二值模式值的二进制表示,再将其转换成十进制数得到最终的局部二值模式值,得到人耳图像的局部二值模式特征。Then, the central pixel of the 3×3 window is defined as the threshold, and the gray values of the remaining 8 pixels are compared with the threshold in turn. If the pixel is larger than the central pixel, it is marked as 1, otherwise it is marked as 0, and the 8-bit binary number is this window. The binary representation of the local binary mode value of . where (x c , y c ) represents the central element of the 3×3 neighborhood, its pixel value is ic , ip represents the value of other pixels in the neighborhood, s(x) is the sign function, and when x≥0, it is 1, otherwise 0, LBP(x c , y c ) is the binary representation of the local binary pattern value of the central pixel, and then convert it into a decimal number to obtain the final local binary pattern value, and obtain the local binary pattern of the human ear image. Value mode feature.
最后,将梯度方向直方图特征和局部二值模式特征的特征向量进行级联,得到新的特征向量,实现特征融合,将其输入到最小距离分类器进行分类、识别。Finally, the gradient direction histogram feature and the feature vector of the local binary pattern feature are cascaded to obtain a new feature vector to realize feature fusion, and input it to the minimum distance classifier for classification and identification.
具体方法为:将待识别人耳图像与若干个已知分类的人耳图像分别经过上述步骤处理后得到的特征向量输入最小距离分类器中,分别计算待识别人耳图像的特征向量与各个已知分类的人耳图像的特征向量之间的距离,最小距离对应的分类即为待识别人耳图像的分类。距离计算公式如下:The specific method is as follows: input the feature vector obtained by processing the above-mentioned steps of the human ear image to be recognized and several known classified human ear images respectively into the minimum distance classifier, and calculate the feature vector of the human ear image to be recognized and each of the recognized human ear images respectively. The distance between the feature vectors of the classified human ear images is known, and the classification corresponding to the smallest distance is the classification of the human ear image to be recognized. The distance calculation formula is as follows:
其中d(i)为待识别人耳图像与第i个已知分类的人耳图像之间的距离,t1(i)为第i个已知分类的人耳图像的第一特征向量,p1为待识别人耳图像的第一特征向量,tn(i)为第i个已知分类的图像的第n个特征向量,pn为待识别图像的第n个特征向量。where d(i) is the distance between the human ear image to be recognized and the i-th known-classified human ear image, t 1 (i) is the first feature vector of the i-th known-classified human ear image, p 1 is the first feature vector of the human ear image to be recognized, t n (i) is the n-th feature vector of the i-th known classified image, and p n is the n-th feature vector of the image to be recognized.
以上所述仅是本发明的优选实施方式,应当指出:对于本技术领域的普通技术人员来说,在不脱离本发明原理的前提下,还可以做出若干改进和润饰,这些改进和润饰也应视为本发明的保护范围。The above is only the preferred embodiment of the present invention, it should be pointed out: for those skilled in the art, under the premise of not departing from the principle of the present invention, several improvements and modifications can also be made, and these improvements and modifications are also It should be regarded as the protection scope of the present invention.
Claims (3)
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910433620.7A CN110188646B (en) | 2019-05-23 | 2019-05-23 | Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910433620.7A CN110188646B (en) | 2019-05-23 | 2019-05-23 | Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern |
Publications (2)
Publication Number | Publication Date |
---|---|
CN110188646A CN110188646A (en) | 2019-08-30 |
CN110188646B true CN110188646B (en) | 2022-08-02 |
Family
ID=67717439
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910433620.7A Active CN110188646B (en) | 2019-05-23 | 2019-05-23 | Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN110188646B (en) |
Families Citing this family (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111461147B (en) * | 2020-04-30 | 2023-05-23 | 柳州智视科技有限公司 | Binary coding organization algorithm based on image features |
CN111967531B (en) * | 2020-08-28 | 2022-09-16 | 南京邮电大学 | A high-precision indoor image localization method based on multi-feature fusion |
CN115547475A (en) * | 2022-12-05 | 2022-12-30 | 医修技术服务(北京)有限公司 | Intelligent automatic control method based on weighing type medical consumable management cabinet |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106599870A (en) * | 2016-12-22 | 2017-04-26 | 山东大学 | Face recognition method based on adaptive weighting and local characteristic fusion |
CN107066958A (en) * | 2017-03-29 | 2017-08-18 | 南京邮电大学 | A kind of face identification method based on HOG features and SVM multi-categorizers |
CN107578007A (en) * | 2017-09-01 | 2018-01-12 | 杭州电子科技大学 | A deep learning face recognition method based on multi-feature fusion |
CN108549868A (en) * | 2018-04-12 | 2018-09-18 | 中国矿业大学 | A kind of pedestrian detection method |
-
2019
- 2019-05-23 CN CN201910433620.7A patent/CN110188646B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN106599870A (en) * | 2016-12-22 | 2017-04-26 | 山东大学 | Face recognition method based on adaptive weighting and local characteristic fusion |
CN107066958A (en) * | 2017-03-29 | 2017-08-18 | 南京邮电大学 | A kind of face identification method based on HOG features and SVM multi-categorizers |
CN107578007A (en) * | 2017-09-01 | 2018-01-12 | 杭州电子科技大学 | A deep learning face recognition method based on multi-feature fusion |
CN108549868A (en) * | 2018-04-12 | 2018-09-18 | 中国矿业大学 | A kind of pedestrian detection method |
Also Published As
Publication number | Publication date |
---|---|
CN110188646A (en) | 2019-08-30 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Sun et al. | Deep learning face representation by joint identification-verification | |
CN102663413B (en) | Multi-gesture and cross-age oriented face image authentication method | |
CN106355138A (en) | Face recognition method based on deep learning and key features extraction | |
CN100426314C (en) | Feature classification based multiple classifiers combined people face recognition method | |
CN104361313B (en) | A kind of gesture identification method merged based on Multiple Kernel Learning heterogeneous characteristic | |
CN102332084B (en) | Identity identification method based on palm print and human face feature extraction | |
CN109902590A (en) | Pedestrian re-identification method based on distance learning of deep multi-view features | |
CN110188646B (en) | Human ear recognition method based on the fusion of gradient direction histogram and local binary pattern | |
CN112560858B (en) | Character and picture detection and rapid matching method combining lightweight network and personalized feature extraction | |
CN101968813A (en) | Method for detecting counterfeit webpage | |
CN107292299B (en) | Side face recognition methods based on kernel specification correlation analysis | |
CN109344856B (en) | Offline signature identification method based on multilayer discriminant feature learning | |
CN103870811A (en) | Method for quickly recognizing front face through video monitoring | |
CN104156690B (en) | A kind of gesture identification method based on image space pyramid feature bag | |
CN106971158A (en) | A kind of pedestrian detection method based on CoLBP symbiosis feature Yu GSS features | |
CN107103266A (en) | The training of two-dimension human face fraud detection grader and face fraud detection method | |
CN110555386A (en) | Face recognition identity authentication method based on dynamic Bayes | |
CN111832405A (en) | A face recognition method based on HOG and deep residual network | |
CN107784263B (en) | Planar rotation face detection method based on improved accelerated robust features | |
CN106682653A (en) | KNLDA-based RBF neural network face recognition method | |
CN103942545A (en) | Method and device for identifying faces based on bidirectional compressed data space dimension reduction | |
CN103942572A (en) | Method and device for extracting facial expression features based on bidirectional compressed data space dimension reduction | |
CN106599778A (en) | Rapid accurate human face detection method | |
CN104077608B (en) | Activity recognition method based on the slow characteristic function of sparse coding | |
CN114549267A (en) | Zero watermarking method for multi-feature clustering |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |