CN110969089A - Lightweight face recognition system and recognition method under noise environment - Google Patents

Lightweight face recognition system and recognition method under noise environment Download PDF

Info

Publication number
CN110969089A
CN110969089A CN201911059976.5A CN201911059976A CN110969089A CN 110969089 A CN110969089 A CN 110969089A CN 201911059976 A CN201911059976 A CN 201911059976A CN 110969089 A CN110969089 A CN 110969089A
Authority
CN
China
Prior art keywords
denoising
features
face image
face recognition
convolution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911059976.5A
Other languages
Chinese (zh)
Other versions
CN110969089B (en
Inventor
白慧慧
郭璐璐
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Jiaotong University
Original Assignee
Beijing Jiaotong University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Jiaotong University filed Critical Beijing Jiaotong University
Priority to CN201911059976.5A priority Critical patent/CN110969089B/en
Publication of CN110969089A publication Critical patent/CN110969089A/en
Application granted granted Critical
Publication of CN110969089B publication Critical patent/CN110969089B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D10/00Energy efficient computing, e.g. low power processors, power management or thermal management

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • General Physics & Mathematics (AREA)
  • Human Computer Interaction (AREA)
  • Multimedia (AREA)
  • Biomedical Technology (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a lightweight face recognition system and a recognition method under a noise environment, which belong to the technical field of computer face recognition, firstly, a noise picture is input, a 3 x 3 convolution kernel is used for extracting features, then a denoising block is added, the features extracted by a convolution layer are denoised, then a depth separable convolution and ten layers of bottleneck structures are used, and finally, output features are obtained through 1 x 1 convolution and linear global depth separable convolution. The invention provides an effective independent de-noising module-LD-MobileFaceNet which can be conveniently combined with any convolutional neural network structure; in the denoising operation, a non-local mean denoising method is adopted to improve the performance of the MobileFaceNet and improve the accuracy of face recognition in a noise environment; a part of bottleneck layers in the network are removed, so that the system is lighter, and better robustness is realized on a noise data set; the nonlinear activation function swish replaces PReLU, loss of identification precision is made up, and identification accuracy under different noise levels is guaranteed.

Description

Lightweight face recognition system and recognition method under noise environment
Technical Field
The invention relates to the technical field of computer face recognition, in particular to a lightweight face recognition system and a recognition method in a noise environment.
Background
Nowadays, face recognition has been widely applied to the fields of face login, mobile payment, identity authentication, and the like. These application scenarios often require high accuracy, however, face recognition still faces a great challenge due to interference of complex environments such as occlusion, illumination, noise, and the like. With the development of artificial intelligence, deep learning has made a major breakthrough in face recognition research, and some high-performance methods based on convolutional neural networks are proposed.
For example, deep face achieves high accuracy on a Labeled Wild face dataset (LFW). The appearance of triple loss allows further development of face recognition research, and the method considers the relative difference of the distances between matched pairs and non-matched pairs and can extract better features. The angular margin Loss (ArcFace Loss) utilizes the arc-cos function, an additional angular edge is added to obtain a target, and therefore accurate correspondence between the angle and the circular arc is achieved on the normalized hypersphere.
The above identification methods based on convolutional neural networks all provide more distinctive features for face identification, but some practical problems also exist. The deep neural network based on deep learning has a large number of parameters and layers, and occupies a large memory space.
For example, when comparing the deep face with the classical image classification task, the human face has strong similarity, simple features cannot accurately distinguish the human face, and a proper loss function needs to be designed to improve the recognition capability. The model size of an ArcFace (LResNet100E-IR) network reaches 250 megabits. In practical face recognition applications, the face recognition network is often used in mobile terminals and embedded devices, and due to the limitations of memory and computing resources, the large face recognition network is difficult to be widely applied.
Therefore, a lightweight convolutional neural network is produced. Such as MobileNetV1, MobileNetV2, which are widely used in engineering and industrial production. The MoileNetV2 model is designed by using structures such as depth separable convolution, inverse residual error, linear bottleneck and the like, so that an efficient and lightweight model is realized. The MobileFaceNet adopts an ArcFace Loss function, and improves the mobileNet V2 algorithm, so that the face recognition speed is improved by more than two times, and meanwhile, the accuracy of an LFW data set reaches 99.55%.
However, the above-mentioned real-time face recognition based on the lightweight convolutional neural network often faces complicated background, such as noise, occlusion and insufficient contrast, which all affect the application of the face recognition network model in practice. In order to apply the network model to our real life more widely, it is indispensable to improve the robustness of the algorithm in a complicated and varied environment. The common method for improving the robustness of the algorithm is to add a denoising network, and remove image noise and clean data in an image preprocessing stage. For example, a clear image is output by using an end-to-end denoising network, and the method increases a large number of parameters and makes the network structure more complex. But is not applicable in real-time face recognition when used in the image pre-processing stage.
Disclosure of Invention
The invention aims to provide a lightweight face recognition system and a recognition method in a noise environment, so as to solve the technical problems in the background technology.
In order to achieve the purpose, the invention adopts the following technical scheme:
in one aspect, the present invention provides a lightweight face recognition system in a noise environment, including:
the image acquisition module is used for acquiring an original face image to be recognized;
the characteristic extraction module is used for extracting the identification characteristics of the collected original face image;
the denoising module is used for denoising the face image with the extracted identification features;
the depth separable convolution module is used for carrying out channel correspondence on the face image subjected to the denoising processing and the original face image;
and the characteristic output module is used for outputting the final identification characteristic.
Preferably, the feature extraction module is a 3 × 3 convolutional layer.
Preferably, the bottle neck structure comprises ten layers.
Preferably, the film further comprises a 1 × 1 convolutional layer.
Preferably, the feature output module comprises a linear global depth separable convolutional layer.
In another aspect, the present invention provides a method for performing lightweight face recognition in a noise environment by using the system described above, including:
extracting identification characteristics of an input original face image with noise;
carrying out feature denoising processing on the extracted identification features;
performing channel correspondence on the face image subjected to noise removal processing and the original face image;
and sequentially obtaining the final identification features through ten layers of bottleneck structures, 1 × 1 convolution and linear global depth separable convolution.
Preferably, the characteristic drying treatment includes:
matching the images by combining the Euclidean distance and a weighted subregion matching method, and reflecting local and global characteristics of the images by the calculated similarity, wherein the expression is as follows:
Figure BDA0002257657410000031
in formula (1), i is an index of an output position, j represents all possible positions, x is an input original face image, y is a denoised feature map with the same pixels as those corresponding to x,
Figure BDA0002257657410000032
is the feature obtained by the down-sampling operation of x, w is a gaussian function,
Figure BDA0002257657410000033
is a normalization operation, S denotes all spatial positions,
Figure BDA0002257657410000034
representing the mapping of input x to output y.
Preferably, the first and second liquid crystal materials are,
Figure BDA0002257657410000035
wherein e is an exponential function, x is an input pixel, T is a transpose matrix of x, and the estimation value of the current pixel is obtained by weighted averaging of neighborhood pixels through a Gaussian function.
Preferably, the activation function in the bottleneck structure is a swish activation function:
f(x)=x·sigmoid(x) (3)。
preferably, the penalty function for the linear global depth separable convolutional layer is Arcfacelos:
Figure BDA0002257657410000041
the loss function is expressed as L1:θjIs the angle between the current weight and the target feature, N is a batch parameter, yiFor the class to which the input features belong,
Figure BDA0002257657410000042
and representing an included angle between the input features and the real weight, n representing the number of the total classes, t representing an additional angle margin penalty to enhance the intra-class compactness and the inter-class difference, and all the extracted face features are distributed on a space with the radius of r.
The invention has the beneficial effects that: an effective independent de-noising module-LD-MobileFaceNet which can be conveniently combined with any convolutional neural network structure is provided; in the denoising operation, a non-local mean denoising method is adopted to improve the performance of the MobileFaceNet and improve the accuracy of face recognition in a noise environment; a part of bottleneck layers in the network are removed, so that the system is lighter, and better robustness is realized on a noise data set; the nonlinear activation function swish is used for replacing the PReLU, loss of identification precision is made up, and identification accuracy under different noise levels is guaranteed.
Additional aspects and advantages of the invention will be set forth in part in the description which follows, and in part will be obvious from the description, or may be learned by practice of the invention.
Drawings
In order to more clearly illustrate the technical solutions of the embodiments of the present invention, the drawings needed to be used in the description of the embodiments are briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic block diagram of a lightweight face recognition system in a noise environment according to an embodiment of the present invention.
Fig. 2 is a schematic block diagram of a drying module according to an embodiment of the present invention.
FIG. 3 is a graph comparing performance of LD-MobileFaceNet (swish) and MobileFaceNet according to the embodiment of the present invention.
Fig. 4 is a performance comparison graph of different recognition methods using different loss functions according to the embodiment of the present invention.
Fig. 5 is a schematic diagram illustrating the recognition accuracy comparison of three models, namely D-MobileFaceNet (4.0MB), LD-MobileFaceNet (3.0MB) and LD-MobileFaceNet (swish) (3.0MB), with an original picture size of 112 × 112 according to an embodiment of the present invention.
Detailed Description
The embodiments described below with reference to the accompanying drawings are illustrative only for the purpose of explaining the present invention, and are not to be construed as limiting the present invention.
As used herein, the singular forms "a", "an", "the" and "the" are intended to include the plural forms as well, unless the context clearly indicates otherwise. It will be further understood that the terms "comprises" and/or "comprising," when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or modules, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, modules, and/or groups thereof.
It will be understood by those skilled in the art that, unless otherwise defined, all terms (including technical and scientific terms) used herein have the same meaning as commonly understood by one of ordinary skill in the art to which this invention belongs. It will be further understood that terms, such as those defined in commonly used dictionaries, should be interpreted as having a meaning that is consistent with their meaning in the context of the prior art and will not be interpreted in an idealized or overly formal sense unless expressly so defined herein.
For the convenience of understanding of the embodiments of the present invention, the following description will be further explained by taking specific embodiments as examples with reference to the drawings, and the embodiments are not to be construed as limiting the embodiments of the present invention.
It will be understood by those of ordinary skill in the art that the figures are merely schematic representations of one embodiment and that the elements or devices in the figures are not necessarily required to practice the present invention.
Example 1
An embodiment 1 of the present invention provides a lightweight face recognition system in a noise environment, including:
the image acquisition module is used for acquiring an original face image to be recognized;
the characteristic extraction module is used for extracting the identification characteristics of the collected original face image;
the denoising module is used for denoising the face image with the extracted identification features;
the depth separable convolution module is used for carrying out channel correspondence on the face image subjected to the denoising processing and the original face image;
and the characteristic output module is used for outputting the final identification characteristic.
The feature extraction module is a 3 × 3 convolutional layer.
Also includes ten layers of bottleneck structure.
Also included are 1 × 1 convolutional layers.
The feature output module includes a linear global depth separable convolutional layer.
When the system is used for carrying out lightweight face recognition in a noise environment, extracting recognition features of an input original face image with noise; carrying out feature denoising processing on the extracted identification features; performing channel correspondence on the face image subjected to noise removal processing and the original face image; and sequentially obtaining the final identification features through ten layers of bottleneck structures, 1 × 1 convolution and linear global depth separable convolution.
The characteristic drying treatment comprises the following steps:
matching the images by combining the Euclidean distance and a weighted subregion matching method, and reflecting local and global characteristics of the images by the calculated similarity, wherein the expression is as follows:
Figure BDA0002257657410000061
in formula (1), i is an index of an output position, j represents all possible positions, x is an input original face image, y is a denoised feature map with the same pixels as those corresponding to x,
Figure BDA0002257657410000062
is the feature obtained by the down-sampling operation of x, w is a gaussian function,
Figure BDA0002257657410000063
is a normalization operation, S denotes all spatial positions,
Figure BDA0002257657410000064
representing the mapping of input x to output y.
Figure BDA0002257657410000065
Wherein e is an exponential function, x is an input pixel, T is a transpose matrix of x, and the estimation value of the current pixel is obtained by weighted averaging of neighborhood pixels through a Gaussian function.
The activation function in the bottleneck structure is the swish activation function:
f(x)=x·sigmoid(x) (3)。
the penalty function for a linear global depth separable convolutional layer is Arcfacelos, denoted L1
Figure BDA0002257657410000071
θjIs the angle between the current weight and the target feature, N is a batch parameter, yiFor the class to which the input features belong,
Figure BDA0002257657410000072
and representing an included angle between the input features and the real weight, n representing the number of the total classes, t representing an additional angle margin penalty to enhance the intra-class compactness and the inter-class difference, and all the extracted face features are distributed on a space with the radius of r.
Example 2
As shown in fig. 1, a structure diagram of an LD-mobilefacenet (swish) network proposed in embodiment 2 of the present invention. Firstly, inputting a noise picture, extracting features by using a 3 multiplied by 3 convolution kernel, then adding a denoising block, carrying out feature denoising on the features extracted by the convolution layer, then using a depth separable convolution and a ten-layer bottleneck structure, and finally obtaining output features through 1 multiplied by 1 convolution and linear global depth separable convolution, so that a network can learn better, and the accuracy of face matching is improved.
In embodiment 2 of the present invention, the idea of a non-local mean algorithm is used, taking into account every position information in space. A denoising module is defined, and denoising representation of extracting features is realized as follows:
Figure BDA0002257657410000073
in formula (1), i is an index of an output position, j represents all possible positions, x is an input original face image, y is a denoised feature map with the same pixels as those corresponding to x,
Figure BDA0002257657410000074
is the feature obtained by the down-sampling operation of x, w is a gaussian function,
Figure BDA0002257657410000075
is a normalization operation, S denotes all spatial positions,
Figure BDA0002257657410000081
representing the mapping of input x to output y.
The formula (1) is combined with the Euclidean distance and weighted subregion matching method to match the images, and the similarity obtained by calculation can reflect the local and global characteristics of the images.
Equation (2) is a gaussian equation for smooth image denoising. For images with random noise, the function can achieve a good denoising effect.
Figure BDA0002257657410000082
Wherein e is an exponential function, x is an input pixel, T is a transpose matrix of x, and the estimation value of the current pixel is obtained by weighted averaging of neighborhood pixels through a Gaussian function.
As shown in fig. 2, a structure of the denoising module is shown, in the graph, the size of an input feature graph x is 64 channels, maximum pooling layers are added at positions a and b, the size of the picture is reduced, the calculation amount is reduced, and the denoising module is realized by using a formula (2) through matrix multiplication. Through convolution of 1 multiplied by 1, a denoised characteristic diagram with an output channel consistent with an input channel can be obtained. This operation keeps the input and output channels unchanged, so the denoising block can be applied to any layer without affecting the network structure. Because the denoising operation can remove part of useful information at the same time, more original features can be stored by using residual concatenation.
In embodiment 2 of the present invention, in consideration of the compatibility of practical applications, a lightweight robust face recognition network LD-mobilefacenet (swish) is designed based on the above denoising module. Firstly, extracting features through a 3 multiplied by 3 convolution kernel with the step length of 2, and then carrying out feature denoising by using a denoising block. The deep separable convolution is a building block of MobileNetV1 that is used to reduce computational expense.
In embodiment 2 of the present invention, in order to reduce the weight, a part of redundant layers is deleted, and only a bottleneck structure of ten layers is used. In a bottleneck structure, adopting swish activation function to replace PReLU as a nonlinear activation function to improve the precision of a neural network, and defining the function as follows:
f(x)=x·sigmoid(x) (3)。
the swish function uses a sigmoid method, and only a simple scalar needs to be input by using self-gating. It is a smooth non-monotonic function, and experiments show that the performance is superior to that of ReLU.
Finally, the features are output by using a linear global depth separable convolution and a linear 1 x 1 convolution.
In embodiment 2 of the invention, Arcface loss with better effect is used as a loss function, and the characteristics of better distinction are extracted, wherein the loss function is expressed as L1
Figure BDA0002257657410000091
θjIs the angle between the current weight and the target feature, N is a batch parameter, yiFor the class to which the input features belong,
Figure BDA0002257657410000092
and representing an included angle between the input features and the real weight, n representing the number of the total classes, t representing an additional angle margin penalty to enhance the intra-class compactness and the inter-class difference, and all the extracted face features are distributed on a space with the radius of r.
Example 3
In embodiment 3 of the present invention, a denoising module is used to train on a CASIA-Webface data set with a noise level of 25.
The LFW data set with different noise levels is used as a test, 13233 pictures are contained in the data set, 6000 pairs of human faces are randomly selected to form a human face recognition picture pair, 3000 pairs of human face pictures belong to the same person, and 3000 pairs of human face pictures belong to different persons, wherein 1 human face picture of each person is tested and verified.
The data set processed in the experiment is added with white gaussian noise, noise with the noise level of 25 is added to the training set, and the test is carried out on the test sets with the noise levels of 0,15,25,35 and 50. The size of the input picture is 112 × 96, a human face is detected through the MTCNN, and a human face alignment operation is performed according to the marked five feature points. Also for better performance, a 112 × 112 dimensional training test was used. During training, the SGD optimizer was used with a batch size of 256 and a momentum parameter of 0.9. The initial learning rate was set to 0.1, decreasing by a factor of 0.1 for 20 iterations, for a total of 70 iterations.
Table 1 is a performance comparison table of the denoising model according to the embodiment of the present invention and CosFace, ArcFace, and MobileFaceNet, where table 1 is to denoise the features extracted from the first convolutional layer. The method is characterized in that the MobileFaceNet is used as a basic model, the L-MobileFaceNet is used as a bottleneck layer with 10 layers, the lightweight of the network is realized, and the D-MobileFaceNet is used for adding a denoising module on the basis of the basic model and comparing the denoising module. LD-MobileFaceNet adds denoising block on L-MobileFaceNet model, LD-MobileFaceNet (swish) further uses swish nonlinear activation function to replace PReLU activation function. These models were trained on a 112 × 96 CASIA-Webface dataset with a noise level of 25, and tested on LFW datasets with different noise levels.
Table 1: the denoising model was compared to the performance of CosFace, ArcFace, MobileFaceNet. Models using noise blocks were all trained on the CASIA-Webface dataset (σ 25, size 112 × 96) and tested on LFW datasets of different noise levels.
Figure BDA0002257657410000101
Compared with the table 1, the effectiveness of the denoising block can be obviously seen, and the swish activation function has certain help for improving the precision.
Table 2 shows the comparison results of different positions and numbers of the denoising blocks, wherein the denoising blocks are used after the first layer of convolution layer by Model A, the denoising blocks are added after the depth separable convolution layer by Model B, and the denoising blocks are added at two positions by Model C. It can be seen from table 2 that the effect of the different positions and numbers of the denoising blocks on the result is not too large, but a certain amount of calculation and parameters are increased by using a plurality of denoising blocks, and the performance of adding the denoising blocks at the front position is better. Meanwhile, as shown in fig. 3, which is a performance comparison graph of LD-MobileFaceNet (swish) and MobileFaceNet, a comparative experiment was performed on a 112 × 112 data set, and the generalization of the method according to the embodiment of the present invention was demonstrated. Fig. 3(a) shows a performance comparison graph in which the pixel size of the original image is 112 × 96, and fig. 3(b) shows a performance comparison graph in which the pixel size of the original image is 112 × 112.
TABLE 2
Figure BDA0002257657410000111
For fair comparison of the denoising models according to the embodiments of the present invention, different loss functions are used to test the performance of LD-mobilefacenet (swish), as shown in fig. 4, where ArcFace loss, CosFace loss, spheerface loss, and Softmax loss are compared, respectively. Wherein the performance of ArcFace loss and CosFace loss is better. FIG. 5 is a visual display of the accuracy of the three models D-MobileFaceNet, LD-MobileFaceNet and LD-MobileFaceNet (swish) on a 112X 112 data set.
Therefore, the denoising module provided by the embodiment of the invention effectively improves the identification precision. LD-MobileFaceNet (swish) shows better robustness under the noise environment, obviously surpasses MobileFaceNet, and its model is smaller, so the light-weight, robust face recognition network has more practical application meaning.
The above description is only for the preferred embodiment of the present invention, but the scope of the present invention is not limited thereto, and any changes or substitutions that can be easily conceived by those skilled in the art within the technical scope of the present invention are included in the scope of the present invention. Therefore, the protection scope of the present invention shall be subject to the protection scope of the claims.

Claims (10)

1. A lightweight face recognition system in a noisy environment, comprising:
the image acquisition module is used for acquiring an original face image to be recognized;
the characteristic extraction module is used for extracting the identification characteristics of the collected original face image;
the denoising module is used for denoising the face image with the extracted identification features;
the depth separable convolution module is used for carrying out channel correspondence on the face image subjected to the denoising processing and the original face image;
and the characteristic output module is used for outputting the final identification characteristic.
2. The system of claim 1, wherein the feature extraction module is a 3 x 3 convolutional layer.
3. The system of claim 1, further comprising a ten-layer bottleneck structure.
4. The lightweight face recognition system in a noisy environment of claim 2, further comprising a 1 x 1 convolutional layer.
5. The lightweight face recognition system in a noisy environment according to claim 4, wherein said feature output module comprises a linear global depth separable convolutional layer.
6. A method for lightweight face recognition in a noisy environment using a system according to any of claims 1-5, characterized by:
extracting identification characteristics of an input original face image with noise;
carrying out feature denoising processing on the extracted identification features;
performing channel correspondence on the face image subjected to noise removal processing and the original face image;
and sequentially obtaining the final identification features through ten layers of bottleneck structures, 1 × 1 convolution and linear global depth separable convolution.
7. The method of claim 6, wherein:
the characteristic drying treatment comprises the following steps:
matching the images by combining the Euclidean distance and a weighted subregion matching method, and reflecting local and global characteristics of the images by the calculated similarity, wherein the expression is as follows:
Figure FDA0002257657400000021
in formula (1), i is an index of an output position, j represents all possible positions, x is an input original face image, y is a denoised feature map with the same pixels as those corresponding to x,
Figure FDA0002257657400000022
is the feature obtained by the down-sampling operation of x, w is a gaussian function,
Figure FDA0002257657400000023
is a normalization operation, S denotes all spatial positions,
Figure FDA0002257657400000024
representing the mapping of input x to output y.
8. The method of claim 7, wherein:
Figure FDA0002257657400000025
wherein e is an exponential function, x is an input pixel, T is a transpose matrix of x, and the estimation value of the current pixel is obtained by weighted averaging of neighborhood pixels through a Gaussian function.
9. The method of claim 8, wherein:
the activation function in the bottleneck structure is the swish activation function:
f(x)=x·sigmoid(x) (3)。
10. the method of claim 9, wherein:
the penalty function for a linear global depth separable convolutional layer is Arcfacelos, denoted L1
Figure FDA0002257657400000026
θjIs the angle between the current weight and the target feature, N is a batch parameter, yiFor the class to which the input features belong,
Figure FDA0002257657400000027
and representing an included angle between the input features and the real weight, n representing the total number of classes to which the input features belong, t representing the penalty of adding angle margin, and all the extracted face features are distributed on a space with the radius of r.
CN201911059976.5A 2019-11-01 2019-11-01 Lightweight face recognition system and recognition method in noise environment Active CN110969089B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911059976.5A CN110969089B (en) 2019-11-01 2019-11-01 Lightweight face recognition system and recognition method in noise environment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911059976.5A CN110969089B (en) 2019-11-01 2019-11-01 Lightweight face recognition system and recognition method in noise environment

Publications (2)

Publication Number Publication Date
CN110969089A true CN110969089A (en) 2020-04-07
CN110969089B CN110969089B (en) 2023-08-18

Family

ID=70029993

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911059976.5A Active CN110969089B (en) 2019-11-01 2019-11-01 Lightweight face recognition system and recognition method in noise environment

Country Status (1)

Country Link
CN (1) CN110969089B (en)

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111563858A (en) * 2020-05-14 2020-08-21 大连理工大学 Denoising method of human embryo heart ultrasonic image based on deep convolutional neural network
CN111680595A (en) * 2020-05-29 2020-09-18 新疆爱华盈通信息技术有限公司 Face recognition method and device and electronic equipment
CN111898413A (en) * 2020-06-16 2020-11-06 深圳市雄帝科技股份有限公司 Face recognition method, face recognition device, electronic equipment and medium
CN112115911A (en) * 2020-09-28 2020-12-22 安徽大学 Light-weight SAR image target detection method based on deep learning
CN112733665A (en) * 2020-12-31 2021-04-30 中科院微电子研究所南京智能技术研究院 Face recognition method and system based on lightweight network structure design
CN113343801A (en) * 2021-05-26 2021-09-03 郑州大学 Automatic wireless signal modulation and identification method based on lightweight convolutional neural network

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016110005A1 (en) * 2015-01-07 2016-07-14 深圳市唯特视科技有限公司 Gray level and depth information based multi-layer fusion multi-modal face recognition device and method
US20180137406A1 (en) * 2016-11-15 2018-05-17 Google Inc. Efficient Convolutional Neural Networks and Techniques to Reduce Associated Computational Costs
CN109063666A (en) * 2018-08-14 2018-12-21 电子科技大学 The lightweight face identification method and system of convolution are separated based on depth
CN109800795A (en) * 2018-12-29 2019-05-24 广州市贺氏办公设备有限公司 A kind of fruit and vegetable recognition method and system
CN109948573A (en) * 2019-03-27 2019-06-28 厦门大学 A kind of noise robustness face identification method based on cascade deep convolutional neural networks

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2016110005A1 (en) * 2015-01-07 2016-07-14 深圳市唯特视科技有限公司 Gray level and depth information based multi-layer fusion multi-modal face recognition device and method
US20180137406A1 (en) * 2016-11-15 2018-05-17 Google Inc. Efficient Convolutional Neural Networks and Techniques to Reduce Associated Computational Costs
CN109063666A (en) * 2018-08-14 2018-12-21 电子科技大学 The lightweight face identification method and system of convolution are separated based on depth
CN109800795A (en) * 2018-12-29 2019-05-24 广州市贺氏办公设备有限公司 A kind of fruit and vegetable recognition method and system
CN109948573A (en) * 2019-03-27 2019-06-28 厦门大学 A kind of noise robustness face identification method based on cascade deep convolutional neural networks

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
LULU GUO等: "A Lightweight and Robust Face Recognition Network on Noisy Condition", 《PROCEEDINGS OF APSIPA ANNUAL SUMMIT AND CONFERENCE 2019》, pages 1964 - 1968 *
YANPENG SUN等: "An Object Detection Network for Embedded System", 《2019 IEEE INTERNATIONAL CONFERENCE ON UBIQUITOUS COMPUTING & COMMUNICATIONS (IUCC) AND DATA SCIENCE AND COMPUTATIONAL INTELLIGENCE (DSCI) AND SMART COMPUTING, NETWORKING AND SERVICES (SMARTCNS)》》 *
江琨;郑禄;帖军;: "基于LBP和PCA算法的智能人脸识别", 计算机与数字工程, no. 10 *

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111563858A (en) * 2020-05-14 2020-08-21 大连理工大学 Denoising method of human embryo heart ultrasonic image based on deep convolutional neural network
CN111563858B (en) * 2020-05-14 2023-08-22 大连理工大学 Denoising method of human embryo heart ultrasonic image based on depth convolution neural network
CN111680595A (en) * 2020-05-29 2020-09-18 新疆爱华盈通信息技术有限公司 Face recognition method and device and electronic equipment
CN111898413A (en) * 2020-06-16 2020-11-06 深圳市雄帝科技股份有限公司 Face recognition method, face recognition device, electronic equipment and medium
CN112115911A (en) * 2020-09-28 2020-12-22 安徽大学 Light-weight SAR image target detection method based on deep learning
CN112733665A (en) * 2020-12-31 2021-04-30 中科院微电子研究所南京智能技术研究院 Face recognition method and system based on lightweight network structure design
CN112733665B (en) * 2020-12-31 2024-05-28 中科南京智能技术研究院 Face recognition method and system based on lightweight network structure design
CN113343801A (en) * 2021-05-26 2021-09-03 郑州大学 Automatic wireless signal modulation and identification method based on lightweight convolutional neural network
CN113343801B (en) * 2021-05-26 2022-09-30 郑州大学 Automatic wireless signal modulation and identification method based on lightweight convolutional neural network

Also Published As

Publication number Publication date
CN110969089B (en) 2023-08-18

Similar Documents

Publication Publication Date Title
Tian et al. Deep learning on image denoising: An overview
CN110969089A (en) Lightweight face recognition system and recognition method under noise environment
Yang et al. BM3D-Net: A convolutional neural network for transform-domain collaborative filtering
CN109558832B (en) Human body posture detection method, device, equipment and storage medium
CN112288011B (en) Image matching method based on self-attention deep neural network
WO2022000420A1 (en) Human body action recognition method, human body action recognition system, and device
Zhang et al. Adaptive residual networks for high-quality image restoration
Gai et al. New image denoising algorithm via improved deep convolutional neural network with perceptive loss
Kussul et al. Improved method of handwritten digit recognition tested on MNIST database
WO2022105125A1 (en) Image segmentation method and apparatus, computer device, and storage medium
WO2022217746A1 (en) High-resolution hyperspectral calculation imaging method and system, and medium
CN106548159A (en) Reticulate pattern facial image recognition method and device based on full convolutional neural networks
CN105205453B (en) Human eye detection and localization method based on depth self-encoding encoder
CN111476249B (en) Construction method of multi-scale large-receptive-field convolutional neural network
CN109978077B (en) Visual recognition method, device and system and storage medium
CN111104941B (en) Image direction correction method and device and electronic equipment
CN116740362B (en) Attention-based lightweight asymmetric scene semantic segmentation method and system
Chaurasiya et al. Deep dilated CNN based image denoising
Xu et al. AutoSegNet: An automated neural network for image segmentation
Liu et al. Iris recognition in visible spectrum based on multi-layer analogous convolution and collaborative representation
Uddin et al. A perceptually inspired new blind image denoising method using $ L_ {1} $ and perceptual loss
CN115995027A (en) Self-supervision scene change detection method, system and computer readable medium
CN110163095B (en) Loop detection method, loop detection device and terminal equipment
CN115439669A (en) Feature point detection network based on deep learning and cross-resolution image matching method
Rana et al. MSRD-CNN: Multi-scale residual deep CNN for general-purpose image manipulation detection

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant