CN112446266B - Face recognition network structure suitable for front end - Google Patents

Face recognition network structure suitable for front end Download PDF

Info

Publication number
CN112446266B
CN112446266B CN201910830546.2A CN201910830546A CN112446266B CN 112446266 B CN112446266 B CN 112446266B CN 201910830546 A CN201910830546 A CN 201910830546A CN 112446266 B CN112446266 B CN 112446266B
Authority
CN
China
Prior art keywords
feature map
convolution
processing
module
depth
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910830546.2A
Other languages
Chinese (zh)
Other versions
CN112446266A (en
Inventor
田凤彬
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Ingenic Semiconductor Co Ltd
Original Assignee
Beijing Ingenic Semiconductor Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Ingenic Semiconductor Co Ltd filed Critical Beijing Ingenic Semiconductor Co Ltd
Priority to CN201910830546.2A priority Critical patent/CN112446266B/en
Publication of CN112446266A publication Critical patent/CN112446266A/en
Application granted granted Critical
Publication of CN112446266B publication Critical patent/CN112446266B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Health & Medical Sciences (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Health & Medical Sciences (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Human Computer Interaction (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Mathematical Physics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Molecular Biology (AREA)
  • Evolutionary Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention provides a face recognition network structure suitable for a front end, wherein the network uses the idea of downsampling to reach a full connection layer, and a feature map of each small module result is directly connected with the full connection layer after downsampling, so that the extracted features of each small module are effectively utilized. In each convolution, the result of the convolution above is used to effectively avoid the reinforcement of the local feature and the disappearance of the feature. The results of the above modules are spliced in the input of each module, thereby avoiding the disappearance of gradients.

Description

Face recognition network structure suitable for front end
Technical Field
The invention relates to the technical field of face image recognition, in particular to a face recognition network structure suitable for a front end.
Background
Along with the continuous development of technology, particularly the development of computer vision technology, the face recognition technology is widely applied to various fields such as information security, electronic authentication and the like, and the image feature extraction method has good recognition performance. Face recognition refers to a technique of recognizing one or more faces from a static or dynamic scene using image processing and/or pattern recognition techniques based on a known face sample library. However, the current face recognition technology comprises 1, various networks applied to face recognition, inception, resNet, VGG, mobileNet, alexNet, googLeNet, inception-ResNet; 2. the framework structure of face recognition training comprises a facenet and an instmctation. The existing face recognition technology has the following defects: 1. the current front-end network uses independent convolution, feature map addition, convolution kernels with various sizes and the like, and is not suitable for many AI chips, so that the AI chips cannot realize the network. 2. The calculation amount is relatively large. 3. The recognition rate is relatively low.
Disclosure of Invention
In order to solve the above problems in the prior art, an object of the present invention is to:
1. few operators are used, only 3 x 3 convolution kernels, employing a feature map stitching (concat) operation, a downsampling operation, and a full join operation.
2. The calculated amount is small, the calculated amount is less than multiplication of 1G, and the model (the parameter is floating point) size is only 29M.
3. The face recognition rate is improved to 99.8 percent.
The invention provides a face recognition network structure suitable for a front end, wherein the network uses a form of downsampling to reach a full connection layer, a network module is arranged in the network, a feature map of a result of each module is directly connected with the full connection layer after downsampling, and in each convolution process, the result of the convolution process above is spliced by using a cross splice, and the result of each module above is spliced in the input of each module.
The convolution processing is that the convolution used by each module is the processing of a convolution kernel of 3 multiplied by 3, and the normalization processing is carried out after the convolution, and then the processing of the excitation function is carried out.
In the design of each layer of convolution, the number of output characteristic diagrams is independently designed.
The downsampling process is that the processing core size is 2×2, and the step size is 2.
Further comprising firstly performing equalization processing on input image data; and carrying out convolution processing with a convolution kernel of 3 multiplied by 3 and a step length of 1 on the data to generate a feature map.
The network module further comprises: subjecting to a first feature pattern length and width reduction process; the first feature map length and width reduction processing includes:
[1.1] a feature map generated by convolution processing of image data is subjected to convolution processing with a convolution kernel of 3×3 and a step length of 2, the depth of the generated feature map is 32, and the generated feature map is 48×48×32;
[1.2] the feature map obtained by convolving the image data described in [1.1] is downsampled, convolved with a processing kernel size of 2×2 and a step size of 2, the depth of the generated feature map is 32, and the generated feature map is 48×48×64.
The network module further comprises: subjecting to a second feature pattern length and width reduction process; the second feature pattern length and width reduction processing further comprises:
input data are respectively in_data1, in_data2, in_data3, N;
[2.1] performing convolution processing with the convolution kernel of in_data1 being 3×3 and the step size being 2, generating a feature map with a depth of N, and generating a feature map with a depth of W×H×N;
[2.2] downsampling the in_data2 with a processing kernel size of 2×2 and a step size of 2 to generate a feature map, which is set to be M as the depth of the input feature map, and is set to be W×H×M;
[2.3] the in_data3 is subjected to downsampling processing, the processing core size is 2×2, the step size is 2, the generated feature map is M as the depth of the input feature map, and the generated feature map is W×H×M.
The network module includes: the first module, the second module, the third module and the fourth module.
The last module of the network is ultimately fully connected to 192 data.
The application has the advantages that:
1. the idea of down sampling to the full-connection layer is used in the network, and the feature map of each small module result is directly connected with the full-connection layer after down sampling, so that the extracted features of each small module are effectively utilized. In each convolution, the result of the convolution above is used to effectively avoid the reinforcement of the local feature and the disappearance of the feature. The results of the above modules are spliced in the input of each module, thereby avoiding the disappearance of gradients.
2. By using a scheme of independently designing the number of the output characteristic graphs, the calculated amount can be effectively reduced, and the efficiency of convolution extracting the characteristics can be improved to the greatest extent.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this application, illustrate and together with the description serve to explain the invention.
Fig. 1 is a schematic block diagram of the network architecture of the present invention.
Fig. 2-1 through 2-22 are flow diagrams of embodiments of the present invention.
Detailed Description
In the field of face recognition, some related technical terms currently include:
1. deep learning: the concept of deep learning is derived from the study of artificial neural networks. The multi-layer sensor with multiple hidden layers is a deep learning structure. Deep learning forms more abstract high-level representation attribute categories or features by combining low-level features to discover distributed feature representations of data.
2. Convolution kernel: the convolution kernel is a matrix used in image processing and is a parameter for operation with the original image. The convolution kernel is typically a matrix of columns (e.g., a matrix of 3*3) with a weight value for each square in the region. The matrix shapes are typically 1×1,3×3,5×5,7×7,1×3,3×1,2×2,1×5,5×1, ….
3. Convolution: the center of the convolution kernel is placed over the pixel to be calculated, and the products of each element in the kernel and its covered image pixel values are calculated and summed once to obtain a structure that is the new pixel value for that location, a process called convolution.
4. Feature map: the result obtained by convolution calculation of input data is called a feature map, and the result generated by full connection of the data is also called a feature map. The feature map size is generally expressed as length x width x depth, or 1 x depth.
5. Step size: the center position of the convolution kernel is moved by the length of the movement in the coordinates.
6: and (3) performing two-end misalignment treatment: processing an image or data with a convolution kernel size of 3 and a step size of 2 may result in insufficient data on both sides, where discarding data on both sides or on one side is used, a phenomenon called both sides not processing it.
7. Front-end face detection: the face detection function used on the AI chip or the common chip is called front-end face detection, and the speed and the accuracy of the front-end face detection are lower than those of the cloud server.
8. Excitation function: each node in the neural network accepts the input value and passes the input value to the next layer, and the input node passes the input attribute value directly to the next layer (hidden layer or output layer). In a neural network, there is a functional relationship between the inputs and outputs of hidden and output layer nodes, this function being called the excitation function. Common excitation functions are: linear excitation functions, threshold or step excitation functions, sigmoid excitation functions, hyperbolic tangent excitation functions, gaussian excitation functions, and the like.
In this application, as shown in fig. 1, the network structure that is set specifically includes:
1. network structure
1) Network characteristics. The idea of down sampling to the full-connection layer is used in the network, and the feature map of each small module result is directly connected with the full-connection layer after down sampling, so that the extracted features of each small module are effectively utilized. In each convolution, the result of the convolution above is used to effectively avoid the reinforcement of the local feature and the disappearance of the feature. The results of the above modules are spliced in the input of each module, thereby avoiding the disappearance of gradients.
2) In the specific design of each layer of convolution, the calculation amount can be effectively reduced by using a scheme of independently designing the number of the output feature graphs, and the efficiency of extracting the features by convolution can be improved to the greatest extent. Feature mapping of each small module is performed once.
As shown in fig. 1, the input image data is subjected to equalization processing, subtraction is performed 128 to generate data1, and division is performed on the data1 by 128 to generate data2; then, the convolution processing with the convolution kernel of 3×3 and the step size of 1 is performed on data1, and the depth of the generated feature map is 64. The generated feature map is 96×96×64 and named conv0; subjecting to a first reduced feature pattern length and width process, said first reduced feature pattern length and width process being as follows:
[1.1] Conv0 was subjected to convolution processing with a convolution kernel of 3×3 and a step size of 2, and the depth of the generated feature map was 32. The generated feature map was 48×48×32, and named conv1.
[1.2] Conv0 was subjected to downsampling, and convolution processing with a processing kernel size of 2×2 and a step size of 2 was performed to generate a feature map with a depth of 32. The generated feature map was 48×48×64 and named pool1.
And then performing layer-by-layer convolution processing, excitation function processing and downsampling processing to generate a first module, a second module, a third module and a fourth module, wherein the first module, the second module, the third module and the fourth module undergo second characteristic diagram length and width reduction processing. The second feature pattern length and width reduction processing further comprises:
input data are respectively in_data1, in_data2, in_data3, N;
[2.1] performing convolution processing with the convolution kernel of in_data1 being 3×3 and the step size being 2, generating a feature map with a depth of N, and generating a feature map with a depth of W×H×N;
[2.2] downsampling the in_data2 with a processing kernel size of 2×2 and a step size of 2 to generate a feature map, which is set to be M as the depth of the input feature map, and is set to be W×H×M;
[2.3] the in_data3 is subjected to downsampling processing, the processing core size is 2×2, the step size is 2, the generated feature map is M as the depth of the input feature map, and the generated feature map is W×H×M.
The last module of the network, the fourth module, is ultimately all connected to 192 data.
Further, the specific setting method referred to in the present application includes the following as shown in fig. 2-1 to 2-22:
2. network module
1) The convolution used by each module is the convolution processing of a convolution kernel of 3×3, and the normalization processing is performed after the convolution, and then the excitation function processing is performed. This process is hereinafter simply referred to as convolution process.
2) Carrying out equalization treatment on input image data, firstly carrying out 128 subtraction to generate data1, and then dividing the data1 by 128 to generate data2;
3) The convolution processing with the convolution kernel of 3×3 and the step size of 1 is performed on data1, and the depth of the generated feature map is 64. The generated feature map is 96×96×64 and named conv0;
4) The first type of feature pattern length and width reduction processing.
[1] Conv0 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 2, and a depth of the generated feature map is 32. The generated feature map was 48×48×32, and named conv1.
[2] The conv0 is subjected to downsampling processing, convolution processing with a processing kernel size of 2×2 and a step size of 2 is performed, and a depth of the generated feature map is 32. The generated feature map was 48×48×64 and named pool1.
5) A first module.
[1] Pool1 was subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and the depth of the generated feature map was 32. The generated feature map is 48×48×32, and is named as block1_conv1.
[2] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block1_conv1, and the depth of the generated feature map is 64. The generated feature map is 48×48×32, and is named as block1_conv2.
[3] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block1_conv2, and the depth of the generated feature map is 64. The generated feature map is 48×48×32, and is named as block1_conv3.
[4] And splicing the block1_conv3 and the conv1 to generate a result named as block1_concat1.
6) And the second type of feature pattern length and width reduction treatment.
The input data are in_data1, in_data2, in_data3, N, respectively.
[1] And performing convolution processing with the convolution kernel of 3 multiplied by 3 and the step length of 2 on the in_data1 to generate a feature map with the depth of N. The generated feature map is w×h×n, and is named out_conv1.
[2] The in_data2 is subjected to downsampling, the processing core size is 2×2, and the step size is 2, and the generated feature map is the same as the depth of the input feature map, here denoted by M. The generated feature map is w×h×m, and is named out_pool1.
[3] The in_data3 is subjected to downsampling, the processing core size is 2×2, and the step size is 2, and the generated feature map is the same as the depth of the input feature map, here denoted by M. The generated feature map is w×h×m, and is named out_pool2.
7) And a second module.
[1] Through 6) a second reduced profile length and width process, where in_data1=block1_concat1, in_data2=block1_concat1, n=48. The output result is named as block2_conv0=out_conv1, block2_pool 1=out_pool 1.
[2] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block2_conv0, and the depth of the generated feature map is 64. The generated feature map is 24×24×64, and is named as block2_conv1.
[3] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block2_conv1, and the depth of the generated feature map is 48. The generated feature map is 24×24×48 and named block2_conv2.
[4] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block2_conv2, and the depth of the generated feature map is 48. The generated feature map is 24×24×48 and named block2_conv3.
[5] And splicing the block2_conv3 and the block2_conv0 to generate a result named block2_concat1.
[6] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block2_conv2, and the depth of the generated feature map is 64. The generated feature map is 24×24×64 and is named block2_conv4.
8) And a third module.
A first part:
[1] through the above 6) second feature map length-width reduction processing, in_data1=block2_concat1, in_data2=block2_conv4, in_data3=block2_concat1, n=48. The output result is named as block31_conv0=out_conv1, block31_poo1=out_poo1, and block31_poo2=out_poo2.
[2] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block31—conv0, and the depth of the generated feature map is 128. The generated feature map is 12×12×128, and is named block31_conv1.
[3] The block31_conv1 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of the feature map is 80. The generated feature map is 12×12×80, and is named block31_conv2.
[4] The result of splicing the block31_conv2 and the block31_pool1 is named as block31_concat1.
[5] The block31_concat1 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of the generated feature map is 64. The generated feature map is 12×12×64 and is named block31_conv3.
[6] The result of splicing the block31_conv3 and the block31_pool2 is named as block31_concat2.
[7] The convolution processing with the convolution kernel of 3×3 and the step size of 1 is performed on the block31_conca1, the depth of the generated feature map is 96, and the generated feature map is 12×12×96 and is named as block32_conv1.
A second part:
[1] the block31_concat2 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of the generated feature map is 48. The generated feature map is 12×12×48 and named block32_conv4.
[2] The block32_conv1 and the block31_conv4 are spliced, and the generated result is named as block 32_conc1.
[3] The block32_concat1 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of a feature map is generated to be 96. The generated feature map is 12×12×96 and named block32_conv2.
[4] And splicing the block32_conv2 and the block32_conv4 to generate a result named block32_conc2.
[5] The block32_concat2 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of a feature map is generated to be 96. The generated feature map is 12×12×96 and named block32_conv3.
[6] And splicing the block32_conv3 with the block31_conca2 to generate a result named as block32_conca3.
Third section:
[1] the block32_concat3 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of the feature map is generated to be 48. The generated feature map is 12×12×48 and named block33_conv1.
[2] The result of splicing the block33_conv1 and the block32_conv3 is named as block33_conc1.
[3] The block33_concat1 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of a feature map is generated to be 96. The generated feature map is 12×12×96 and named block33_conv2.
[4] And splicing the block33_conv2 and the block33_conv1 to generate a result named block33_conc2.
[5] The block33_concat2 is subjected to convolution processing with a convolution kernel of 3×3 and a step size of 1, and a depth of a feature map is generated to be 96. The generated feature map is 12×12×96 and named block33_conv3.
[6] And splicing the block33_conv3 with the block32_conc3 to generate a result named block33_conc3.
9) And a fourth module.
[1] Through 6) a second reduced profile length-width process, where in_data1=block33_concat3, in_data2=block33_conv3, in_data3=block33_concat3, n=64. The output result is named as block4_conv0=out_conv1, block4_pog1=out_pog1, block4_pog2=out_pog2.
[2] The convolution processing with the convolution kernel 3×3 and the step size 1 is performed on the block4_conv0, and the depth of the generated feature map is 160. The generated feature map is 6×6×160 and is named block4_conv1.
[3] And splicing the block4_conv1 and the block4_pool1, and naming the generated result as block4_conca1.
[4] The convolution processing with 3×3 convolution kernel and 1 step size is performed on the block4_concat1, and the depth of the generated feature map is 128. The generated feature map is 6×6×128 and is named block4_conv2.
[5] And splicing the block4_conv2 and the block4_conv0 to generate a result named block4_conc2.
[6] The convolution processing with 3×3 convolution kernel and 1 step size is performed on the block4_concat2, and the depth of the generated feature map is 192. The generated feature map is 6×6×192 and is named block4_conv3.
[7] And splicing the block4_conv3 with the block31_pool2 to generate a result named block4_concat3.
10 Full connection)
The block4_conv3 is fully connected to 192 data.
The above description is only of the preferred embodiments of the present invention and is not intended to limit the present invention, and various modifications and variations can be made to the embodiments of the present invention by those skilled in the art. Any modification, equivalent replacement, improvement, etc. made within the spirit and principle of the present invention should be included in the protection scope of the present invention.

Claims (1)

1. The face recognition network structure suitable for the front end is characterized in that the network uses a form of downsampling to reach a full connection layer, a network module is arranged in the network, a feature map of a result of each module is directly connected with the full connection layer after downsampling, and in each convolution process, the result of the convolution process above is spliced by using the cross splice, and the result of each module above is spliced in the input of each module;
the convolution processing is the processing that the convolution used by each module is a convolution kernel of 3 multiplied by 3, and the normalization processing is carried out after the convolution, and then the excitation function processing is carried out;
in the design of each layer of convolution, the number of the output characteristic diagrams is independently designed;
the downsampling process is that the processing kernel size is 2 multiplied by 2, and the step length is 2;
firstly, carrying out equalization treatment on input image data; then, carrying out convolution processing with a convolution kernel of 3 multiplied by 3 and a step length of 1 on the data to generate a feature map; subjecting to a first feature pattern length and width reduction process; the first feature map length and width reduction processing includes:
[1.1] a feature map generated by convolution processing of image data is subjected to convolution processing with a convolution kernel of 3×3 and a step length of 2, the depth of the generated feature map is 32, and the generated feature map is 48×48×32;
[1.2] downsampling the feature map obtained by convolving the image data described in [1.1], convolving the feature map with a processing kernel size of 2×2 and a step size of 2, generating a feature map with a depth of 32, and generating a feature map with a depth of 48×48×64;
the network module comprises a first module, a second module, a third module and a fourth module; the modules are subjected to a second feature pattern length and width reduction process, wherein the second feature pattern length and width reduction process comprises the following steps:
the input data are in_data1, in_data2, in_data3 and N respectively;
[2.1] performing convolution processing with the convolution kernel of in_data1 being 3×3 and the step size being 2, generating a feature map with a depth of N, and generating a feature map with a depth of W×H×N;
[2.2] downsampling the in_data2 with a processing kernel size of 2×2 and a step size of 2 to generate a feature map, which is set to be M as the depth of the input feature map, and is set to be W×H×M;
[2.3] downsampling the in_data3 with a processing kernel size of 2×2 and a step size of 2 to generate a feature map, which is set to M as the depth of the input feature map, and is set to W×H×M;
the fourth module of the network is ultimately fully connected to 192 data.
CN201910830546.2A 2019-09-04 2019-09-04 Face recognition network structure suitable for front end Active CN112446266B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910830546.2A CN112446266B (en) 2019-09-04 2019-09-04 Face recognition network structure suitable for front end

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910830546.2A CN112446266B (en) 2019-09-04 2019-09-04 Face recognition network structure suitable for front end

Publications (2)

Publication Number Publication Date
CN112446266A CN112446266A (en) 2021-03-05
CN112446266B true CN112446266B (en) 2024-03-29

Family

ID=74734921

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910830546.2A Active CN112446266B (en) 2019-09-04 2019-09-04 Face recognition network structure suitable for front end

Country Status (1)

Country Link
CN (1) CN112446266B (en)

Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105975931A (en) * 2016-05-04 2016-09-28 浙江大学 Convolutional neural network face recognition method based on multi-scale pooling
CN106570521A (en) * 2016-10-24 2017-04-19 中国科学院自动化研究所 Multi-language scene character recognition method and recognition system
CN106782602A (en) * 2016-12-01 2017-05-31 南京邮电大学 Speech-emotion recognition method based on length time memory network and convolutional neural networks
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN107154037A (en) * 2017-04-20 2017-09-12 西安交通大学 Fan blade fault recognition method based on depth level feature extraction
CN107622261A (en) * 2017-11-03 2018-01-23 北方工业大学 Face age estimation method and device based on deep learning
CN107679487A (en) * 2017-09-29 2018-02-09 中国科学院福建物质结构研究所 Missing Persons' discrimination method and system
WO2018028255A1 (en) * 2016-08-11 2018-02-15 深圳市未来媒体技术研究院 Image saliency detection method based on adversarial network
KR101845769B1 (en) * 2017-03-21 2018-04-06 경남대학교 산학협력단 Car rear detection system using convolution neural network, and method thereof
CN107977456A (en) * 2017-12-15 2018-05-01 清华大学 A kind of multi-source big data analysis method based on multitask depth network
CN108319962A (en) * 2018-01-29 2018-07-24 安徽大学 A kind of Tool Wear Monitoring method based on convolutional neural networks
CN108764317A (en) * 2018-05-21 2018-11-06 浙江工业大学 A kind of residual error convolutional neural networks image classification method based on multichannel characteristic weighing
WO2018214195A1 (en) * 2017-05-25 2018-11-29 中国矿业大学 Remote sensing imaging bridge detection method based on convolutional neural network
CN109376625A (en) * 2018-10-10 2019-02-22 东北大学 A kind of human facial expression recognition method based on convolutional neural networks
CN109522807A (en) * 2018-10-22 2019-03-26 深圳先进技术研究院 Satellite image identifying system, method and electronic equipment based on self-generating feature
US10262214B1 (en) * 2018-09-05 2019-04-16 StradVision, Inc. Learning method, learning device for detecting lane by using CNN and testing method, testing device using the same
US10304193B1 (en) * 2018-08-17 2019-05-28 12 Sigma Technologies Image segmentation and object detection using fully convolutional neural network
CN109886341A (en) * 2019-02-25 2019-06-14 厦门美图之家科技有限公司 A kind of trained method for generating Face datection model
CN110097003A (en) * 2019-04-29 2019-08-06 中南民族大学 Check class attendance method, equipment, storage medium and device neural network based
CN110096964A (en) * 2019-04-08 2019-08-06 厦门美图之家科技有限公司 A method of generating image recognition model

Patent Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017133009A1 (en) * 2016-02-04 2017-08-10 广州新节奏智能科技有限公司 Method for positioning human joint using depth image of convolutional neural network
CN105975931A (en) * 2016-05-04 2016-09-28 浙江大学 Convolutional neural network face recognition method based on multi-scale pooling
WO2018028255A1 (en) * 2016-08-11 2018-02-15 深圳市未来媒体技术研究院 Image saliency detection method based on adversarial network
CN106570521A (en) * 2016-10-24 2017-04-19 中国科学院自动化研究所 Multi-language scene character recognition method and recognition system
CN106782602A (en) * 2016-12-01 2017-05-31 南京邮电大学 Speech-emotion recognition method based on length time memory network and convolutional neural networks
KR101845769B1 (en) * 2017-03-21 2018-04-06 경남대학교 산학협력단 Car rear detection system using convolution neural network, and method thereof
CN107154037A (en) * 2017-04-20 2017-09-12 西安交通大学 Fan blade fault recognition method based on depth level feature extraction
WO2018214195A1 (en) * 2017-05-25 2018-11-29 中国矿业大学 Remote sensing imaging bridge detection method based on convolutional neural network
CN107679487A (en) * 2017-09-29 2018-02-09 中国科学院福建物质结构研究所 Missing Persons' discrimination method and system
CN107622261A (en) * 2017-11-03 2018-01-23 北方工业大学 Face age estimation method and device based on deep learning
CN107977456A (en) * 2017-12-15 2018-05-01 清华大学 A kind of multi-source big data analysis method based on multitask depth network
CN108319962A (en) * 2018-01-29 2018-07-24 安徽大学 A kind of Tool Wear Monitoring method based on convolutional neural networks
CN108764317A (en) * 2018-05-21 2018-11-06 浙江工业大学 A kind of residual error convolutional neural networks image classification method based on multichannel characteristic weighing
US10304193B1 (en) * 2018-08-17 2019-05-28 12 Sigma Technologies Image segmentation and object detection using fully convolutional neural network
US10262214B1 (en) * 2018-09-05 2019-04-16 StradVision, Inc. Learning method, learning device for detecting lane by using CNN and testing method, testing device using the same
CN109376625A (en) * 2018-10-10 2019-02-22 东北大学 A kind of human facial expression recognition method based on convolutional neural networks
CN109522807A (en) * 2018-10-22 2019-03-26 深圳先进技术研究院 Satellite image identifying system, method and electronic equipment based on self-generating feature
CN109886341A (en) * 2019-02-25 2019-06-14 厦门美图之家科技有限公司 A kind of trained method for generating Face datection model
CN110096964A (en) * 2019-04-08 2019-08-06 厦门美图之家科技有限公司 A method of generating image recognition model
CN110097003A (en) * 2019-04-29 2019-08-06 中南民族大学 Check class attendance method, equipment, storage medium and device neural network based

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
基于特征图切分的轻量级卷积神经网络;张雨丰;郑忠龙;刘华文;向道红;何小卫;李知菲;何依然;KHODJA Abd Erraouf;;模式识别与人工智能(03);第237-246页 *

Also Published As

Publication number Publication date
CN112446266A (en) 2021-03-05

Similar Documents

Publication Publication Date Title
CN111462126B (en) Semantic image segmentation method and system based on edge enhancement
CN108647585B (en) Traffic identifier detection method based on multi-scale circulation attention network
CN109190752B (en) Image semantic segmentation method based on global features and local features of deep learning
CN107564025B (en) Electric power equipment infrared image semantic segmentation method based on deep neural network
US20190311223A1 (en) Image processing methods and apparatus, and electronic devices
CN112528976B (en) Text detection model generation method and text detection method
CN109118504B (en) Image edge detection method, device and equipment based on neural network
CN107564009B (en) Outdoor scene multi-target segmentation method based on deep convolutional neural network
CN110346808B (en) Point cloud data processing method and system of laser radar
CN111488901B (en) Method and device for extracting features from input images in multiple modules in CNN
CN113468996B (en) Camouflage object detection method based on edge refinement
CN111476247B (en) CNN method and device using 1xK or Kx1 convolution operation
CN112329801B (en) Convolutional neural network non-local information construction method
CN111553869A (en) Method for complementing generated confrontation network image under space-based view angle
CN111815526A (en) Rain image rainstrip removing method and system based on image filtering and CNN
CN113392728B (en) Target detection method based on SSA sharpening attention mechanism
Jung et al. Extension of convolutional neural network with general image processing kernels
CN113256546A (en) Depth map completion method based on color map guidance
CN112446266B (en) Face recognition network structure suitable for front end
CN112446267B (en) Setting method of face recognition network suitable for front end
CN116452900A (en) Target detection method based on lightweight neural network
CN114519819B (en) Remote sensing image target detection method based on global context awareness
CN113688783B (en) Face feature extraction method, low-resolution face recognition method and equipment
Kaur et al. Deep transfer learning based multiway feature pyramid network for object detection in images
Wang et al. Saliency detection by multilevel deep pyramid model

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant