CN112597823A - Attention recognition method and device, electronic equipment and storage medium - Google Patents
Attention recognition method and device, electronic equipment and storage medium Download PDFInfo
- Publication number
- CN112597823A CN112597823A CN202011438525.5A CN202011438525A CN112597823A CN 112597823 A CN112597823 A CN 112597823A CN 202011438525 A CN202011438525 A CN 202011438525A CN 112597823 A CN112597823 A CN 112597823A
- Authority
- CN
- China
- Prior art keywords
- model
- face
- attention
- classification
- determining
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 81
- 230000004927 fusion Effects 0.000 claims abstract description 86
- 238000013145 classification model Methods 0.000 claims abstract description 81
- 238000012549 training Methods 0.000 claims description 94
- 238000012795 verification Methods 0.000 claims description 37
- 238000001514 detection method Methods 0.000 claims description 25
- 210000001525 retina Anatomy 0.000 claims description 16
- 238000000605 extraction Methods 0.000 claims description 11
- 238000004590 computer program Methods 0.000 claims description 9
- 238000012216 screening Methods 0.000 claims description 2
- 238000010586 diagram Methods 0.000 description 16
- 238000002790 cross-validation Methods 0.000 description 13
- 230000006870 function Effects 0.000 description 10
- 238000007500 overflow downdraw method Methods 0.000 description 8
- 238000004891 communication Methods 0.000 description 6
- 238000010200 validation analysis Methods 0.000 description 5
- 230000000694 effects Effects 0.000 description 4
- 238000002474 experimental method Methods 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000013527 convolutional neural network Methods 0.000 description 3
- 238000013528 artificial neural network Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 210000000887 face Anatomy 0.000 description 2
- 230000003287 optical effect Effects 0.000 description 2
- 230000036544 posture Effects 0.000 description 2
- 238000003066 decision tree Methods 0.000 description 1
- 230000007423 decrease Effects 0.000 description 1
- 210000001061 forehead Anatomy 0.000 description 1
- 238000009434 installation Methods 0.000 description 1
- 238000002372 labelling Methods 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003340 mental effect Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000007084 physiological dysfunction Effects 0.000 description 1
- 230000035790 physiological processes and functions Effects 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 230000005997 psychological dysfunction Effects 0.000 description 1
- 230000008433 psychological processes and functions Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
- 238000012706 support-vector machine Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/25—Fusion techniques
- G06F18/254—Fusion techniques of classification results, e.g. of results related to same input data
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/50—Context or environment of the image
- G06V20/59—Context or environment of the image inside of a vehicle, e.g. relating to seat occupancy, driver state or inner lighting conditions
- G06V20/597—Recognising the driver's state or behaviour, e.g. attention or drowsiness
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/168—Feature extraction; Face representation
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- General Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Cheminformatics (AREA)
- General Engineering & Computer Science (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
Abstract
The application relates to the technical field of computer vision, and provides an attention recognition method, an attention recognition device, electronic equipment and a storage medium, wherein the attention recognition method comprises the following steps: determining a face image of a driver to be recognized; inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model; the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types. The method, the device, the electronic equipment and the storage medium can fully identify the face characteristics of the driver, have strong generalization capability and improve the accuracy of the attention identification of the driver.
Description
Technical Field
The present application relates to the field of computer vision technologies, and in particular, to an attention recognition method and apparatus, an electronic device, and a storage medium.
Background
Fatigue driving refers to the phenomenon of decline of driving performance due to the psychological and physiological functions of the driver after continuous driving. Driving fatigue is generally considered to be technical fatigue involving both mental and physical strength. When a driver drives a vehicle, physiological or psychological dysfunction often occurs due to various reasons, attention is reduced, and driving safety is seriously affected.
In the prior art, the fatigue state of a driver is generally analyzed by collecting electroencephalograms of the forehead of the driver, and then the attention state of the driver is obtained. The accuracy of recognizing the attention state of the driver by the method is low.
Disclosure of Invention
The application provides an attention recognition method, an attention recognition device, electronic equipment and a storage medium, which can fully recognize the human face characteristics of a driver, have strong generalization capability and improve the accuracy of the attention recognition of the driver.
The application provides an attention recognition method, comprising the following steps:
determining a face image of a driver to be recognized;
inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model;
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
According to the attention recognition method provided by the application, the determination method of the attention classification fusion model comprises the following steps:
respectively selecting an attention classification model with the highest performance score from a plurality of attention classification models of each type as a basic model of each type;
and fusing the basic models of each type to obtain the attention classification fusion model.
According to an attention recognition method provided by the present application, the step of respectively selecting an attention classification model with the highest performance score from a plurality of attention classification models of each type as a base model of each type includes:
determining a plurality of face training sets and corresponding face verification sets thereof;
training any type of initial model based on each face training set to obtain an attention classification model corresponding to each face training set;
performing performance scoring on the attention classification model corresponding to each face training set based on the face verification set corresponding to each face training set;
and taking the attention classification model with the highest performance score as a basic model corresponding to any type.
According to the attention recognition method provided by the application, the determining of the plurality of face training sets and the face verification sets corresponding to the face training sets comprises the following steps:
determining a sample face data set;
and splitting the sample face data set into a plurality of subsets, and determining a plurality of face training sets and corresponding face verification sets by adopting a cross verification method.
According to the attention recognition method provided by the application, the fusing of each type of basic model to obtain the attention classification fusion model comprises the following steps:
determining a weight corresponding to each basic model based on the performance score of each type of basic model;
and fusing the plurality of basic models based on the weight corresponding to each basic model to obtain the attention classification fusion model.
According to the attention recognition method provided by the application, the step of determining the face image of the driver to be recognized comprises the following steps:
acquiring image data containing the face of the driver;
inputting the image data into a face detection model to obtain a face image output by the face detection model;
the Face detection model is established based on a Retina Face model, and a feature extraction layer of the Retina Face model is increment V4.
According to an attention recognition method provided by the application, the plurality of different types of attention classification models comprise at least one of DenseNet121, EfficientNet-B0, EfficientNet-B3 and EfficientNet-B7.
The present application also provides an attention recognition device, comprising:
a determination unit for determining a face image of a driver to be recognized;
the recognition unit is used for inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model;
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
The present application further provides an electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, wherein the processor implements the steps of the attention recognition method as described in any of the above when executing the program.
The present application also provides a non-transitory computer readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of the method of attention recognition as described in any of the above.
According to the attention recognition method, the attention recognition device, the electronic equipment and the storage medium, the input face image of the driver to be recognized is recognized through the attention classification fusion model, and the face direction angle of the driver is obtained. The attention classification fusion model is obtained by fusing a plurality of attention classification models of different types, so that the face characteristics of the driver can be fully recognized, the generalization capability is strong, and the accuracy of the driver attention recognition is improved.
Drawings
In order to more clearly illustrate the technical solutions in the present application or the prior art, the drawings needed for the description of the embodiments or the prior art will be briefly introduced below, and it is obvious that the drawings in the following description are some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic flow chart of an attention recognition method provided in the present application;
FIG. 2 is a schematic flow chart of a method for determining an attention classification fusion model provided in the present application;
FIG. 3 is a schematic flow chart of a method for determining a base model provided herein;
FIG. 4 is a schematic flow chart of a cross-validation method provided herein;
FIG. 5 is a schematic flow chart of a model fusion method provided herein;
fig. 6 is a schematic flow chart of a face image determination method provided in the present application;
FIG. 7 is a block diagram of a basic module of a feature extraction layer provided herein;
FIG. 8 is a schematic flow chart of a training method of the attention classification fusion model provided in the present application;
FIG. 9 is a schematic structural diagram of an attention recognition device provided herein;
fig. 10 is a schematic structural diagram of a fusion model determination unit provided in the present application;
FIG. 11 is a schematic structural diagram of a basic model determination subunit provided in the present application;
FIG. 12 is a block diagram of a training validation determination module provided herein;
FIG. 13 is a schematic diagram of the structure of a base model fusion subunit provided herein;
fig. 14 is a schematic structural diagram of a determination unit provided in the present application;
fig. 15 is a schematic structural diagram of an electronic device provided in the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present application clearer, the technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are some embodiments of the present application, but not all embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
Fig. 1 is a schematic flow chart of an attention recognition method provided in the present application, and as shown in fig. 1, the method includes:
Specifically, the image of the face of the driver can be acquired in real time by a camera mounted inside the vehicle. A plurality of cameras can be installed at different positions in the vehicle, and each camera acquires images of the face of a driver in different postures. For example, the installation position of the camera can be the positions of a rearview mirror, a center console, an instrument panel, a windshield and the like. The number of cameras may be set to 9.
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
Specifically, the human face image of the driver is input into the attention classification fusion model, and the attention classification fusion model is used for identifying the human face posture in the human face image to obtain the human face direction angle. The face direction angle is used to characterize the direction of attention of the driver. The attention direction of the driver can be judged according to the face direction angle, and the attention focusing time of the driver can be judged according to the time of the face direction angle.
A plurality of attention classification models of different types can be obtained through pre-training, and the models are fused to obtain an attention classification fusion model.
The attention classification model of a certain type can be obtained through pre-training, and specifically, the attention classification model can be obtained through the following training modes: firstly, collecting a large number of sample face images of a driver, determining a face direction angle label of each sample face image in a manual labeling mode, and then inputting the large number of sample face images and the face direction angle labels corresponding to the sample face images into an initial model for training, so as to obtain the type of attention classification model by aiming at improving the recognition capability of the initial model on the face direction angles in the sample face images.
The initial model of the attention classification model may be a decision tree model, an artificial neural network model, a support vector machine model, or the like, which is not specifically limited in this embodiment of the present application.
The attention classification fusion model obtained after fusion can also be subjected to joint training, for example, the attention classification fusion model can be trained by adopting a global loss function with the aim of reducing the overall loss of the attention classification fusion model.
According to the attention recognition method, the input face image of the driver to be recognized is recognized through the attention classification fusion model, and the face direction angle of the driver is obtained. The attention classification fusion model is obtained by fusing a plurality of attention classification models of different types, so that the face characteristics of the driver can be fully recognized, the generalization capability is strong, and the accuracy of the driver attention recognition is improved.
Based on the above embodiment, fig. 2 is a schematic flow chart of a method for determining an attention classification fusion model provided in the present application, and as shown in fig. 2, the method for determining an attention classification fusion model includes:
Specifically, the base model is a model with optimal performance in each type of the multiple attention classification models, and can be used as a basic component unit of the attention classification fusion model.
For each type of initial model, multiple attention classification models may be trained using multiple data sets. For example, a convolutional neural network is selected as an initial model, and 4 different data sets are used for training to obtain 4 attention classification models, which are model 1, model 2, model 3 and model 4 in sequence. Due to sample differences in the data sets, the accuracy, convergence speed, etc. of the 4 attention classification models are different.
The accuracy can be set as a performance index according to needs, the convergence rate can also be set as a performance index, and a value obtained by weighting and summing the accuracy and the convergence rate can also be set as a performance index. And according to the set performance indexes, performing performance grading on each attention classification model, and taking the attention classification model with the highest performance grade as a basic model of each type. For example, the accuracy is set as a performance index, the performance of the 4 attention classification models obtained according to the convolutional neural network is scored, the accuracy of the model 3 is the highest, the performance score is also the highest, and the model 3 can be used as a basic model in a model of the convolutional neural network type.
And step 220, fusing the basic models of each type to obtain an attention classification fusion model.
Specifically, model fusion is to train multiple models, and fuse the multiple models into one model according to a certain method. The model fusion method comprises a linear weighted fusion method, a cross fusion method, a waterfall fusion method, a feature fusion method, a prediction fusion method and the like.
By fusing each type of basic model, the advantages of each different type of model can be fully utilized, the adaptability of the fusion model to different samples is improved, and the overall performance of the fusion model is improved.
Based on any of the above embodiments, fig. 3 is a schematic flowchart of the method for determining a basic model provided in the present application, and as shown in fig. 3, step 210 includes:
Specifically, the face training set is used for training each type of initial model, and the face verification set is used for adjusting the hyper-parameters of the initial model and evaluating the training effect of the initial model. The hyper-parameters are parameters which are preset before the initial model is trained.
The face training set and the face verification set may be determined from the same face data set. For example, a face training set and a face verification set can be obtained by splitting the same face data set. The sizes of the face training set and the face verification set can be set according to needs. According to different splitting modes, a plurality of face training sets and corresponding face verification sets can be obtained.
Specifically, for any type of initial model, a plurality of face training sets are adopted for training, and an attention classification model corresponding to each face training set is obtained respectively. For example, the initial model selection may be selected as DenseNet121, and different face training sets a1, a2, A3, a4 and a5 are used to train the DenseNet121, so as to obtain an attention classification model M1 corresponding to a1, an attention classification model M2 corresponding to a2, an attention classification model M3 corresponding to A3, an attention classification model M4 corresponding to a4 and an attention classification model M5 corresponding to a5, respectively.
And 2103, performing performance scoring on the attention classification model corresponding to each face training set based on the face verification set corresponding to each face training set.
Specifically, the face verification sets corresponding to the face training sets a1, a2, A3, a4 and a5 are B1, B2, B3, B4 and B5, respectively. Performance scoring was performed on the attention classification model M1 using B1, performance scoring was performed on the attention classification model M2 using B2, performance scoring was performed on the attention classification model M3 using B3, performance scoring was performed on the attention classification model M4 using B4, and performance scoring was performed on the attention classification model M5 using B5.
At step 2104, the attention classification model with the highest performance score is used as the base model corresponding to the type.
Specifically, the performance scores of the attention classification models M1, M2, M3, M4 and M5 were compared, and the attention classification model M3 with the highest score was used as the base model corresponding to DenseNet 121.
According to the attention recognition method provided by the embodiment of the application, the performance grading is carried out on the plurality of attention classification models of the same type to obtain the basic model corresponding to the type, and the influence of the plurality of training sets on the performance of the attention classification models is fully considered.
Based on any of the above embodiments, fig. 4 is a schematic flow chart of the cross-validation method provided in the present application, and as shown in fig. 4, step 2101 includes:
Specifically, the sample face data set may be obtained from a public data source, or a real photograph of the driver may be acquired to obtain the sample face data set.
Cross validation (Cross validation) takes out most samples from a given modeling sample to build a model, and leaves a small portion of samples to be forecasted by the just built model. K-fold cross validation (K-fold cross validation) may be used. The K-fold cross validation is to divide a data set into K parts in equal proportion, wherein one part is used as a validation set, and the other K-1 parts are used as a training set. Then, the experiment is performed once, and the K-fold cross validation is completed only after K times of the experiment, that is, the cross validation actually repeats the experiment for K times, each time of the experiment selects a different data part from the K parts as a validation set, the data of the K parts are ensured to be respectively tested, and the rest K-1 parts are taken as training sets.
For example, the sample face data set may be split into 5 subsets by an average splitting method, each subset is respectively used as a primary verification set, and the remaining 5 subsets are used as training sets, so as to obtain 5 training sets and face verification sets corresponding to the training sets.
By adopting the K-fold cross validation, an average value obtained by scoring a plurality of attention classification models of the same type by the K validation sets may be used as the score of the attention classification model of the type, which is not specifically limited in the embodiment of the present application.
According to the attention recognition method provided by the embodiment of the application, the cross verification method is adopted to determine the face training sets and the face verification sets corresponding to the face training sets, so that the over-fitting and under-fitting in model training are effectively avoided, and the accuracy of the attention classification model in driver attention recognition is improved.
Based on any of the above embodiments, fig. 5 is a schematic flow chart of the model fusion method provided in the present application, and as shown in fig. 5, step 220 includes:
and 2202, fusing the plurality of basic models based on the weight corresponding to each basic model to obtain an attention classification fusion model.
Specifically, a linear weighting method may be adopted to fuse a plurality of base models into one attention classification fusion model.
And determining a weight corresponding to each basic model according to the performance score of each type of basic model, and fusing the basic models by using a weighted summation mode to obtain the attention classification fusion model.
According to the attention recognition method provided by the embodiment of the application, the attention classification fusion model is obtained by fusing the plurality of basic models by adopting a linear weighting method, the complexity of the algorithm is reduced, and the method is simple and easy to implement.
Based on any of the above embodiments, fig. 6 is a schematic flow chart of the face image determination method provided by the present application, and as shown in fig. 6, step 110 includes:
Specifically, the image data including the face of the driver is sent to a face detection model, and the face detection model detects the face region in the image data to obtain a face image.
The face detection model may be pre-trained before step 1102 is performed. First, a large amount of sample image data including the face of the driver is collected. Secondly, a human face area mark in each sample image data is determined in a manual marking mode. Then, a large amount of sample image data and the face region marks in each sample image data are input into the initial model for training, so that the recognition capability of the initial model on the face regions in the image data is improved, and a face detection model is obtained.
Here, the initial model may select a Retina Face model. The Retina Face model adopts the technology of a characteristic pyramid, realizes multi-scale fusion and has an important effect on detecting small objects.
The inclusion V4 can be used as a feature extraction layer of the Retina Face model. The Incep V4 adopts a plurality of basic modules connected in series, each basic module comprises convolution kernels with different scales and different connection modes, and the features of input image data are extracted, so that human faces with different sizes can be detected.
For example, 3 basic modules, module 1, module 2, and module 3, respectively, may be provided in the feature extraction layer.
Fig. 7 is a schematic structural diagram of basic modules of the feature extraction layer provided in the present application, and as shown in fig. 7, the modules 1, 2, and 3 respectively employ convolution kernels of different scales, such as 1 × 1 convolution kernel, 3 × 3 convolution kernel, 7 × 1 convolution kernel, and 3 × 1 convolution kernel, and overlap in different connection manners. Firstly, the image data I is adjusted to 299 × 299 (pixels), then the input module 1 obtains a feature map F1, the feature map F1 is input into the module 2 to obtain a feature map F2, and the feature map F2 is input into the module 3 to obtain a feature map F3. And respectively proposing (proposal) to each layer in the three-layer feature maps of F1, F2 and F3, generating detection frames on 3 different scales, wherein each scale has anchors (anchors) with different sizes, and ensuring that human faces with different sizes can be detected.
In the training process of the face detection model, the loss function can adopt Focal loss to solve the problem of serious unbalance of positive and negative samples in target detection, and the loss function reduces the weight occupied by a large number of simple samples in training. Focal loss function LflCan be formulated as:
wherein y is a real label of the sample, y' is a predicted value of the sample, gamma is an influence factor, and gamma is more than 0, so that the loss of the sample which is easy to classify can be reduced, and the sample which is difficult to classify and is wrongly classified is concerned more; alpha is a balance factor to balance the importance of positive and negative samples. For example, γ is 2, and for the positive type sample, the prediction result is 0.95, which is a simple sample, so the γ power of (1-0.95) is small, and the loss function value becomes smaller. Whereas the loss is relatively large for samples with a prediction probability of 0.3. For negative class samples as well, the result of predicting 0.1 should be much smaller than the sample loss value of predicting 0.7. For a prediction probability of 0.5, the loss is reduced by only 0.25 times, so this indistinguishable sample is of greater interest. Therefore, the influence of simple samples is reduced, and the effect of overlapping a large number of samples with small prediction probability is more effective.
According to the attention recognition method provided by the embodiment of the application, the Face detection model is established based on the Retina Face model, the characteristic extraction layer is set as the inclusion V4, the Face image of a driver can be accurately extracted, and interference factors of image data are eliminated.
Based on any of the above embodiments, the plurality of different types of attention classification models includes at least one of DenseNet121, EffectientNet-B0, EffectientNet-B3, and EffectientNet-B7.
Specifically, DenseNet proposes a more aggressive dense connection mechanism than ResNet: i.e. interconnecting all layers, in particular each layer will accept as its additional input all layers in front of it. ResNet is the short-circuiting of each layer to a previous layer (typically 2-3 layers) by element-level addition. In DenseNet, each layer is connected (concat) dimensionally with all previous layers and serves as input for the next layer. For a L-layer network, DenseNet contains L (L +1)/2 connections in total, which is a dense connection compared to ResNet. And DenseNet is directly connected with feature maps from different layers, so that feature reuse can be realized, and efficiency is improved.
The EfficientNet has an excellent effect on image classification. A better Backbone network (Backbone) can be searched by utilizing the neural network, namely EfficientNet-B0. On the basis, in order to improve the complexity of the model and further improve the performance of the model, the model can be scaled appropriately. Common scaling methods are: increasing the width of the model, increasing the depth of the model, increasing the resolution of the input picture of the model, etc. The combined scaling method can be formulated as:
depth:d=αφ
width:w=βφ
resolution:r=γφ
s.t.α·β2·γ2≈2,α≥1,β≥1,γ≥1
where phi is a single combined scaling factor, scaling the width d, depth w and picture resolution r simultaneously. α, β, γ are scaling bases for depth, width, resolution.
In the above formula, α · β is restricted2·γ2≈2,α≥1,β≥1,γ≥1。
For EfficientNet-B0, the corresponding scaling basis is α ═ 1.2, β ═ 1.1, and γ ═ 1.15. EfficientNet-B0 through EfficientNet-B7 can be derived from the scaling factors of Table 1.
TABLE 1 Scale factor Table
DenseNet121, EfficientNet-B0, EfficientNet-B3 and EfficientNet-B7 can be selected as initial models to be trained to obtain a plurality of attention classification models of different types, and the training method comprises the following steps:
(1) adjusting the size of the face image to be the same size, such as 224 × 224 (pixel points), and obtaining eye marks (landmarks) in the face image;
(2) inputting the adjusted face image into DenseNet121, adding an eye mark, selecting cross entropy loss as a loss function for training, and obtaining a basic model 1 corresponding to DenseNet 121;
(3) inputting the adjusted face image into EfficientNet-B0, adding eye marks, selecting cross entropy loss as a loss function for training, and obtaining a basic model 2 corresponding to EfficientNet-B0;
(4) inputting the adjusted face image into EfficientNet-B3, adding eye marks, selecting cross entropy loss as a loss function for training, and obtaining a basic model 3 corresponding to EfficientNet-B3;
(5) inputting the adjusted face image into EfficientNet-B7, adding eye marks, selecting cross entropy loss as a loss function for training, and obtaining a basic model 4 corresponding to EfficientNet-B7;
(6) and inputting the adjusted face image into EfficientNet-B7, selecting cross entropy loss as a loss function for training without adding eye marks, and obtaining a basic model 5 corresponding to EfficientNet-B7.
According to the attention recognition method provided by the embodiment of the application, the DenseNet121, the EfficientNet-B0, the EfficientNet-B3 and the EfficientNet-B7 are adopted for training to obtain a plurality of attention classification models of different types, so that the attention classification fusion model obtained after fusion has high accuracy and generalization capability.
Based on any of the above embodiments, fig. 8 is a schematic flowchart of a training method of an attention classification fusion model provided in the present application, and as shown in fig. 8, the method includes:
inputting a plurality of image data containing the Face of a driver into a Retina Face model to obtain a Face image corresponding to each image data, wherein a feature extraction layer of the Retina Face model is expression V4;
performing five-fold cross validation on the obtained multiple face images to obtain 5 groups of training sets and validation sets corresponding to the training sets;
and step three, inputting the 5 groups of training sets and the corresponding verification sets into DenseNet121, EfficientNet-B0, EfficientNet-B3 and EfficientNet-B7 respectively for training. Wherein, 2 groups of training sets and corresponding verification sets are respectively input into EfficientNet-B7 for training, the first group is added with eye marks during training, and the second group is not added with eye marks during training. Obtaining 5 attention classification models of each type after training;
performing performance grading on the 5 attention classification models of each type, and selecting the attention classification model with the highest performance grading as a basic model of each type to obtain 5 basic models;
and step five, fusing the 5 basic models by adopting a linear weighting method to obtain an attention classification fusion model.
And inputting the face image of any driver into the attention classification fusion model to obtain the face direction angle of the driver.
The following describes the attention recognition device provided in the present application, and the attention recognition device described below and the attention recognition method described above may be referred to correspondingly.
Based on any of the above embodiments, fig. 9 is a schematic structural diagram of an attention recognition device provided in the present application, and as shown in fig. 9, the device includes:
a determination unit 910 configured to determine a face image of a driver to be recognized;
the recognition unit 920 is configured to input the face image into the attention classification fusion model to obtain a face direction angle output by the attention classification fusion model;
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
Specifically, the determination unit 910 determines a face image of the driver to be recognized. The recognition unit 920 inputs the face image into the attention classification fusion model to obtain the face direction angle output by the attention classification fusion model.
The attention recognition device provided by the application recognizes the input face image of the driver to be recognized through the attention classification fusion model to obtain the face direction angle of the driver. The attention classification fusion model is obtained by fusing a plurality of attention classification models of different types, so that the face characteristics of the driver can be fully recognized, the generalization capability is strong, and the accuracy of the driver attention recognition is improved.
Based on any of the above embodiments, fig. 10 is a schematic structural diagram of a fusion model determining unit provided in the present application, as shown in fig. 10, the apparatus further includes a fusion model determining unit 930, and the fusion model determining unit 930 includes:
a basic model determining subunit 9301, configured to select, from the multiple attention classification models of each type, an attention classification model with the highest performance score as a basic model of each type;
and a basic model fusion subunit 9302 for fusing the basic models of each type to obtain an attention classification fusion model.
Based on any of the above embodiments, fig. 11 is a schematic structural diagram of the basic model determining subunit provided in the present application, and as shown in fig. 11, the basic model determining subunit 9301 includes:
a training verification determining module 93011, configured to determine a plurality of face training sets and face verification sets corresponding to the face training sets;
the model training module 93012 is configured to train any type of initial model based on each face training set to obtain an attention classification model corresponding to each face training set;
the model scoring module 93013 is configured to perform performance scoring on the attention classification model corresponding to each face training set based on the face verification set corresponding to each face training set;
and the model screening module 93014 is used for taking the attention classification model with the highest performance score as a basic model corresponding to any type.
Based on any of the above embodiments, fig. 12 is a schematic structural diagram of a training verification determining module provided in the present application, and as shown in fig. 12, the training verification determining module 93011 includes:
a data set determination submodule 930111 for determining a sample face data set;
and the data set splitting submodule 930112 is configured to split the sample face data set into multiple subsets, and determine multiple face training sets and corresponding face verification sets by using a cross-validation method.
Based on any of the above embodiments, fig. 13 is a schematic structural diagram of the basic model fusion subunit provided in the present application, and as shown in fig. 13, the basic model fusion subunit 9302 includes:
a weight determination module 93021, configured to determine a weight corresponding to each basic model based on the performance score of each type of basic model;
and the model fusion module 93022 is configured to fuse the multiple basic models based on the weight corresponding to each basic model to obtain an attention classification fusion model.
Based on any of the above embodiments, fig. 14 is a schematic structural diagram of a determining unit provided in the present application, and as shown in fig. 14, the determining unit 910 includes:
an image acquiring subunit 9101, configured to acquire image data including a face of a driver;
a face detection subunit 9102, configured to input image data to a face detection model, so as to obtain a face image output by the face detection model; the Face detection model is established based on a Retina Face model, and the feature extraction layer of the Retina Face model is increment V4.
Based on any of the above embodiments, the plurality of different types of attention classification models includes at least one of DenseNet121, EffectientNet-B0, EffectientNet-B3, and EffectientNet-B7.
The attention recognition device provided in the embodiment of the present application is used for executing the above-mentioned attention recognition method, and the implementation manner of the attention recognition device is consistent with that of the attention recognition method provided in the present application, and the same beneficial effects can be achieved, and details are not described here.
Based on any of the above embodiments, fig. 15 is a schematic structural diagram of an electronic device provided in the present application, and as shown in fig. 15, the electronic device may include: a Processor (Processor)1510, a communication Interface (Communications Interface)1520, a Memory (Memory)1530 and a communication Bus (Communications Bus)1540, wherein the Processor 1510, the communication Interface 1520 and the Memory 1530 communicate with each other via the communication Bus 1540. The processor 1510 may call the logic command in the memory 1530 to execute the method provided by the above embodiments, the method includes:
determining a face image of a driver to be recognized; inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model; the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
In addition, the logic commands in the memory 1530 can be implemented in the form of software functional units and stored in a computer readable storage medium when the logic commands are sold or used as independent products. Based on such understanding, the technical solution of the present application or portions thereof that substantially contribute to the prior art may be embodied in the form of a software product stored in a storage medium and including commands for causing a computer device (which may be a personal computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The processor in the electronic device provided in the embodiment of the present application may call a logic instruction in a memory to implement the above attention identification method, and a specific implementation manner of the method is consistent with the method implementation manner and may achieve the same beneficial effects, which is not described herein again.
The present application also provides a non-transitory computer-readable storage medium, which is described below, and the non-transitory computer-readable storage medium described below and the attention recognition method described above are referred to in correspondence.
Embodiments of the present application provide a non-transitory computer-readable storage medium, on which a computer program is stored, where the computer program is implemented to perform the method provided in the foregoing embodiments when executed by a processor, and the method includes:
determining a face image of a driver to be recognized; inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model; the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
When the computer program stored on the non-transitory computer readable storage medium provided in the embodiment of the present application is executed, the above attention identification method is implemented, and the specific implementation manner is consistent with the method implementation manner and can achieve the same beneficial effects, which is not described herein again.
The above-described embodiments of the apparatus are merely illustrative, and the units described as separate parts may or may not be physically separate, and parts displayed as units may or may not be physical units, may be located in one place, or may be distributed on a plurality of network units. Some or all of the modules may be selected according to actual needs to achieve the purpose of the solution of the present embodiment. One of ordinary skill in the art can understand and implement it without inventive effort.
Through the above description of the embodiments, those skilled in the art will clearly understand that each embodiment can be implemented by software plus a necessary general hardware platform, and certainly can also be implemented by hardware. With this understanding in mind, the above technical solutions may be embodied in the form of a software product, which can be stored in a computer-readable storage medium, such as ROM/RAM, magnetic disk, optical disk, etc., and includes commands for causing a computer device (which may be a personal computer, a server, or a network device, etc.) to execute the method according to the embodiments or some parts of the embodiments.
Finally, it should be noted that: the above embodiments are only used to illustrate the technical solutions of the present application, and not to limit the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions in the embodiments of the present application.
Claims (16)
1. An attention recognition method, comprising:
determining a face image of a driver to be recognized;
inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model;
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
2. The attention recognition method according to claim 1, wherein the determination method of the attention classification fusion model comprises:
respectively selecting an attention classification model with the highest performance score from a plurality of attention classification models of each type as a basic model of each type;
and fusing the basic models of each type to obtain the attention classification fusion model.
3. The attention recognition method according to claim 2, wherein the step of selecting the attention classification model with the highest performance score from the plurality of attention classification models of each type as the base model of each type comprises:
determining a plurality of face training sets and corresponding face verification sets thereof;
training any type of initial model based on each face training set to obtain an attention classification model corresponding to each face training set;
performing performance scoring on the attention classification model corresponding to each face training set based on the face verification set corresponding to each face training set;
and taking the attention classification model with the highest performance score as a basic model corresponding to any type.
4. The attention recognition method of claim 3, wherein the determining a plurality of face training sets and corresponding face verification sets comprises:
determining a sample face data set;
and splitting the sample face data set into a plurality of subsets, and determining a plurality of face training sets and corresponding face verification sets by adopting a cross verification method.
5. The attention recognition method of claim 2, wherein the fusing the base models of each type to obtain the attention classification fusion model comprises:
determining a weight corresponding to each basic model based on the performance score of each type of basic model;
and fusing the plurality of basic models based on the weight corresponding to each basic model to obtain the attention classification fusion model.
6. The attention recognition method according to any one of claims 1 to 5, wherein the determining a face image of a driver to be recognized includes:
acquiring image data containing the face of the driver;
inputting the image data into a face detection model to obtain a face image output by the face detection model;
the Face detection model is established based on a Retina Face model, and a feature extraction layer of the Retina Face model is increment V4.
7. The attention recognition method of any one of claims 1 to 5, wherein the plurality of different types of attention classification models comprise at least one of DenseNet121, EfficientNet-B0, EfficientNet-B3, and EfficientNet-B7.
8. An attention recognition device, comprising:
a determination unit for determining a face image of a driver to be recognized;
the recognition unit is used for inputting the face image into an attention classification fusion model to obtain a face direction angle output by the attention classification fusion model;
the attention classification fusion model is obtained by fusing a plurality of attention classification models of different types.
9. The attention recognition device according to claim 8, further comprising a fusion model determination unit including:
the basic model determining subunit is used for respectively selecting the attention classification model with the highest performance score from the plurality of attention classification models of each type as the basic model of each type;
and the basic model fusion subunit is used for fusing the basic models of each type to obtain the attention classification fusion model.
10. The attention recognition device of claim 9, wherein the base model determining subunit comprises:
the training verification determining module is used for determining a plurality of face training sets and corresponding face verification sets;
the model training module is used for training any type of initial model based on each face training set to obtain an attention classification model corresponding to each face training set;
the model scoring module is used for performing performance scoring on the attention classification model corresponding to each face training set based on the face verification set corresponding to each face training set;
and the model screening module is used for taking the attention classification model with the highest performance score as the basic model corresponding to any type.
11. The attention recognition device of claim 10, wherein the training verification determination module comprises:
a data set determining submodule for determining a sample face data set;
and the data set splitting submodule is used for splitting the sample face data set into a plurality of subsets and determining a plurality of face training sets and corresponding face verification sets by adopting a cross verification method.
12. The attention recognition device of claim 9, wherein the base model fusion subunit comprises:
the weight determination module is used for determining the weight corresponding to each basic model based on the performance score of each type of basic model;
and the model fusion module is used for fusing the plurality of basic models based on the weight corresponding to each basic model to obtain the attention classification fusion model.
13. The attention recognition device according to claim 8, wherein the determination unit includes:
an image acquisition subunit configured to acquire image data including a face of the driver;
the face detection subunit is used for inputting the image data into a face detection model to obtain a face image output by the face detection model;
the Face detection model is established based on a Retina Face model, and a feature extraction layer of the Retina Face model is increment V4.
14. The attention recognition device of any one of claims 8 to 13, wherein the plurality of different types of attention classification models comprises at least one of DenseNet121, EfficientNet-B0, EfficientNet-B3, and EfficientNet-B7.
15. An electronic device comprising a memory, a processor and a computer program stored on the memory and executable on the processor, characterized in that the processor implements the steps of the attention recognition method according to any one of claims 1 to 7 when executing the computer program.
16. A non-transitory computer readable storage medium, having stored thereon a computer program, characterized in that the computer program, when being executed by a processor, realizes the steps of the attention recognition method according to any one of claims 1 to 7.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011438525.5A CN112597823A (en) | 2020-12-07 | 2020-12-07 | Attention recognition method and device, electronic equipment and storage medium |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011438525.5A CN112597823A (en) | 2020-12-07 | 2020-12-07 | Attention recognition method and device, electronic equipment and storage medium |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112597823A true CN112597823A (en) | 2021-04-02 |
Family
ID=75192308
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011438525.5A Pending CN112597823A (en) | 2020-12-07 | 2020-12-07 | Attention recognition method and device, electronic equipment and storage medium |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112597823A (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113743236A (en) * | 2021-08-11 | 2021-12-03 | 交控科技股份有限公司 | Passenger portrait analysis method, device, electronic equipment and computer readable storage medium |
CN114125155A (en) * | 2021-11-15 | 2022-03-01 | 天津市国瑞数码安全系统股份有限公司 | Crank call detection method and system based on big data analysis |
Citations (15)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104143327A (en) * | 2013-07-10 | 2014-11-12 | 腾讯科技(深圳)有限公司 | Acoustic model training method and device |
CN106250858A (en) * | 2016-08-05 | 2016-12-21 | 重庆中科云丛科技有限公司 | A kind of recognition methods merging multiple face recognition algorithms and system |
CN108776768A (en) * | 2018-04-19 | 2018-11-09 | 广州视源电子科技股份有限公司 | Image-recognizing method and device |
CN108875521A (en) * | 2017-12-20 | 2018-11-23 | 北京旷视科技有限公司 | Method for detecting human face, device, system and storage medium |
CN109409222A (en) * | 2018-09-20 | 2019-03-01 | 中国地质大学(武汉) | A kind of multi-angle of view facial expression recognizing method based on mobile terminal |
CN109446892A (en) * | 2018-09-14 | 2019-03-08 | 杭州宇泛智能科技有限公司 | Human eye notice positioning method and system based on deep neural network |
CN110110603A (en) * | 2019-04-10 | 2019-08-09 | 天津大学 | A kind of multi-modal labiomaney method based on facial physiologic information |
CN110796109A (en) * | 2019-11-05 | 2020-02-14 | 哈尔滨理工大学 | Driver distraction behavior identification method based on model fusion |
CN111339801A (en) * | 2018-12-19 | 2020-06-26 | 杭州海康威视系统技术有限公司 | Method, device, equipment and system for detecting attention of people |
US20200210768A1 (en) * | 2018-12-18 | 2020-07-02 | Slyce Acquisition Inc. | Training data collection for computer vision |
CN111523398A (en) * | 2020-03-30 | 2020-08-11 | 西安交通大学 | Method and device for fusing 2D face detection and 3D face recognition |
CN111680546A (en) * | 2020-04-26 | 2020-09-18 | 北京三快在线科技有限公司 | Attention detection method, attention detection device, electronic equipment and storage medium |
CN111796681A (en) * | 2020-07-07 | 2020-10-20 | 重庆邮电大学 | Self-adaptive sight estimation method and medium based on differential convolution in man-machine interaction |
US20200342214A1 (en) * | 2018-06-05 | 2020-10-29 | Tencent Technology (Shenzhen) Company Limited | Face recognition method and apparatus, classification model training method and apparatus, storage medium and computer device |
CN111860316A (en) * | 2020-07-20 | 2020-10-30 | 上海汽车集团股份有限公司 | Driving behavior recognition method and device and storage medium |
-
2020
- 2020-12-07 CN CN202011438525.5A patent/CN112597823A/en active Pending
Patent Citations (16)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104143327A (en) * | 2013-07-10 | 2014-11-12 | 腾讯科技(深圳)有限公司 | Acoustic model training method and device |
CN106250858A (en) * | 2016-08-05 | 2016-12-21 | 重庆中科云丛科技有限公司 | A kind of recognition methods merging multiple face recognition algorithms and system |
CN108875521A (en) * | 2017-12-20 | 2018-11-23 | 北京旷视科技有限公司 | Method for detecting human face, device, system and storage medium |
CN108776768A (en) * | 2018-04-19 | 2018-11-09 | 广州视源电子科技股份有限公司 | Image-recognizing method and device |
WO2019200902A1 (en) * | 2018-04-19 | 2019-10-24 | 广州视源电子科技股份有限公司 | Image recognition method and device |
US20200342214A1 (en) * | 2018-06-05 | 2020-10-29 | Tencent Technology (Shenzhen) Company Limited | Face recognition method and apparatus, classification model training method and apparatus, storage medium and computer device |
CN109446892A (en) * | 2018-09-14 | 2019-03-08 | 杭州宇泛智能科技有限公司 | Human eye notice positioning method and system based on deep neural network |
CN109409222A (en) * | 2018-09-20 | 2019-03-01 | 中国地质大学(武汉) | A kind of multi-angle of view facial expression recognizing method based on mobile terminal |
US20200210768A1 (en) * | 2018-12-18 | 2020-07-02 | Slyce Acquisition Inc. | Training data collection for computer vision |
CN111339801A (en) * | 2018-12-19 | 2020-06-26 | 杭州海康威视系统技术有限公司 | Method, device, equipment and system for detecting attention of people |
CN110110603A (en) * | 2019-04-10 | 2019-08-09 | 天津大学 | A kind of multi-modal labiomaney method based on facial physiologic information |
CN110796109A (en) * | 2019-11-05 | 2020-02-14 | 哈尔滨理工大学 | Driver distraction behavior identification method based on model fusion |
CN111523398A (en) * | 2020-03-30 | 2020-08-11 | 西安交通大学 | Method and device for fusing 2D face detection and 3D face recognition |
CN111680546A (en) * | 2020-04-26 | 2020-09-18 | 北京三快在线科技有限公司 | Attention detection method, attention detection device, electronic equipment and storage medium |
CN111796681A (en) * | 2020-07-07 | 2020-10-20 | 重庆邮电大学 | Self-adaptive sight estimation method and medium based on differential convolution in man-machine interaction |
CN111860316A (en) * | 2020-07-20 | 2020-10-30 | 上海汽车集团股份有限公司 | Driving behavior recognition method and device and storage medium |
Non-Patent Citations (5)
Title |
---|
A. GHOFRANI 等: "Attention-Based Face AntiSpoofing of RGB Camera using a Minimal End-2-End Neural Network", 《2020 INTERNATIONAL CONFERENCE ON MACHINE VISION AND IMAGE PROCESSING (MVIP), IRAN, 2020》, 15 June 2020 (2020-06-15), pages 1 - 6 * |
J. DENG 等: "RetinaFace: Single-Shot Multi-Level Face Localisation in the Wild", 《 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), SEATTLE, WA, USA, 2020》, 5 August 2020 (2020-08-05), pages 5202 - 5211 * |
W. HUANG 等: "Video-Based Abnormal Driving Behavior Detection via Deep Learning Fusions", 《IEEE ACCESS》, vol. 7, 16 May 2019 (2019-05-16), pages 64571 - 64582, XP011726072, DOI: 10.1109/ACCESS.2019.2917213 * |
YANJIA ZHU 等: "TinaFace: Strong but Simple Baseline for Face Detection", 《ARXIV:2011.13183.V1》, 26 November 2020 (2020-11-26), pages 1 - 9 * |
YU Z 等: "A Multi-Modal Approach for Driver Gaze Prediction to Remove Identity Bias", 《22ND ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, ICMI 2020. ASSOCIATION FOR COMPUTING MACHINERY, INC,2020》, 29 October 2020 (2020-10-29), pages 768 - 776, XP058728706, DOI: 10.1145/3382507.3417961 * |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN113743236A (en) * | 2021-08-11 | 2021-12-03 | 交控科技股份有限公司 | Passenger portrait analysis method, device, electronic equipment and computer readable storage medium |
CN114125155A (en) * | 2021-11-15 | 2022-03-01 | 天津市国瑞数码安全系统股份有限公司 | Crank call detection method and system based on big data analysis |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3779774B1 (en) | Training method for image semantic segmentation model and server | |
US11842487B2 (en) | Detection model training method and apparatus, computer device and storage medium | |
CN110826530B (en) | Face detection using machine learning | |
CN108304876B (en) | Classification model training method and device and classification method and device | |
CN106960195B (en) | Crowd counting method and device based on deep learning | |
US9779492B1 (en) | Retinal image quality assessment, error identification and automatic quality correction | |
CN104239858B (en) | A kind of method and apparatus of face characteristic checking | |
CN110276745B (en) | Pathological image detection algorithm based on generation countermeasure network | |
CN107851192B (en) | Apparatus and method for detecting face part and face | |
CN106127164A (en) | The pedestrian detection method with convolutional neural networks and device is detected based on significance | |
CN110503000B (en) | Teaching head-up rate measuring method based on face recognition technology | |
CN103136504A (en) | Face recognition method and device | |
CN107992807B (en) | Face recognition method and device based on CNN model | |
CN109919252A (en) | The method for generating classifier using a small number of mark images | |
CN112990097A (en) | Face expression recognition method based on countermeasure elimination | |
CN113111968B (en) | Image recognition model training method, device, electronic equipment and readable storage medium | |
CN112597823A (en) | Attention recognition method and device, electronic equipment and storage medium | |
CN110598638A (en) | Model training method, face gender prediction method, device and storage medium | |
CN110543848B (en) | Driver action recognition method and device based on three-dimensional convolutional neural network | |
CN110991513A (en) | Image target recognition system and method with human-like continuous learning capability | |
CN111160239A (en) | Concentration degree evaluation method and device | |
CN112307984B (en) | Safety helmet detection method and device based on neural network | |
CN109993021A (en) | The positive face detecting method of face, device and electronic equipment | |
CN114333046A (en) | Dance action scoring method, device, equipment and storage medium | |
CN111797705A (en) | Action recognition method based on character relation modeling |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |