WO2021012526A1 - Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage - Google Patents

Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage Download PDF

Info

Publication number
WO2021012526A1
WO2021012526A1 PCT/CN2019/118461 CN2019118461W WO2021012526A1 WO 2021012526 A1 WO2021012526 A1 WO 2021012526A1 CN 2019118461 W CN2019118461 W CN 2019118461W WO 2021012526 A1 WO2021012526 A1 WO 2021012526A1
Authority
WO
WIPO (PCT)
Prior art keywords
face recognition
image
output
network
feature extraction
Prior art date
Application number
PCT/CN2019/118461
Other languages
English (en)
Chinese (zh)
Inventor
姚旭峰
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021012526A1 publication Critical patent/WO2021012526A1/fr

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/30Noise filtering
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/172Classification, e.g. identification

Definitions

  • This application relates to the field of biometrics, and in particular to a method for training a face recognition model, a face recognition method, device, equipment and storage medium.
  • Face recognition technology refers to the recognition of the position of a face in a picture or a video through a face recognition model.
  • the existing face recognition model mainly adopts the transfer learning method for training to accelerate the training speed.
  • a classification layer is often added after the presentation layer of the network. Because the parameter distributions of the presentation layer and the classification layer are inconsistent, there is a problem that gradient explosions are prone to occur, resulting in poor model stability.
  • This application provides a face recognition model training method, face recognition method, device, equipment and storage medium.
  • the method can increase the speed of face recognition and avoid gradients caused by inconsistent parameter distributions between the feature extraction network and the classification network
  • the problem of explosion improves the stability of the model.
  • this application provides a method for training a face recognition model, the method including:
  • this application also provides a face recognition method, which includes:
  • first prompt information for prompting the user to successfully recognize the image to be recognized is displayed.
  • the present application also provides a training device for a face recognition model, the training device includes:
  • the feature training unit is used to train a preset convolutional neural network according to the first sample image to construct a feature extraction network
  • a network connection unit configured to establish a connection between the feature extraction network and a preset classification network to obtain a first convolutional neural network model
  • the classification training unit is configured to iteratively train the classification network in the first convolutional neural network model according to the second sample image to adjust the weight parameters of the classification network in the first convolutional neural network model , Thereby obtaining the second convolutional neural network model;
  • a network unfreezing unit configured to unfreeze the weight parameters of the feature extraction network of the second convolutional neural network model
  • the model training unit is used to train the thawed second convolutional neural network model according to the third sample image to obtain the face recognition model.
  • the present application also provides a face recognition device, which includes:
  • the image recognition unit is used to obtain the image to be recognized
  • the image input unit is configured to input the image to be recognized into a preset face recognition model to obtain a face recognition result, and the face recognition model is determined by the face recognition according to any one of claims 1-5 Trained by the training method of the model;
  • the first information display unit is configured to, if the face recognition result indicates that the image to be recognized is successfully recognized, display first prompt information for prompting the user to successfully recognize the image to be recognized.
  • the present application also provides a computer device, the computer device includes a memory and a processor; the memory is used to store a computer program; the processor is used to execute the computer program and execute the The computer program implements the above-mentioned face recognition model training method.
  • the present application also provides a computer-readable storage medium that stores a computer program, and when the computer program is executed by a processor, the processor realizes the aforementioned face recognition The training method of the model.
  • This application discloses a method, device, device and storage medium for training a face recognition model.
  • a preset convolutional neural network is trained according to the first sample image information to construct a feature extraction network;
  • the feature extraction network establishes a connection with the preset classification network to obtain the first convolutional neural network model; freezes the weight parameters of the feature extraction network of the first convolutional neural network model; according to the second sample image information, the
  • the classification network in the first convolutional neural network model performs iterative training to obtain the second convolutional neural network model; unfreezes the weight parameters of the feature extraction network of the second convolutional neural network model; according to the third sample image information, Training the thawed second convolutional neural network model to obtain the face recognition model.
  • FIG. 1 is a schematic flowchart of a method for training a face recognition model provided by an embodiment of the present application
  • FIG. 2 is a schematic flowchart of a method for training a face recognition model provided by another embodiment of the present application
  • FIG. 3 is a schematic flowchart of sub-steps of a method for training a face recognition model provided by an embodiment in FIG. 2;
  • FIG. 4 is a schematic flowchart of the sub-steps of the method for training a face recognition model provided in another embodiment of FIG. 2;
  • FIG. 5 is a schematic flowchart of sub-steps of the training method of the face recognition model in FIG. 2;
  • Fig. 6 is a schematic flowchart of a method for training a face recognition model provided by still another embodiment of the present application.
  • FIG. 7 is a schematic flowchart of a face recognition method provided by an embodiment of the present application.
  • FIG. 8 is a schematic diagram of an application scenario of a face recognition method provided by an embodiment of the present application.
  • FIG. 9 is a schematic diagram of an application scenario of a face recognition method provided by another embodiment of the present application.
  • FIG. 10 is a schematic block diagram of a training device for a face recognition model provided by an embodiment of the present application.
  • FIG. 11 is a schematic block diagram of an apparatus for training a face recognition model provided by another embodiment of the present application.
  • FIG. 12 is a schematic block diagram of a subunit of a training device for a face recognition model provided by an embodiment of the present application.
  • FIG. 13 is a schematic block diagram of a face recognition device provided by an embodiment of the present application.
  • FIG. 14 is a schematic block diagram of the structure of a computer device according to an embodiment of the application.
  • the embodiments of the present application provide a method for training a face recognition model, a face recognition method, device, equipment, and storage medium.
  • the training method of the face recognition model can be used to train the face recognition model, which can increase the speed of face recognition and avoid the problem of gradient explosion caused by the inconsistent parameter distribution of the face recognition model, thereby improving the stability of the face recognition model.
  • FIG. 1 is a schematic flowchart of steps of a method for training a face recognition model according to an embodiment of the present application.
  • the face recognition model training method is used to train the face recognition model to avoid the problem of gradient explosion caused by the inconsistent parameter distribution of the face recognition model, thereby improving the stability of the model.
  • the training method of the face recognition model specifically includes: step S110 to step S160.
  • the first sample image is an image collected in advance.
  • the pre-collected image can be a directly collected image or an image obtained from a video.
  • the position of the face in the first sample image is marked as the first true label.
  • the feature extraction network is used to extract image features from the image of the input feature extraction network.
  • the feature extraction network can include a number of convolutional layers. Of course, the pooling layer may or may not be included. After the image is input to the feature extraction network, each convolutional layer in the feature extraction network performs convolution processing on the input image layer by layer, and the last convolution layer in the feature extraction network outputs the image features of the input image.
  • the feature extraction network includes five convolutional layers
  • the first convolutional layer conv1 includes 96 11 ⁇ 11 convolution kernels
  • the second convolutional layer conv2 includes 256 5 ⁇ 5 convolution kernels
  • the third The convolutional layer conv3 and the fourth convolutional layer conv4 both include 384 3 ⁇ 3 convolution kernels
  • the fifth convolutional layer conv5 includes 256 3 ⁇ 3 convolution kernels.
  • the convolutional layer and the fifth convolutional layer are connected to a 2 ⁇ 2 pooling layer, and each layer is connected to a modified linear unit.
  • a pre-trained model such as YOLO9000 can be used as the preset convolutional neural network.
  • the method before training a preset convolutional neural network based on the first sample image information to construct a feature extraction network, the method further includes:
  • S101 Obtain a sample video, and determine a sample image set in the sample video, where the sample image set includes first sample image information, second sample image information, and third sample image information.
  • a camera can be used to collect sample video on the target task. After the camera collects the sample video, the terminal or the server can obtain the sample video and determine the sample image set in the sample video.
  • the sample image set may be divided into at least three subsets, which are a first image subset, a second image subset, and a third image subset.
  • the first image subset is a set of first sample image information.
  • the second image subset is a set of second sample image information.
  • the third image subset is a set of third sample image information.
  • the determining the sample image set in the sample video includes:
  • S1011a Perform framing processing on the sample video to obtain several single-frame images.
  • the sample video is composed of successive pictures, and each picture is a frame.
  • S1011b If there is a face image in the single frame image, perform wavelet threshold denoising processing on the single frame image.
  • denoising processing is performed on a single frame image with a face image to effectively remove noise and reduce the influence of noise generated by the imaging device and the external environment, thereby improving the quality of the sample image set.
  • the method further includes: judging whether there is a face image in each single frame image.
  • the judging whether there is a face image in each single frame image specifically includes: detecting whether there is a position of a key part of the face in each single frame image; if there is a preset in each single frame image For key parts of the face, it is determined that there is a face image in each of the single frame images; if there is no preset key part of the face in each of the single frame images, it is determined that there is no face image in each of the single frame images.
  • a single frame image without a face image is removed to ensure that all the sample images in the sample image set have face images, thereby improving the effectiveness of the sample image set.
  • the first sample image information may be an original image directly collected by an image collection device such as a camera.
  • training a preset convolutional neural network according to the first sample image information to construct a feature extraction network specifically includes:
  • S1012a Acquire first original image information, second original image information, and third original image information.
  • the first original image information, the second original image information, and the third original image information are images directly collected in advance, and may also be images obtained in advance from a video.
  • S1012b If there is a face image in the first original image information, perform wavelet threshold denoising processing on the first original image information to obtain first sample image information.
  • denoising processing is performed on the first original image information with the face image to effectively remove noise and reduce the influence of noise generated by the imaging device and the external environment, thereby improving the quality of the first sample image information.
  • the method further includes: determining whether there is a face image in the first original image information.
  • the judging whether there is a face image in the first original image information specifically includes: detecting whether there is a position of a key part of the face in the first original image information; if each of the first original image information exists Preset key parts of the face, and determine that there is a face image in the first original image information; if the preset key parts of the face do not exist in the first original image information, it is determined that there is no face image in the first original image information. If there is no face image in the first original image information, remove the first original image information to ensure that the first sample image information has a face image, thereby improving the validity of the first sample image information .
  • S1012c If there is a face image in the second original image information, perform wavelet threshold denoising processing on the second original image information to obtain second sample image information.
  • denoising processing is performed on the second original image information with the face image to effectively remove noise and reduce the influence of noise generated by the imaging device and the external environment, thereby improving the quality of the second sample image information.
  • the method further includes: determining whether there is a face image in the second original image information.
  • the judging whether there is a face image in the second original image information specifically includes: detecting whether there is a position of a key part of the face in the second original image information; if each of the second original image information exists Preset key parts of the face, and determine that there is a face image in the second original image information; if the preset key parts of the face do not exist in the second original image information, determine that there is no face image in the second original image information. If there is no face image in the second original image information, the second original image information is removed to ensure that the second sample image information has a face image, thereby improving the effectiveness of the second sample image information.
  • S1012d If there is a face image in the third original image information, perform wavelet threshold denoising processing on the third original image information to obtain third sample image information.
  • denoising processing is performed on the third original image information with the face image to effectively remove noise and reduce the influence of noise generated by the imaging device and the external environment, thereby improving the quality of the third sample image information.
  • the method further includes: determining whether there is a face image in the third original image information.
  • the judging whether there is a face image in the third original image information specifically includes: detecting whether there is a position of a key part of the face in the third original image information; if each of the third original image information exists Presetting the key parts of the face, and determining that there is a face image in the third original image information; if the preset key parts of the face do not exist in the third original image information, it is determined that there is no face image in the third original image information. If there is no face image in the third original image information, remove the third original image information to ensure that all the third sample image information has a face image, thereby improving the effectiveness of the third sample image information.
  • a preset classification network is added after the feature extraction network, and the output of the feature extraction network is used as the input of the classification network, so that the feature extraction network establishes a connection with the classification network to obtain The first convolutional neural network model.
  • the classification network includes a convolutional layer, a fully connected layer, and a classifier that are sequentially connected.
  • the step S120 to establish a connection between the feature extraction network and the preset classification network specifically includes sub-step S121, sub-step S122, and sub-step S123.
  • Sub-step S121 input the output of the feature extraction network to the convolutional layer.
  • the output of the feature extraction network can be input to the convolutional layer of the classification network.
  • Sub-step S122 input the output of the convolutional layer to the fully connected layer, so as to reduce the dimensionality of the output of the convolutional layer.
  • the output of the convolutional layer of the classification network is input to the fully connected layer of the classification network, so that the dimensionality of the output of the convolutional layer is reduced.
  • the inputting the output of the convolutional layer to the fully connected layer to reduce the dimensionality of the output of the convolutional layer includes:
  • the weight calculation formula Based on the weight calculation formula, a fully connected layer operation is performed on each feature value of the output of the convolution layer to reduce the dimensionality of the output of the convolution layer; the weight calculation formula is:
  • the loss function is the mean square error MSE function
  • W represents the weight of the convolutional layer
  • W i represents the ith weight of the convolutional layer
  • h represents the bias of the convolutional layer
  • h i represents the ith bias of the convolutional layer
  • X represents the entire sample image set
  • X(i) represents the first true label corresponding to the i-th sample image
  • represents the learning efficiency of the back propagation algorithm.
  • a fully connected layer operation is performed on each feature value of the output of the convolutional layer through a backpropagation algorithm, thereby reducing the dimensionality of the output of the convolutional layer.
  • S123 Use the classifier to classify the output of the fully connected layer to establish a connection between the feature extraction network and the classification network.
  • the method before establishing a connection between the feature extraction network and a preset classification network to obtain the first convolutional neural network model, the method further includes:
  • the weight parameter of the feature extraction network is composed of the weight parameters of each layer of the feature extraction network, that is, each layer of the feature extraction network has a weight parameter, and the set of weight parameters of each layer forms the weight parameter of the feature extraction network.
  • the method before establishing a connection between the feature extraction network and a preset classification network to obtain the first convolutional neural network model, the method further includes:
  • S103 Determine whether the error value between the output of the feature extraction network and the first real tag is less than a first preset threshold.
  • the position of the target sample face in the first sample image is labeled as the first real label.
  • the first preset threshold can be set according to actual needs, for example, set to 0.01.
  • step S120 is executed, that is, the feature extraction network is connected to the preset classification network .
  • step S110 If the error value between the output of the feature extraction network and the first real label is greater than or equal to the first preset threshold, return to step S110 and continue to train the preset convolutional neural network until all The error value between the output of the feature extraction network and the first real tag is less than the first preset threshold.
  • S130 Freeze the weight parameters of the feature extraction network of the first convolutional neural network model.
  • the first The weight parameters of the feature extraction network of the convolutional neural network model will not change accordingly.
  • S140 Perform iterative training on the classification network in the first convolutional neural network model according to the second sample image to obtain a second convolutional neural network model.
  • the second sample image is a pre-collected image including the target sample face.
  • the pre-collected image can be a directly collected image or an image obtained from a video.
  • step S150 before the unfreezing the weight parameters of the feature extraction network of the second convolutional neural network model, further includes:
  • S104 Determine whether the error value between the output of the classification network of the second convolutional neural network model and the second real label is less than a second preset threshold.
  • the target sample face region in the second sample image is labeled as the second real label.
  • the second preset threshold can be set according to actual needs, for example, set to 0.005.
  • step S150 is executed, that is, the second convolutional neural network is unfreezed
  • the features of the network model extract the weight parameters of the network.
  • step S140 If the error value between the output of the classification network of the second convolutional neural network model and the second real label is greater than or equal to the second preset threshold, return to step S140 and continue to compare the first volume
  • the classification network in the product neural network model is iteratively trained until the error value between the output of the classification network of the second convolutional neural network model and the second true label is less than the second preset threshold.
  • the third sample image is a pre-collected image including the target sample face.
  • the pre-collected image can be a directly collected image or an image obtained from a video.
  • the feature extraction network and the classification network of the thawed second convolutional neural network are jointly trained according to the third sample image, so as to improve the performance of the second convolutional neural network.
  • the weight parameters of the feature extraction network and the weight parameters of the classification network are jointly adjusted until convergence, and a face recognition model is obtained.
  • the weight parameters of the feature extraction training network and the weight parameters of the classification network are determined, so as to determine all the parameters of the face recognition model, and obtain the Face recognition model. Specifically, the face area in the third sample image is labeled to obtain the third true label.
  • the aforementioned method for training a face recognition model is to train a preset convolutional neural network according to the first sample image information to construct a feature extraction network; establish a connection between the feature extraction network and the preset classification network, To obtain the first convolutional neural network model; freeze the weight parameters of the feature extraction network of the first convolutional neural network model; perform the classification network in the first convolutional neural network model according to the second sample image information Iterative training to obtain the second convolutional neural network model; unfreeze the weight parameters of the feature extraction network of the second convolutional neural network model; perform processing on the thawed second convolutional neural network model according to the third sample image information Training to obtain the face recognition model.
  • the weight parameter update of the face recognition model is smoother during training, so that the face recognition model is more robust. At the same time, it is easy to reach the optimal value in the process of backpropagation update parameters, which improves the stability of the model.
  • FIG. 8 is a schematic flowchart of steps of a face recognition method according to an embodiment of the present application.
  • FIG. 9 is a schematic diagram of an application scenario of a face recognition method provided by an embodiment of the present application.
  • the face recognition method can be applied to a system including terminal devices 310 and 320, network 330 and server 340.
  • the network 340 is used to provide a medium of communication links between the terminal devices 310 and 320 and the server 340.
  • the network 330 may include various connection types, such as wired, wireless communication links, or fiber optic cables.
  • the user can use the terminal devices 310 and 320 to interact with the server 340 via the network 330 to receive or send request instructions and the like.
  • Various communication client applications such as image processing applications, web browser applications, search applications, instant messaging tools, etc., may be installed on the terminal devices 310 and 320.
  • the terminal devices 310 and 320 may be various electronic devices with a display screen and supporting web browsing, including but not limited to smart phones, tablet computers, laptop portable computers, desktop computers, and so on.
  • the server 340 may be a server that provides various services, for example, a background management server that provides support for teaching websites browsed by users using the terminal devices 310 and 320.
  • the background management server can analyze and process the received product information query request and other data, and feed back the processing result to the terminal devices 310 and 320.
  • the face recognition method specifically includes: step 210 to step 230.
  • the image to be recognized includes a face target to be recognized, which may be a visible light image, such as an image in an RGB (Red Green Blue) mode.
  • a visible light image such as an image in an RGB (Red Green Blue) mode.
  • the aforementioned image to be recognized may also be a near infrared (Near Infrared, NIR) image.
  • the execution subject of this embodiment may be installed with a camera for collecting visible light images and a camera for collecting near-infrared images.
  • the user can select the camera to be turned on, and then use the selected camera to take a picture (using a self-portrait of the user's head or face) to obtain the image to be recognized.
  • the image to be recognized After the image to be recognized is obtained, the image to be recognized can be input to a pre-trained face recognition model to obtain a face recognition result.
  • the preset face recognition model is a face recognition model obtained by training using the training method of the aforementioned face recognition model.
  • the terminal device may display first prompt information for prompting the user to successfully recognize the image to be recognized. For example, the character string "Recognition passed" is displayed.
  • step S230 if the result of the face recognition indicates that the face to be recognized is successfully recognized.
  • the image after displaying the first prompt information for prompting the user to successfully recognize the image to be recognized, further includes:
  • the execution subject may display second prompt information for prompting the user to reacquire the image to be recognized. For example, the character string "Please reacquire the image" is displayed.
  • the user's characteristic information may be pre-stored in the terminal device, and the pre-stored characteristic information may be extracted from the face image uploaded by the user during registration.
  • the terminal device may use the feature information extracted from the image to be recognized using the aforementioned face recognition model as the face recognition result. If the face recognition result does not match the pre-stored feature information (for example, the similarity is less than a certain predetermined Set the value), then it can be determined that the recognition of the image to be recognized fails.
  • the face recognition model can use the image to be detected.
  • the extracted feature information is quite different from the pre-stored feature information. At this time, the face recognition result may indicate that the image to be recognized cannot be recognized.
  • the aforementioned face recognition method may be used to perform face recognition login.
  • the camera of the terminal device can collect the face image of the user to be logged in, and compare the face image of the user to be logged in with the facial images of all users who have registered on the teaching application platform or teaching website to control the user log in.
  • the face image of the user to be logged in can be used as the image to be recognized.
  • the image to be recognized can be preprocessed.
  • the preprocessing process here may include a face image alignment process.
  • the face alignment process mainly includes face detection, face key point positioning, and then the detected face key points in all images are as close as possible to the preset face key point positions, and finally the person is cut out from the image Face area and adjust the resolution of the face area to a predetermined size, such as 224 ⁇ 224. Next, you can perform specific operations on the preprocessed image to be recognized.
  • the face recognition method described above obtains an image to be recognized; inputs the image to be recognized into a preset face recognition model to obtain a face recognition result; if the face recognition result indicates that the image to be recognized is successfully recognized, Display the first prompt message for prompting the user to successfully recognize the image to be recognized.
  • This method can quickly recognize the face of the image to be recognized, and at the same time has the advantages of high recognition accuracy.
  • FIG. 10 is a schematic block diagram of an apparatus for training a face recognition model provided by an embodiment of the present application.
  • the training apparatus for a face recognition model may be configured in a server for performing any of the foregoing The training method of face recognition model.
  • the training device 300 for a face recognition model includes:
  • the feature training unit 310 is configured to train a preset convolutional neural network according to the first sample image to construct a feature extraction network
  • the network connection unit 320 is configured to connect the feature extraction network with a preset classification network to obtain a first convolutional neural network model
  • the parameter freezing unit 330 is configured to freeze the weight parameters of the feature extraction network of the first convolutional neural network model
  • the classification training unit 340 is configured to perform iterative training on the classification network in the first convolutional neural network model according to the second sample image to perform weight parameters of the classification network in the first convolutional neural network model Adjust to obtain the second convolutional neural network model;
  • the network unfreezing unit 350 is configured to unfreeze the weight parameters of the feature extraction network of the second convolutional neural network model
  • the model training unit 360 is configured to train the thawed second convolutional neural network model according to the third sample image to obtain the face recognition model.
  • the training device 300 for the face recognition model further includes an output judgment unit 370 for judging whether the error value between the output of the feature extraction network and the first real label is smaller than the first true label.
  • a preset threshold for judging whether the error value between the output of the feature extraction network and the first real label is smaller than the first true label.
  • the network connection unit 320 is specifically configured to establish the feature extraction network and the classification network if the error value between the output of the feature extraction network and the first real label is less than the first preset threshold Connected to obtain the first convolutional neural network model.
  • the classification network includes a convolutional layer, a fully connected layer, and a classifier that are sequentially connected.
  • the network connection unit 320 includes a convolution input subunit 321, a connection input subunit 322, and a classification processing subunit 323.
  • the convolution input subunit 321 is configured to input the output of the feature extraction network to the convolution layer;
  • connection input subunit 322 is configured to input the output of the convolutional layer to the fully connected layer, so as to reduce the dimensionality of the output of the convolutional layer;
  • the classification processing subunit 323 is configured to use the classifier to classify the output of the fully connected layer to establish a connection between the feature extraction network and the classification network.
  • connection input subunit 322 is specifically configured to perform a fully connected layer operation on each feature value of the output of the convolutional layer based on a weight calculation formula, so as to perform a calculation on the output of the convolutional layer. Dimensionality reduction.
  • the weight calculation formula is:
  • the loss function is the mean square error MSE function
  • W represents the weight of the convolutional layer
  • W i represents the ith weight of the convolutional layer
  • h represents the bias of the convolutional layer
  • h i represents the ith bias of the convolutional layer
  • X represents the entire sample image set
  • X(i) represents the first true label corresponding to the i-th sample image
  • represents the learning efficiency of the back propagation algorithm.
  • FIG. 13 is a schematic block diagram of a face recognition device according to an embodiment of the present application, and the face recognition device is used to perform any of the aforementioned methods for training a face recognition model.
  • the face recognition device can be configured in a server or a terminal.
  • the face recognition device 400 includes: an image acquisition unit 410, an image input unit 420, and an information display unit 430.
  • the image acquisition unit 410 is configured to acquire an image to be recognized
  • the image input unit 420 is configured to input the image to be recognized into a preset face recognition model to obtain a face recognition result;
  • the information display unit 430 is configured to display first prompt information for prompting the user to successfully recognize the image to be recognized if the face recognition result indicates that the image to be recognized is successfully recognized.
  • the above-mentioned apparatus can be implemented in the form of a computer program, and the computer program can be run on the computer device as shown in FIG. 14.
  • FIG. 14 is a schematic block diagram of a computer device according to an embodiment of the present application.
  • the computer equipment can be a server or a terminal.
  • the computer device includes a processor, a memory, and a network interface connected through a system bus, where the memory may include a non-volatile storage medium and an internal memory.
  • the non-volatile storage medium can store an operating system and a computer program.
  • the computer program includes program instructions.
  • the processor can execute a method for training a face recognition model.
  • the processor is used to provide computing and control capabilities and support the operation of the entire computer equipment.
  • the internal memory provides an environment for the operation of the computer program in the non-volatile storage medium.
  • the processor can execute a method for training a face recognition model.
  • the network interface is used for network communication, such as sending assigned tasks.
  • the network interface is used for network communication, such as sending assigned tasks.
  • FIG. 14 is only a block diagram of part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
  • the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
  • the processor may be a central processing unit (Central Processing Unit, CPU), and the processor may also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), and application specific integrated circuits (Application Specific Integrated Circuits). Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
  • the general-purpose processor may be a microprocessor or the processor may also be any conventional processor.
  • the processor is used to run a computer program stored in the memory to implement the following steps:
  • the processor before the processor realizes the establishment of the connection between the feature extraction network and the preset classification network to obtain the first convolutional neural network model, the processor is configured to realize:
  • the processor realizes the connection between the feature extraction network and the preset classification network to obtain the first convolutional neural network model, it is used to realize:
  • the classification network includes a convolutional layer, a fully connected layer, and a classifier that are sequentially connected.
  • the processor realizes the establishment of the connection between the feature extraction network and the preset classification network, it is used to realize:
  • the output of the fully connected layer is classified to establish a connection between the feature extraction network and the classification network.
  • the processor when the processor implements the input of the output of the convolutional layer to the fully connected layer to reduce the dimensionality of the output of the convolutional layer, the processor is used to implement:
  • a fully connected layer operation is performed on each feature value of the output of the convolutional layer, so as to reduce the dimensionality of the output of the convolutional layer.
  • the weight calculation formula is:
  • the loss function is the mean square error MSE function
  • W represents the weight of the convolutional layer
  • W i represents the ith weight of the convolutional layer
  • h represents the bias of the convolutional layer
  • h i represents the ith bias of the convolutional layer
  • X represents the entire sample image set
  • X(i) represents the first true label corresponding to the i-th sample image
  • represents the learning efficiency of the back propagation algorithm.
  • the processor is used to run a computer program stored in the memory to implement the following steps:
  • the embodiments of the present application also provide a computer-readable storage medium, the computer-readable storage medium stores a computer program, the computer program includes program instructions, and the processor executes the program instructions to implement the present application Any one of face recognition model training methods or face recognition methods provided in the embodiments.
  • the computer-readable storage medium may be the internal storage unit of the computer device described in the foregoing embodiment, such as the hard disk or memory of the computer device.
  • the computer-readable storage medium may also be an external storage device of the computer device, such as a plug-in hard disk, a smart memory card (SMC), or a secure digital (Secure Digital, SD) equipped on the computer device. ) Card, Flash Card, etc.

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Health & Medical Sciences (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Oral & Maxillofacial Surgery (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Biomedical Technology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Human Computer Interaction (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

La présente invention concerne un procédé d'apprentissage de modèle de reconnaissance faciale, un procédé et un appareil de reconnaissance faciale, un dispositif, et un support de stockage. Le procédé comprend les étapes consistant à : entraîner un réseau neuronal convolutif prédéfini ; établir une connexion entre un réseau d'extraction de caractéristiques et un réseau de classification ; geler un paramètre de pondération du réseau d'extraction de caractéristiques ; effectuer un apprentissage itératif sur le réseau de classification ; dégeler un paramètre de pondération d'un réseau d'extraction de caractéristiques d'un second modèle de réseau neuronal convolutif ; et entraîner le second modèle de réseau neuronal convolutif non gelé pour obtenir un modèle de reconnaissance faciale.
PCT/CN2019/118461 2019-07-22 2019-11-14 Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage WO2021012526A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910663230.9 2019-07-22
CN201910663230.9A CN110543815B (zh) 2019-07-22 2019-07-22 人脸识别模型的训练方法、人脸识别方法、装置、设备及存储介质

Publications (1)

Publication Number Publication Date
WO2021012526A1 true WO2021012526A1 (fr) 2021-01-28

Family

ID=68709979

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/118461 WO2021012526A1 (fr) 2019-07-22 2019-11-14 Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage

Country Status (2)

Country Link
CN (1) CN110543815B (fr)
WO (1) WO2021012526A1 (fr)

Cited By (30)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113065512A (zh) * 2021-04-21 2021-07-02 深圳壹账通智能科技有限公司 人脸微表情识别方法、装置、设备及存储介质
CN113168573A (zh) * 2021-03-02 2021-07-23 深圳市锐明技术股份有限公司 一种模型的训练方法、装置、终端设备及存储介质
CN113156439A (zh) * 2021-05-08 2021-07-23 中国石油大学(华东) 一种基于数据驱动的sar风场海浪联合反演方法及系统
CN113158850A (zh) * 2021-04-07 2021-07-23 大连海事大学 基于深度学习的船舶驾驶员疲劳检测方法和系统
CN113221830A (zh) * 2021-05-31 2021-08-06 平安科技(深圳)有限公司 一种超分活体识别方法、系统、终端及存储介质
CN113283978A (zh) * 2021-05-06 2021-08-20 北京思图场景数据科技服务有限公司 基于生物基础与行为特征及业务特征的金融风险评估方法
CN113420579A (zh) * 2021-06-29 2021-09-21 北大方正集团有限公司 标识码位置定位模型的训练及定位方法、装置、电子设备
CN113436231A (zh) * 2021-06-30 2021-09-24 平安科技(深圳)有限公司 行人轨迹生成方法、装置、设备及存储介质
CN113469111A (zh) * 2021-07-16 2021-10-01 中国银行股份有限公司 图像关键点检测方法及系统、电子设备、存储介质
CN113570689A (zh) * 2021-07-28 2021-10-29 杭州网易云音乐科技有限公司 人像卡通化方法、装置、介质和计算设备
CN113657220A (zh) * 2021-08-03 2021-11-16 南方电网数字电网研究院有限公司 电能质量扰动信号分类模型的训练方法及装置
CN113705685A (zh) * 2021-08-30 2021-11-26 平安科技(深圳)有限公司 疾病特征识别模型训练、疾病特征识别方法、装置及设备
CN113723448A (zh) * 2021-07-16 2021-11-30 北京工业大学 图像中对象分类和计数方法、装置、电子设备及介质
CN113724163A (zh) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 基于神经网络的图像矫正方法、装置、设备及介质
CN113781729A (zh) * 2021-07-23 2021-12-10 广东电网有限责任公司广州供电局 基于智能网关的输电线路外力破坏监测方法和装置
CN113850207A (zh) * 2021-09-29 2021-12-28 中国平安财产保险股份有限公司 基于人工智能的微表情分类方法、装置、电子设备及介质
CN113902018A (zh) * 2021-10-12 2022-01-07 深圳壹账通智能科技有限公司 图像样本生成方法、装置、计算机可读介质及电子设备
CN114005019A (zh) * 2021-10-29 2022-02-01 北京有竹居网络技术有限公司 一种翻拍图像识别方法及其相关设备
CN114005015A (zh) * 2021-12-28 2022-02-01 北京的卢深视科技有限公司 模型训练方法、电子设备和计算机可读存储介质
CN114255354A (zh) * 2021-12-31 2022-03-29 智慧眼科技股份有限公司 人脸识别模型训练方法、人脸识别方法、装置及相关设备
CN114359949A (zh) * 2021-12-23 2022-04-15 华南理工大学 针对电网接线图文字的识别方法
CN114387553A (zh) * 2022-01-18 2022-04-22 桂林电子科技大学 一种基于帧结构感知聚合的视频人脸识别方法
CN114445899A (zh) * 2022-01-30 2022-05-06 中国农业银行股份有限公司 一种表情识别方法、装置、设备及存储介质
CN114463559A (zh) * 2022-01-29 2022-05-10 新疆爱华盈通信息技术有限公司 图像识别模型的训练方法、装置、网络和图像识别方法
CN114789870A (zh) * 2022-05-20 2022-07-26 深圳市信成医疗科技有限公司 一种创新型模块化药储仓管理实现模式
CN115546848A (zh) * 2022-10-26 2022-12-30 南京航空航天大学 一种对抗生成网络训练方法、跨设备掌纹识别方法及系统
CN116112932A (zh) * 2023-02-20 2023-05-12 南京航空航天大学 一种数据知识双驱动的射频指纹识别方法与系统
CN116214524A (zh) * 2023-05-08 2023-06-06 国网浙江省电力有限公司宁波供电公司 用于油样回收的无人机抓载方法、装置及存储介质
CN116630832A (zh) * 2023-07-21 2023-08-22 江西现代职业技术学院 一种无人机目标识别方法、系统、计算机及可读存储介质
CN113724163B (zh) * 2021-08-31 2024-06-07 平安科技(深圳)有限公司 基于神经网络的图像矫正方法、装置、设备及介质

Families Citing this family (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111104897A (zh) * 2019-12-18 2020-05-05 深圳市捷顺科技实业股份有限公司 儿童人脸识别模型的训练方法、装置以及存储介质
CN111241998B (zh) * 2020-01-09 2023-04-28 中移(杭州)信息技术有限公司 人脸识别的方法、装置、电子设备和存储介质
CN111242217A (zh) * 2020-01-13 2020-06-05 支付宝实验室(新加坡)有限公司 图像识别模型的训练方法、装置、电子设备及存储介质
CN111524118B (zh) * 2020-04-22 2021-06-29 广东电网有限责任公司东莞供电局 变压器的运行状态检测方法、装置、计算机设备和存储介质
CN113642353A (zh) * 2020-04-27 2021-11-12 Tcl科技集团股份有限公司 一种人脸检测模型的训练方法、存储介质及终端设备
CN111581620A (zh) * 2020-04-30 2020-08-25 新浪网技术(中国)有限公司 一种用户识别方法及装置
CN111582381B (zh) * 2020-05-09 2024-03-26 北京市商汤科技开发有限公司 确定性能参数的方法及装置、电子设备和存储介质
CN111680597B (zh) * 2020-05-29 2023-09-01 北京百度网讯科技有限公司 人脸识别模型处理方法、装置、设备和存储介质
CN111783601B (zh) * 2020-06-24 2024-04-26 北京百度网讯科技有限公司 人脸识别模型的训练方法、装置、电子设备及存储介质
CN111985434B (zh) * 2020-08-28 2023-07-28 厦门市易联众易惠科技有限公司 模型增强的人脸识别方法、装置、设备及存储介质
CN112016702B (zh) * 2020-09-09 2023-07-28 平安科技(深圳)有限公司 基于迁移学习的医疗数据处理方法、装置、设备及介质
CN112561480A (zh) * 2020-12-16 2021-03-26 中国平安人寿保险股份有限公司 智能推动工作流方法、设备及计算机存储介质
CN112668637B (zh) * 2020-12-25 2023-05-23 苏州科达科技股份有限公司 网络模型的训练方法、识别方法、装置及电子设备
CN112967243B (zh) * 2021-02-26 2023-01-13 清华大学深圳国际研究生院 一种基于yolo的深度学习芯片封装裂纹缺陷检测方法
CN113011387B (zh) * 2021-04-20 2024-05-24 上海商汤科技开发有限公司 网络训练及人脸活体检测方法、装置、设备及存储介质
CN113627265A (zh) * 2021-07-13 2021-11-09 深圳市创客火科技有限公司 无人机控制方法、装置及计算机可读存储介质
CN113627361B (zh) * 2021-08-13 2023-08-08 北京百度网讯科技有限公司 人脸识别模型的训练方法、装置及计算机程序产品

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107247949A (zh) * 2017-08-02 2017-10-13 北京智慧眼科技股份有限公司 基于深度学习的人脸识别方法、装置和电子设备
US20180089803A1 (en) * 2016-03-21 2018-03-29 Boe Technology Group Co., Ltd. Resolving Method and System Based on Deep Learning
CN108133238A (zh) * 2017-12-29 2018-06-08 国信优易数据有限公司 一种人脸识别模型训练方法和装置及人脸识别方法和装置
CN109086868A (zh) * 2018-07-09 2018-12-25 南京邮电大学 一种抽象图像情感识别方法
CN109117897A (zh) * 2018-08-09 2019-01-01 百度在线网络技术(北京)有限公司 基于卷积神经网络的图像处理方法、装置及可读存储介质
CN109190476A (zh) * 2018-08-02 2019-01-11 福建工程学院 一种蔬菜识别的方法及装置

Family Cites Families (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105989375A (zh) * 2015-01-30 2016-10-05 富士通株式会社 对手写字符图像进行分类的分类器、分类装置和分类方法
CN105426963B (zh) * 2015-12-01 2017-12-26 北京天诚盛业科技有限公司 用于人脸识别的卷积神经网络的训练方法、装置及应用
US10032067B2 (en) * 2016-05-28 2018-07-24 Samsung Electronics Co., Ltd. System and method for a unified architecture multi-task deep learning machine for object recognition
CN106600538A (zh) * 2016-12-15 2017-04-26 武汉工程大学 一种基于区域深度卷积神经网络的人脸超分辨率算法
US11354577B2 (en) * 2017-03-15 2022-06-07 Samsung Electronics Co., Ltd System and method for designing efficient super resolution deep convolutional neural networks by cascade network training, cascade network trimming, and dilated convolutions
CN108304788B (zh) * 2018-01-18 2022-06-14 陕西炬云信息科技有限公司 基于深度神经网络的人脸识别方法
CN108510485B (zh) * 2018-03-27 2022-04-05 福州大学 一种基于卷积神经网络的无参照图像质量评估方法
CN108805137A (zh) * 2018-04-17 2018-11-13 平安科技(深圳)有限公司 牲畜特征向量的提取方法、装置、计算机设备和存储介质
CN109284669A (zh) * 2018-08-01 2019-01-29 辽宁工业大学 基于Mask RCNN的行人检测方法
CN109934115B (zh) * 2019-02-18 2021-11-02 苏州市科远软件技术开发有限公司 人脸识别模型的构建方法、人脸识别方法及电子设备

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180089803A1 (en) * 2016-03-21 2018-03-29 Boe Technology Group Co., Ltd. Resolving Method and System Based on Deep Learning
CN107247949A (zh) * 2017-08-02 2017-10-13 北京智慧眼科技股份有限公司 基于深度学习的人脸识别方法、装置和电子设备
CN108133238A (zh) * 2017-12-29 2018-06-08 国信优易数据有限公司 一种人脸识别模型训练方法和装置及人脸识别方法和装置
CN109086868A (zh) * 2018-07-09 2018-12-25 南京邮电大学 一种抽象图像情感识别方法
CN109190476A (zh) * 2018-08-02 2019-01-11 福建工程学院 一种蔬菜识别的方法及装置
CN109117897A (zh) * 2018-08-09 2019-01-01 百度在线网络技术(北京)有限公司 基于卷积神经网络的图像处理方法、装置及可读存储介质

Cited By (46)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113168573A (zh) * 2021-03-02 2021-07-23 深圳市锐明技术股份有限公司 一种模型的训练方法、装置、终端设备及存储介质
CN113168573B (zh) * 2021-03-02 2024-04-16 深圳市锐明技术股份有限公司 一种模型的训练方法、装置、终端设备及存储介质
CN113158850A (zh) * 2021-04-07 2021-07-23 大连海事大学 基于深度学习的船舶驾驶员疲劳检测方法和系统
CN113158850B (zh) * 2021-04-07 2024-01-05 大连海事大学 基于深度学习的船舶驾驶员疲劳检测方法和系统
CN113065512A (zh) * 2021-04-21 2021-07-02 深圳壹账通智能科技有限公司 人脸微表情识别方法、装置、设备及存储介质
CN113283978B (zh) * 2021-05-06 2024-05-10 北京思图场景数据科技服务有限公司 基于生物基础与行为特征及业务特征的金融风险评估方法
CN113283978A (zh) * 2021-05-06 2021-08-20 北京思图场景数据科技服务有限公司 基于生物基础与行为特征及业务特征的金融风险评估方法
CN113156439A (zh) * 2021-05-08 2021-07-23 中国石油大学(华东) 一种基于数据驱动的sar风场海浪联合反演方法及系统
CN113221830A (zh) * 2021-05-31 2021-08-06 平安科技(深圳)有限公司 一种超分活体识别方法、系统、终端及存储介质
CN113221830B (zh) * 2021-05-31 2023-09-01 平安科技(深圳)有限公司 一种超分活体识别方法、系统、终端及存储介质
CN113420579A (zh) * 2021-06-29 2021-09-21 北大方正集团有限公司 标识码位置定位模型的训练及定位方法、装置、电子设备
CN113436231B (zh) * 2021-06-30 2023-09-15 平安科技(深圳)有限公司 行人轨迹生成方法、装置、设备及存储介质
CN113436231A (zh) * 2021-06-30 2021-09-24 平安科技(深圳)有限公司 行人轨迹生成方法、装置、设备及存储介质
CN113469111A (zh) * 2021-07-16 2021-10-01 中国银行股份有限公司 图像关键点检测方法及系统、电子设备、存储介质
CN113723448A (zh) * 2021-07-16 2021-11-30 北京工业大学 图像中对象分类和计数方法、装置、电子设备及介质
CN113781729A (zh) * 2021-07-23 2021-12-10 广东电网有限责任公司广州供电局 基于智能网关的输电线路外力破坏监测方法和装置
CN113570689A (zh) * 2021-07-28 2021-10-29 杭州网易云音乐科技有限公司 人像卡通化方法、装置、介质和计算设备
CN113570689B (zh) * 2021-07-28 2024-03-01 杭州网易云音乐科技有限公司 人像卡通化方法、装置、介质和计算设备
CN113657220A (zh) * 2021-08-03 2021-11-16 南方电网数字电网研究院有限公司 电能质量扰动信号分类模型的训练方法及装置
CN113705685A (zh) * 2021-08-30 2021-11-26 平安科技(深圳)有限公司 疾病特征识别模型训练、疾病特征识别方法、装置及设备
CN113705685B (zh) * 2021-08-30 2023-08-01 平安科技(深圳)有限公司 疾病特征识别模型训练、疾病特征识别方法、装置及设备
CN113724163B (zh) * 2021-08-31 2024-06-07 平安科技(深圳)有限公司 基于神经网络的图像矫正方法、装置、设备及介质
CN113724163A (zh) * 2021-08-31 2021-11-30 平安科技(深圳)有限公司 基于神经网络的图像矫正方法、装置、设备及介质
CN113850207B (zh) * 2021-09-29 2024-05-03 中国平安财产保险股份有限公司 基于人工智能的微表情分类方法、装置、电子设备及介质
CN113850207A (zh) * 2021-09-29 2021-12-28 中国平安财产保险股份有限公司 基于人工智能的微表情分类方法、装置、电子设备及介质
CN113902018A (zh) * 2021-10-12 2022-01-07 深圳壹账通智能科技有限公司 图像样本生成方法、装置、计算机可读介质及电子设备
CN114005019B (zh) * 2021-10-29 2023-09-22 北京有竹居网络技术有限公司 一种翻拍图像识别方法及其相关设备
CN114005019A (zh) * 2021-10-29 2022-02-01 北京有竹居网络技术有限公司 一种翻拍图像识别方法及其相关设备
CN114359949B (zh) * 2021-12-23 2023-04-25 华南理工大学 针对电网接线图文字的识别方法
CN114359949A (zh) * 2021-12-23 2022-04-15 华南理工大学 针对电网接线图文字的识别方法
CN114005015A (zh) * 2021-12-28 2022-02-01 北京的卢深视科技有限公司 模型训练方法、电子设备和计算机可读存储介质
CN114255354A (zh) * 2021-12-31 2022-03-29 智慧眼科技股份有限公司 人脸识别模型训练方法、人脸识别方法、装置及相关设备
CN114387553A (zh) * 2022-01-18 2022-04-22 桂林电子科技大学 一种基于帧结构感知聚合的视频人脸识别方法
CN114387553B (zh) * 2022-01-18 2024-03-22 桂林电子科技大学 一种基于帧结构感知聚合的视频人脸识别方法
CN114463559B (zh) * 2022-01-29 2024-05-10 芯算一体(深圳)科技有限公司 图像识别模型的训练方法、装置、网络和图像识别方法
CN114463559A (zh) * 2022-01-29 2022-05-10 新疆爱华盈通信息技术有限公司 图像识别模型的训练方法、装置、网络和图像识别方法
CN114445899A (zh) * 2022-01-30 2022-05-06 中国农业银行股份有限公司 一种表情识别方法、装置、设备及存储介质
CN114789870A (zh) * 2022-05-20 2022-07-26 深圳市信成医疗科技有限公司 一种创新型模块化药储仓管理实现模式
CN115546848A (zh) * 2022-10-26 2022-12-30 南京航空航天大学 一种对抗生成网络训练方法、跨设备掌纹识别方法及系统
CN115546848B (zh) * 2022-10-26 2024-02-02 南京航空航天大学 一种对抗生成网络训练方法、跨设备掌纹识别方法及系统
CN116112932B (zh) * 2023-02-20 2023-11-10 南京航空航天大学 一种数据知识双驱动的射频指纹识别方法与系统
CN116112932A (zh) * 2023-02-20 2023-05-12 南京航空航天大学 一种数据知识双驱动的射频指纹识别方法与系统
CN116214524B (zh) * 2023-05-08 2023-10-03 国网浙江省电力有限公司宁波供电公司 用于油样回收的无人机抓载方法、装置及存储介质
CN116214524A (zh) * 2023-05-08 2023-06-06 国网浙江省电力有限公司宁波供电公司 用于油样回收的无人机抓载方法、装置及存储介质
CN116630832B (zh) * 2023-07-21 2023-09-29 江西现代职业技术学院 一种无人机目标识别方法、系统、计算机及可读存储介质
CN116630832A (zh) * 2023-07-21 2023-08-22 江西现代职业技术学院 一种无人机目标识别方法、系统、计算机及可读存储介质

Also Published As

Publication number Publication date
CN110543815A (zh) 2019-12-06
CN110543815B (zh) 2024-03-08

Similar Documents

Publication Publication Date Title
WO2021012526A1 (fr) Procédé d'apprentissage de modèle de reconnaissance faciale, procédé et appareil de reconnaissance faciale, dispositif, et support de stockage
US11908239B2 (en) Image recognition network model training method, image recognition method and apparatus
CN109101602B (zh) 图像检索模型训练方法、图像检索方法、设备及存储介质
WO2019233421A1 (fr) Procédé et dispositif de traitement d'image, appareil informatique, et support de stockage
CN109284729B (zh) 基于视频获取人脸识别模型训练数据的方法、装置和介质
WO2020119350A1 (fr) Procédé et appareil de classification de vidéos, dispositif informatique et support d'informations
WO2019119505A1 (fr) Procédé et dispositif de reconnaissance faciale, dispositif informatique et support d'enregistrement
KR102385463B1 (ko) 얼굴 특징 추출 모델 학습 방법, 얼굴 특징 추출 방법, 장치, 디바이스 및 저장 매체
WO2019109526A1 (fr) Procédé et dispositif de reconnaissance de l'âge de l'image d'un visage, et support de stockage
CN109284733B (zh) 一种基于yolo和多任务卷积神经网络的导购消极行为监控方法
WO2016177259A1 (fr) Procédé et dispositif de reconnaissance d'images similaires
CN111814620B (zh) 人脸图像质量评价模型建立方法、优选方法、介质及装置
EP2806374B1 (fr) Procédé et système de sélection automatique d'un ou de plusieurs algorithmes de traitement d'image
US8971591B2 (en) 3D image estimation for 2D image recognition
US8750573B2 (en) Hand gesture detection
US8879803B2 (en) Method, apparatus, and computer program product for image clustering
WO2017161756A1 (fr) Procédé et système d'identification de vidéo
WO2018054329A1 (fr) Procédé et dispositif de détection d'objets, appareil électronique, programme informatique et support de stockage
US20120027252A1 (en) Hand gesture detection
WO2021237570A1 (fr) Procédé et appareil d'audit d'image, dispositif, et support de stockage
WO2020228181A1 (fr) Procédé et appareil de cadrage d'image de paume, dispositif informatique et support de stockage
WO2019056503A1 (fr) Procédé d'évaluation de surveillance de magasin, dispositif, et support d'informations
WO2020155790A1 (fr) Procédé et appareil d'extraction d'informations de règlement de sinistre, et dispositif électronique
US11537636B2 (en) System and method for using multimedia content as search queries
CN112036316B (zh) 手指静脉识别方法、装置、电子设备及可读存储介质

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19938770

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19938770

Country of ref document: EP

Kind code of ref document: A1