WO2021007857A1 - Identity authentication method, terminal device, and storage medium - Google Patents

Identity authentication method, terminal device, and storage medium Download PDF

Info

Publication number
WO2021007857A1
WO2021007857A1 PCT/CN2019/096579 CN2019096579W WO2021007857A1 WO 2021007857 A1 WO2021007857 A1 WO 2021007857A1 CN 2019096579 W CN2019096579 W CN 2019096579W WO 2021007857 A1 WO2021007857 A1 WO 2021007857A1
Authority
WO
WIPO (PCT)
Prior art keywords
text information
recognition
detected
image
lip
Prior art date
Application number
PCT/CN2019/096579
Other languages
French (fr)
Chinese (zh)
Inventor
艾静雅
柳彤
朱大卫
汤慧秀
Original Assignee
深圳海付移通科技有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 深圳海付移通科技有限公司 filed Critical 深圳海付移通科技有限公司
Priority to CN201980010284.3A priority Critical patent/CN111684459A/en
Priority to PCT/CN2019/096579 priority patent/WO2021007857A1/en
Publication of WO2021007857A1 publication Critical patent/WO2021007857A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/161Detection; Localisation; Normalisation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/22Matching criteria, e.g. proximity measures
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F21/00Security arrangements for protecting computers, components thereof, programs or data against unauthorised activity
    • G06F21/30Authentication, i.e. establishing the identity or authorisation of security principals
    • G06F21/31User authentication
    • G06F21/32User authentication using biometric data, e.g. fingerprints, iris scans or voiceprints
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V40/00Recognition of biometric, human-related or animal-related patterns in image or video data
    • G06V40/10Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
    • G06V40/16Human faces, e.g. facial parts, sketches or expressions
    • G06V40/168Feature extraction; Face representation
    • G06V40/171Local features and components; Facial parts ; Occluding parts, e.g. glasses; Geometrical relationships

Definitions

  • This application relates to the technical field of identity verification, in particular to an identity verification method, terminal device, and storage medium.
  • terminal devices With the development of society, people rely more and more on the use of terminal devices. For security and privacy considerations, terminal devices need to be authenticated to identify whether the current user has permission to use them, such as most smart phones. Some private content of the unlocked screen and terminal equipment will be encrypted.
  • fingerprint verification cannot be processed without contact with quick verification
  • character verification has shortcomings such as easy to forget and easy to copy
  • fingerprint verification can only ensure that the characteristics of a person are effectively verified. It is not guaranteed to be a real person, it may be a fingerprint film.
  • the main problem solved by this application is to provide an identity verification method, terminal device, and storage medium, which realize the characteristics of the verification method that is difficult to copy, resist forgetting, and contactless, improve the accuracy of identity verification, and make the use of terminal devices more secure .
  • the technical solution adopted in this application is to provide an identity verification method.
  • the method includes: when the terminal device obtains a setting operation instruction, collects a first image to be detected, and performs processing on the first image to be detected. Face recognition; after the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected image; after the lip shape recognition is passed, the setting operation instruction is responded to.
  • the face recognition is passed, collecting a plurality of consecutive second to-be-detected images, and performing lip shape recognition on the second to-be-detected image includes: after the face recognition is passed, collecting a plurality of consecutive second to-be-detected images Image; perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is text information in the preset whitelist; if so, determine that the lip shape recognition passes.
  • the method further includes: acquiring text information entered by the user; adding the entered text information to the whitelist.
  • after the face recognition is passed, collecting a plurality of consecutive second to-be-detected images, and performing lip-shape recognition on the second to-be-detected image includes: after the face recognition is passed, collecting multiple second to-be-reading images Detect images; perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is text information in the preset blacklist; if so, determine that the lip shape recognition fails.
  • the method further includes: adding at least one piece of text information in the white list to the black list, and deleting the text information from the white list.
  • a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected image, including: after the face recognition is passed, the standard text information is displayed, and the continuous A plurality of second to-be-detected images; perform lip shape recognition on the plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is the same as the standard text information; if so, determine that the lip shape recognition passes.
  • displaying standard text information includes: randomly selecting one text information from a plurality of text information in the database as the standard text information, and displaying the standard text information.
  • performing lip shape recognition on multiple second to-be-detected images to obtain recognized text information includes: extracting face information of multiple second to-be-detected images; extracting multiple continuously changing lip shapes from multiple face information Features: Based on multiple continuously changing lip features, the recognized text information is obtained.
  • based on a plurality of continuously changing lip shape features to obtain recognized text information including: input a plurality of continuously changing lip shape features into the lip shape recognition model, so that the lip shape recognition model can generate corresponding pronunciation information, and based on the pronunciation Information, calculate the corresponding recognized text information.
  • collecting the first image to be detected and performing face recognition on the first image to be detected includes: when the mobile terminal obtains the setting operation instruction, collecting the first to be detected Image; extract the face image in the first image to be detected; perform face recognition on the face image.
  • the face recognition of the face image includes: extracting face feature information from the face image; comparing the face feature information with the pre-stored standard face feature information for similarity; and comparing the result of the similarity When the preset requirements are met, the face recognition is determined to pass.
  • the setting operation instruction is a payment operation instruction; after the lip shape recognition is passed, responding to the setting operation instruction includes: after the lip shape recognition is passed, responding to the payment operation instruction to complete the corresponding payment.
  • a terminal device which includes a processor, a camera module connected to the processor, and a memory; the memory is used to store program data, and the processor is used to execute Program data to implement the method described above.
  • another technical solution adopted in this application is to provide a computer storage medium for storing program data, and the program data is used to implement the above-mentioned method when the program data is executed by a processor.
  • the terminal device includes: a first identification module, when the setting operation instruction is obtained, the first image to be detected is collected, and Perform face recognition on the first image to be detected; the second recognition module is used to collect a plurality of consecutive second images to be detected after the face recognition is passed, and perform lip shape recognition on the second image to be detected; response module , Used to respond to setting operation instructions after lip shape recognition is passed.
  • an identity verification method of the present application combines face recognition and lip shape recognition to achieve verification methods that are difficult to copy, anti-forgetting, non-contact, etc.
  • FIG. 1 is a schematic flowchart of the first embodiment of the identity verification method provided by the present application
  • FIG. 2 is a schematic flowchart of a second embodiment of the identity verification method provided by the present application.
  • FIG. 3 is a schematic flowchart of a third embodiment of the identity verification method provided by the present application.
  • FIG. 4 is a schematic flowchart of a fourth embodiment of the identity verification method provided by the present application.
  • FIG. 5 is a schematic flowchart of a fifth embodiment of the identity verification method provided by the present application.
  • FIG. 6 is a schematic flowchart of a sixth embodiment of the identity verification method provided by the present application.
  • FIG. 7 is a schematic structural diagram of a first embodiment of a terminal device provided by the present application.
  • FIG. 8 is a schematic structural diagram of a second embodiment of a terminal device provided by the present application.
  • FIG. 9 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application.
  • Fig. 1 is a schematic flowchart of a first embodiment of an identity verification method provided by the present application. The method includes:
  • Step 11 When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
  • the terminal device When the terminal device obtains the setting operation instruction, it calls the camera module, collects the first image to be detected, and performs face recognition on the first image to be detected.
  • the terminal device may be a mobile terminal, such as a smart phone, a tablet computer, a wearable device, etc.
  • the setting operation instruction may be an unlock screen instruction.
  • the device terminal obtains the unlock screen instruction, it turns on the camera to collect the current
  • the image information within the shooting range of the camera is used for face recognition on the collected image information. It can be understood that when there is no face information in the collected image information, the terminal device stops the recognition, or prompts to re-collect and recognize again.
  • Face recognition can be divided into face image acquisition and detection, face image preprocessing, face image feature extraction, and matching and recognition.
  • Face image collection Different face images can be collected through the camera lens, such as static images, dynamic images, different positions, different expressions, etc. can be well collected. When the user is within the shooting range of the capture device, the capture device will automatically search for and shoot the user's face image.
  • Face detection In practice, face detection is mainly used for preprocessing of face recognition, that is, to accurately calibrate the position and size of the face in the image.
  • the pattern features contained in face images are very rich, such as histogram features, color features, template features, structural features, and Haar features. Face detection is to pick out the useful information, and use these features to realize face detection.
  • the Adaboost algorithm is used to select some rectangular features (weak classifiers) that best represent the face, and the weak classifier is constructed into a strong classifier according to the weighted voting method, and then several strong classifiers obtained by training A cascade structure of stacked classifiers is formed in series, which effectively improves the detection speed of the classifier.
  • Face image preprocessing The image preprocessing of the face is based on the face detection result, the image is processed and finally serves the process of feature extraction. Due to various conditions and random interference, the original image acquired by the system cannot be used directly. It must be preprocessed by grayscale correction and noise filtering in the early stage of image processing.
  • the preprocessing process mainly includes light compensation, gray scale transformation, histogram equalization, normalization, geometric correction, filtering and sharpening of the face image.
  • Face image feature extraction The features that can be used in face recognition are usually divided into visual features, pixel statistical features, face image transformation coefficient features, and face image algebraic features. Facial feature extraction is based on certain features of the human face. Face feature extraction, also known as face representation, is a process of feature modeling of human faces. Facial feature extraction methods can be summarized into two categories: one is knowledge-based representation methods; the other is based on algebraic features or statistical learning.
  • the knowledge-based representation method is mainly based on the shape description of the face organs and the distance characteristics between them to obtain feature data that is helpful for face classification. Its feature components usually include the Euclidean distance, curvature, and angle between feature points. .
  • the human face is composed of parts such as eyes, nose, mouth, and chin. The geometric description of these parts and the structural relationship between them can be used as important features to recognize the face. These features are called geometric features.
  • Knowledge-based face representation mainly includes geometric feature-based methods and template matching methods.
  • Face image matching and recognition The feature data of the extracted face image is searched and matched with the feature template stored in the database. By setting a threshold, when the similarity exceeds this threshold, the matching result is output. Face recognition is to compare the facial features to be recognized with the obtained facial feature template, and judge the identity information of the face based on the degree of similarity. This process is divided into two categories: one is confirmation, which is a process of one-to-one image comparison, and the other is identification, which is a process of one-to-many image matching and comparison.
  • Step 12 After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected images.
  • the manner of collecting consecutive multiple second to-be-detected images may be that the user speaks a paragraph of text, and the camera module of the terminal device collects consecutive images.
  • the lip shape recognition model uses complex end-to-end deep neural network technology to model the lip sequence and establish a vocabulary.
  • the terminal device can broadcast a paragraph of text by voice, the user repeats this paragraph of text, collects continuous images of the user repeats this paragraph of text, extracts the lip features of the continuous images, and recognizes the corresponding pronunciation through the lip recognition model.
  • the corresponding pronunciation is matched with password characters to obtain text information, and this text information is matched with the text information voiced by the terminal device. If the matching is passed, then step 13 is executed.
  • a section of text can be displayed on the display screen of the terminal device, and the user can read this section of text, collect continuous images of the text read by the user, extract the lip features of the continuous images, and recognize them through the lip recognition model
  • the corresponding pronunciation is generated, the corresponding pronunciation is matched with the password characters, and the text information is obtained, and the text information is matched with the text information voiced by the terminal device. If the matching is passed, then step 13 is executed.
  • the text displayed by the terminal device may be preset text information or random text information.
  • Step 13 After the lip shape recognition is passed, respond to the setting operation instruction.
  • the terminal device After the lip recognition is passed, if the operation instruction is set to unlock the screen, the terminal device will unlock the screen and display the screen content; if the operation instruction is set to unlock the terminal device’s private album, the terminal device will unlock the private album and display the photos in the private album ; If the set operation instruction is a payment instruction, the terminal device completes the corresponding payment; if the set operation instruction is to view private information, the terminal device displays the private information.
  • the terminal device when a user needs to use a terminal device to pay a bill, the terminal device obtains a payment instruction and prompts the user to perform face recognition. After face recognition, the user is prompted to say a paragraph, and the user is synchronized to collect continuous For image information, lip feature extraction is performed on the image information, and lip shape recognition is performed. After the lip shape recognition is passed, the terminal device completes the corresponding payment.
  • the user clicks on an application on the terminal device.
  • the application requires identity verification.
  • the terminal device obtains the operation instruction and prompts the user to perform face recognition. After the face recognition is passed, it is displayed on the display
  • the text information prompts the user to read the text information.
  • the continuous image information is synchronously collected, and the image information is subjected to lip shape extraction and lip shape recognition. After the lip shape recognition is passed, the application is unlocked.
  • the terminal device after obtaining the setting operation instruction, performs face recognition on the user. After the face recognition is passed, it obtains the video information of the text information spoken by the user, and the terminal device splits the video information into audio streams. And the image stream, perform voice recognition on the audio stream, recognize the text information, perform continuous lip feature extraction on the image stream, recognize through the lip recognition model, and calculate the text information contained in the lip feature in the image stream. Audio The text information recognized by the stream is compared with the text information recognized by the image stream. If the same, the recognized text information is matched with the preset text information. If the matching is successful, the identity verification is considered successful, and the terminal device responds to the setting Operating instructions.
  • an application operation method and an identity verification method of the present application include: when a terminal device obtains a setting operation instruction, collecting a first image to be detected, Perform face recognition on an image to be detected; after the face recognition is passed, collect consecutive multiple second to be detected images, and perform lip shape recognition on the second to be detected image; after the lip shape recognition is passed, respond to the setting operation instruction.
  • Figure 2 is a schematic flowchart of a second embodiment of the identity verification method provided by the present application, and the method includes:
  • Step 21 When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
  • Step 22 After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected.
  • the terminal device prompts the user to collect the second to-be-detected image, such as prompting the user to face the camera and speak text information.
  • the identity of the current user is recognized to a certain extent, so when the user is prompted to collect the second to-be-detected image, the user can be guided by relevant information to prompt the user to say the preset whitelist Text messages in.
  • Step 23 Perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information.
  • Lip shape recognition technology is a technology that interprets the content of speech based on the movement of the lips when speaking.
  • it is necessary to collect multiple images containing the lips movement of the speaker or collect a video containing the lips movement of the speaker, and then Combining image processing technology and deep learning technology to identify multi-frame continuous image sequences, by identifying the lip shape in the multi-frame continuous image sequence, mapping the lip shape to the pronunciation, and then determining the corresponding natural language words and sentences based on the pronunciation in a continuous period of time , That is, the content of the speech.
  • the lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain corresponding pronunciation information, and based on the pronunciation information, the corresponding recognized text information is calculated.
  • the lip recognition model may be an end-to-end algorithm model based on the encoder-decoder architecture fusion spatiotemporal convolutional neural network feature extractor and word embedding network, and using the attention mechanism.
  • the feature extractor uses a spatio-temporal convolutional neural network (STCNN)
  • the encoder-decoder subunit uses a long short-term memory network (LSTM)
  • the word embedding (Embedding) encoding method uses Word2vec.
  • the lip shape recognition model can use the lip shape recognition data set of Mandarin Chinese to train the model, use an improved multi-stage convolutional neural network (MTCNN) to extract the lip region in the silent video, and then send the extracted lip region In the spatio-temporal convolutional network STCNN, it is used to extract the visual feature information of the lip action.
  • MTCNN multi-stage convolutional neural network
  • STCNN spatio-temporal convolutional network
  • the encoder-decoder based on LSTM is used to encode lip visual feature information and decode it into relevant text information during model inference.
  • the attention mechanism can make the model decoder pay attention to the coded content of the encoder at a specific location, instead of using the entire coded content as a basis for decoding, thereby improving the decoding effect of the model.
  • THULAC TSU Lexical Analyzer for Chinese, Chinese lexical analysis toolkit
  • the role of this part in the network is essentially to act as a character encoding.
  • the encoder-decoder architecture encodes a variable-length sequence into a fixed-length representation, and represents a given fixed-length vector as a variable-length sequence. From a probabilistic point of view, the model uses a general method to learn the conditional probability distribution of another variable-length sequence under the condition of a variable-length sequence.
  • the lip shape recognition model can use the above scheme or other related schemes to establish different databases according to different languages, so as to be applied to different language regions.
  • Step 24 Determine whether the recognized text information is the text information in the preset whitelist.
  • the user enters some text information in the terminal device, and adds the entered text information to the whitelist.
  • step 25 is executed.
  • the text in the whitelist can have only one paragraph or multiple paragraphs.
  • the text in the whitelist can have only one paragraph or multiple paragraphs.
  • Step 25 Confirm that the lip shape recognition passes.
  • Step 26 After the lip shape recognition is passed, respond to the setting operation instruction.
  • the terminal device contains private short messages, which need to be authenticated before they can be viewed.
  • the terminal device responds to this operation instruction to perform face recognition on the user.
  • face recognition the terminal device uses the camera to collect multiple consecutive images of the text information read by the user for detection.
  • the corresponding text information is calculated through the lip shape recognition model, and the corresponding text information is matched with the preset text information in the whitelist. If the matching is successful, the lip shape recognition is determined to pass, and the terminal device responds to the setting Operation instructions, display private short messages for users to view.
  • Fig. 3 is a schematic flowchart of a third embodiment of an identity verification method provided by the present application, and the method includes:
  • Step 31 When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
  • Step 32 After the face recognition is passed, a plurality of second to-be-detected images that are read consecutively are collected.
  • Steps 31-32 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
  • Step 33 Perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information.
  • the lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain corresponding pronunciation information, and based on the pronunciation information, the corresponding recognized text information is calculated.
  • Step 34 Determine whether the recognized text information is the text information in the preset blacklist.
  • the user enters some text information in the terminal device, and adds the entered text information to the blacklist.
  • step 35 is executed.
  • the text in the blacklist can have only one paragraph or multiple paragraphs.
  • the text in the blacklist can have only one paragraph or multiple paragraphs.
  • At least one piece of text information in the white list is added to the black list, and the text information is deleted from the white list.
  • Step 35 Determine that the lip shape recognition fails.
  • the recognized text information is different from the text information in the blacklist, the recognized text information is matched with the text information in the white list. If they are the same, the lip recognition passes and the terminal device responds to the settings Operating instructions.
  • this part of the text information is deleted from the whitelist and added to the blacklist, so that this part of the text information can be used To verify whether the text information recognized by the lip shape is safe.
  • each piece of text information in the whitelist of the terminal device is time-sensitive.
  • the time limit can be two hours, twenty hours, forty-eight hours, and the specific time limit is set by the system or user requirements.
  • the terminal device will Automatically delete it and add it to the blacklist, and prompt the user or remind the user that the text message has exceeded the time limit may be a security risk, please handle it by yourself.
  • there is a limit on the number of times each piece of text information is used for identity verification the limit can be ten, twenty, fifty, one hundred, and the specific limit is set by the system or user needs).
  • the terminal device When the text information exceeds the limit , The terminal device will automatically delete it and add it to the blacklist and prompt the user or remind the user that the text message has exceeded the number of uses may be a security risk, please handle it by yourself. This ensures the iterative update of the text information in the whitelist, which is easy to ensure information security and is not easy to be stolen. Even if the text information is stolen, the text information is already in the blacklist, and the use of the text information cannot pass identity verification
  • FIG. 4 is a schematic flowchart of a fourth embodiment of an identity verification method provided by the present application, and the method includes:
  • Step 41 When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
  • Step 42 After the face recognition is passed, the standard text information is displayed, and a plurality of consecutive second to-be-detected images are collected.
  • the terminal device displays standard text on the display screen, prompting the user's face to face the camera to read the displayed standard text, and at the same time collect multiple consecutive second to-be-detected images when the user reads the standard text through the camera.
  • the standard text information may be one of multiple text information entered in advance by the user.
  • the standard text information may be randomly selected from a plurality of text information in the database as the standard text information.
  • the standard text information may be randomly selected from the cloud server as the standard text information.
  • Step 43 Perform lip recognition on multiple second to-be-detected images to obtain recognized text information.
  • step 43 is specifically extracting the face information of multiple second to-be-detected images from the continuous collection of multiple second to-be-detected images; extracting multiple continuously changing lip features from the multiple face information; A continuously changing lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain the corresponding pronunciation information, and based on the pronunciation information, the corresponding recognition text information is calculated.
  • Step 44 Determine whether the recognized text information is the same as the standard text information.
  • step 45 It is judged whether the recognized text information is the same as the standard text information, and if the same, step 45 is executed.
  • Step 45 Confirm that the lip shape recognition passes.
  • the terminal device After confirming that the lip shape recognition is passed, the terminal device responds to the setting operation instruction to complete the corresponding operation.
  • Fig. 5 is a schematic flowchart of a fifth embodiment of an identity verification method provided by the present application. The method includes:
  • Step 51 The mobile terminal collects the first image to be detected when acquiring the setting operation instruction.
  • Step 52 Extract the face image in the first image to be detected.
  • the terminal device will re-acquire the first image to be detected and prompt the user to face the camera, so that the collected first image to be detected contains the face image .
  • Step 53 Extract face feature information from the face image.
  • a local feature extraction method can be used for the method for extracting facial feature information from the facial image.
  • feature extraction based on facial organs can be used.
  • feature extraction based on templates can be used.
  • feature extraction based on elastic map matching methods can be used.
  • the method for extracting facial feature information from the facial image may adopt an overall feature extraction method.
  • feature extraction based on algebraic method feature extraction based on neural network
  • feature extraction based on wavelet multi-resolution can be used.
  • Step 54 Perform a similarity comparison between the facial feature information and the pre-stored standard facial feature information.
  • the pre-stored standard facial feature information is the facial feature information extracted from the facial image information collected in advance by the user.
  • the pre-stored standard facial feature information can be grouped as a unit, and the facial feature information in each group A face image is formed so that multiple sets of standard facial feature information can be pre-stored in the terminal device.
  • Step 55 When the result of the similarity comparison meets the preset requirements, it is determined that the face recognition passes.
  • the similarity comparison method can be to compare a single feature with a single standard facial feature, and then multiply the comparison results of multiple single features. When the multiplied result is greater than a preset value, it is determined Face recognition passed. Taking a single feature as an example of nose, eyes, and mouth, the similarity comparison value of nose is 0.95, the similarity comparison value of eyes is 0.85, and the similarity comparison value of mouth is 0.99. Multiply the three comparison values. It is 0.95*0.85*0.99 ⁇ 0.8, the preset value is 0.75, 0.8>0.75, so the similarity comparison result is greater than the preset requirement, and the face recognition is determined to pass.
  • the similarity comparison method may be to compare the overall feature with the overall standard face feature, and when the comparison result is greater than a preset value, it is determined that the face recognition passes.
  • Step 56 After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected images.
  • Step 57 After the lip shape recognition is passed, respond to the setting operation instruction.
  • FIG. 6 is a schematic flowchart of a sixth embodiment of an identity verification method provided by the present application. The method includes:
  • Step 61 When the terminal device obtains the payment operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
  • Step 62 After the face recognition is passed, collect a plurality of consecutive second to-be-detected images, and perform lip shape recognition on the second to-be-detected images.
  • Step 63 After the lip shape recognition is passed, respond to the payment operation instruction to complete the corresponding payment.
  • the terminal device when the terminal device obtains a payment operation instruction, if the payment amount is a small payment, the face recognition in step 61 can be skipped, and the current user only needs to correctly read out the tasks set in the whitelist. A piece of text information. After the lip shape recognition is passed, the terminal device responds to the payment operation instruction to complete the corresponding payment.
  • the terminal device when the terminal device obtains the payment operation instruction, the current user needs to confirm that the current user has the authority in the terminal device through face recognition, and then correctly read any piece of text information set in the whitelist. After the lip shape recognition is passed, the terminal device responds to the payment operation instruction to complete the corresponding payment.
  • the terminal device when the terminal device obtains the payment operation instruction, the current user needs to confirm that the current user has the authority in the terminal device through face recognition, and then correctly read the random text information on the terminal device. After passing, the terminal device responds to the payment operation instruction to complete the corresponding payment.
  • the user finds that the text information in the whitelist in the terminal device has hidden security risks. If it is stolen by others, the text information with hidden security risks will be deleted and added to the blacklist. In this way, even if the face recognition is passed, when the lip shape recognizes that the text information is the text information in the blacklist, the terminal device is immediately locked and all operation instructions are terminated.
  • users of terminal devices are divided into identities with different permissions, and users with the highest permissions can quickly update the white list, determine the list of users with payment permissions, and can change the list of users with payment permissions at any time.
  • FIG. 7 is a schematic structural diagram of a first embodiment of a terminal device provided by the present application.
  • the terminal device 70 includes a processor 71, a camera module 72 connected to the processor 71, and a memory 73; the memory 73 is used to store programs Data, the processor 71 is used to execute program data to implement the following methods:
  • the terminal device When the terminal device obtains the setting operation instruction, it collects the first image to be detected and performs face recognition on the first image to be detected; after the face recognition is passed, it collects a plurality of consecutive second images to be detected, and The second image to be detected performs lip shape recognition; after the lip shape recognition passes, responds to the setting operation instruction.
  • the processor 71 is used to execute the program data to implement the following method: after the face recognition is passed, a plurality of consecutive second to-be-detected images are collected; Recognize to obtain recognized text information; determine whether the recognized text information is the text information in the preset whitelist; if so, determine that the lip shape recognition passes.
  • the processor 71 used to execute the program data is also used to implement the following methods: acquiring text information entered by the user; adding the entered text information to the white list.
  • the processor 71 is configured to execute the program data to implement the following method: after the face recognition is passed, collect a plurality of second to-be-detected images that are read consecutively; Shape recognition to obtain recognized text information; determine whether the recognized text information is the text information in the preset blacklist; if so, determine that the lip shape recognition fails.
  • the processor 71 used to execute the program data is also used to implement the following method: adding at least one piece of text information in the white list to the black list, and deleting the text information from the white list. .
  • the processor 71 is used to execute the program data to implement the following method: after the face recognition is passed, display standard text information, and collect consecutive multiple second to-be-detected images; Perform lip shape recognition on the image to be detected to obtain recognized text information; determine whether the recognized text information is the same as the standard text information; if yes, determine that the lip shape recognition passes.
  • the processor 71 used to execute the program data is also used to implement the following method: randomly select one text information from a plurality of text information in the database as the standard text information, and display the standard text information.
  • the processor 71 is used to execute the program data to implement the following method: extracting face information of multiple second images to be detected; extracting multiple continuously changing lip features from multiple face information; Based on multiple continuously changing lip features, the recognized text information is obtained.
  • the processor 71 used to execute the program data is also used to implement the following method: input a plurality of continuously changing lip features into the lip recognition model, so that the lip recognition model can output corresponding pronunciation information, and Based on the pronunciation information, the corresponding recognized text information is calculated.
  • the processor 71 is configured to execute the program data to implement the following method: when the mobile terminal obtains the setting operation instruction, collect the first image to be detected; extract the face image in the first image to be detected ; Perform face recognition on face images.
  • the processor 71 is used to execute the program data to implement the following method: extracting facial feature information from a face image; comparing the facial feature information with pre-stored standard facial feature information for similarity ; When the result of the similarity comparison meets the preset requirements, the face recognition is determined to pass.
  • the processor 71 used to execute the program data is also used to implement the following method: after the lip shape recognition is passed, respond to the payment operation instruction to complete the corresponding payment.
  • FIG. 8 is a schematic structural diagram of a second embodiment of a terminal device provided by the present application.
  • the terminal device 80 includes a first identification module 81, a second identification module 82 and a response module 83.
  • the first recognition module 81 is configured to collect the first image to be detected and perform face recognition on the first image to be detected when the setting operation instruction is acquired.
  • the second recognition module 82 is configured to collect a plurality of consecutive second to-be-detected images after the face recognition is passed, and perform lip-shape recognition on the second to-be-detected images.
  • the response module 83 is used to respond to the setting operation instruction after the lip shape recognition is passed.
  • Fig. 9 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application.
  • the computer storage medium 90 is used to store program data 91.
  • the program data 91 is executed by a processor, it is used to implement the following methods:
  • the terminal device When the terminal device obtains the setting operation instruction, it collects the first image to be detected and performs face recognition on the first image to be detected; after the face recognition is passed, it collects a plurality of consecutive second images to be detected, and The second image to be detected performs lip shape recognition; after the lip shape recognition passes, responds to the setting operation instruction.
  • the disclosed method and device may be implemented in other ways.
  • the device implementation described above is merely illustrative.
  • the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be Combined or can be integrated into another system, or some features can be ignored or not implemented.
  • the units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
  • the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit.
  • the above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
  • the integrated units in the other embodiments described above are implemented in the form of software functional units and sold or used as independent products, they can be stored in a computer readable storage medium.
  • the technical solution of this application essentially or the part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , Including several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application.
  • the aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk and other media that can store program code .

Abstract

Disclosed are an identity authentication method, a terminal device, and a storage medium. The method comprises: when a terminal device acquires a set operation instruction, collecting a first image to be detected and performing a facial recognition with respect to said first image; when the facial recognition is successful, collecting multiple consecutive second images to be detected and performing a lip recognition with respect to said second images; and when the lip recognition is successful, responding to the set operation instruction. By such means, the characteristics of an authentication scheme of being difficult to reproduce, not easily forgotten, and contactless are implemented, the accuracy of identity authentication is increased, and the security in using the terminal device is strengthened.

Description

一种身份验证方法、终端设备、存储介质Identity verification method, terminal equipment and storage medium 【技术领域】【Technical Field】
本申请涉及身份验证技术领域,具体涉及一种身份验证方法、终端设备、存储介质。This application relates to the technical field of identity verification, in particular to an identity verification method, terminal device, and storage medium.
【背景技术】【Background technique】
随着社会发展,人们对终端设备的使用越来越依赖,而出于安全及隐私考虑,终端设备使用时都需要进行身份验证来识别当前用户是否有权限使用,如大部分的智能手机都需要解锁屏幕、终端设备部分隐私内容会加密。With the development of society, people rely more and more on the use of terminal devices. For security and privacy considerations, terminal devices need to be authenticated to identify whether the current user has permission to use them, such as most smart phones. Some private content of the unlocked screen and terminal equipment will be encrypted.
相关的身份验证方式如指纹验证、字符验证等,指纹验证无法做到不接触快速验证的处理,而字符验证存在易遗忘、易复制等不足,并且指纹验证方式只能保证人的特征被有效验证,不能保证是真人,有可能是指纹膜。Relevant identity verification methods such as fingerprint verification and character verification, fingerprint verification cannot be processed without contact with quick verification, and character verification has shortcomings such as easy to forget and easy to copy, and fingerprint verification can only ensure that the characteristics of a person are effectively verified. It is not guaranteed to be a real person, it may be a fingerprint film.
【发明内容】[Content of the invention]
本申请主要解决的问题是提供一种身份验证方法、终端设备、存储介质,实现了验证方式难复制、抗遗忘、不接触等特点,提高了身份验证的准确性、让终端设备的使用更加安全。The main problem solved by this application is to provide an identity verification method, terminal device, and storage medium, which realize the characteristics of the verification method that is difficult to copy, resist forgetting, and contactless, improve the accuracy of identity verification, and make the use of terminal devices more secure .
为解决上述技术问题,本申请采用的技术方案是提供一种身份验证方法,该方法包括:终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别;在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别;在唇形识别通过后,响应设定操作指令。In order to solve the above technical problems, the technical solution adopted in this application is to provide an identity verification method. The method includes: when the terminal device obtains a setting operation instruction, collects a first image to be detected, and performs processing on the first image to be detected. Face recognition; after the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected image; after the lip shape recognition is passed, the setting operation instruction is responded to.
其中,在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别,包括:在人脸识别通过后,采集连续的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识别文字信息;判断识别文字信息是否为预设白名单中的文字信息;若是,则确定唇形识别通过。Wherein, after the face recognition is passed, collecting a plurality of consecutive second to-be-detected images, and performing lip shape recognition on the second to-be-detected image includes: after the face recognition is passed, collecting a plurality of consecutive second to-be-detected images Image; perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is text information in the preset whitelist; if so, determine that the lip shape recognition passes.
其中,该方法还包括:获取用户录入的文字信息;将录入的文字信息加入白名单中。Wherein, the method further includes: acquiring text information entered by the user; adding the entered text information to the whitelist.
其中,在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别,包括:在人脸识别通过后,采集连读的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识 别文字信息;判断识别文字信息是否为预设黑名单中的文字信息;若是,则确定唇形识别不通过。Among them, after the face recognition is passed, collecting a plurality of consecutive second to-be-detected images, and performing lip-shape recognition on the second to-be-detected image includes: after the face recognition is passed, collecting multiple second to-be-reading images Detect images; perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is text information in the preset blacklist; if so, determine that the lip shape recognition fails.
其中,该方法还包括:将白名单中的至少一段文字信息加入黑名单中,并将文字信息在白名单中删除。Wherein, the method further includes: adding at least one piece of text information in the white list to the black list, and deleting the text information from the white list.
其中,在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别,包括:在人脸识别通过后,显示标准文字信息,并采集连续的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识别文字信息;判断识别文字信息与标准文字信息是否相同;若是,则确定唇形识别通过。Among them, after the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected image, including: after the face recognition is passed, the standard text information is displayed, and the continuous A plurality of second to-be-detected images; perform lip shape recognition on the plurality of second to-be-detected images to obtain recognized text information; determine whether the recognized text information is the same as the standard text information; if so, determine that the lip shape recognition passes.
其中,显示标准文字信息,包括:从数据库中的多个文字信息中随机选择一个文字信息作为标准文字信息,并显示标准文字信息。Wherein, displaying standard text information includes: randomly selecting one text information from a plurality of text information in the database as the standard text information, and displaying the standard text information.
其中,对多个第二待检测图像进行唇形识别,以得到识别文字信息,包括:提取多个第二待检测图像的人脸信息;从多个人脸信息中提取多个连续变化的唇形特征;基于多个连续变化的唇形特征,得到识别文字信息。Wherein, performing lip shape recognition on multiple second to-be-detected images to obtain recognized text information includes: extracting face information of multiple second to-be-detected images; extracting multiple continuously changing lip shapes from multiple face information Features: Based on multiple continuously changing lip features, the recognized text information is obtained.
其中,基于多个连续变化的唇形特征,得到识别文字信息,包括:将多个连续变化的唇形特征输入至唇形识别模型,以使唇形识别模型出对应的发音信息,并基于发音信息,计算出对应的识别文字信息。Among them, based on a plurality of continuously changing lip shape features to obtain recognized text information, including: input a plurality of continuously changing lip shape features into the lip shape recognition model, so that the lip shape recognition model can generate corresponding pronunciation information, and based on the pronunciation Information, calculate the corresponding recognized text information.
其中,移动终端在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别,包括:移动终端在获取到设定操作指令时,采集第一待检测图像;提取第一待检测图像中的人脸图像;对人脸图像进行人脸识别。Wherein, when the mobile terminal obtains the setting operation instruction, collecting the first image to be detected and performing face recognition on the first image to be detected includes: when the mobile terminal obtains the setting operation instruction, collecting the first to be detected Image; extract the face image in the first image to be detected; perform face recognition on the face image.
其中,对人脸图像进行人脸识别,包括:从人脸图像中提取人脸特征信息;将人脸特征信息与预存的标准人脸特征信息进行相似度比对;在相似度比对的结果满足预设要求时,确定人脸识别通过。Among them, the face recognition of the face image includes: extracting face feature information from the face image; comparing the face feature information with the pre-stored standard face feature information for similarity; and comparing the result of the similarity When the preset requirements are met, the face recognition is determined to pass.
其中,设定操作指令为支付操作指令;在唇形识别通过后,响应设定操作指令,包括:在唇形识别通过后,响应支付操作指令,以完成相应的支付。Among them, the setting operation instruction is a payment operation instruction; after the lip shape recognition is passed, responding to the setting operation instruction includes: after the lip shape recognition is passed, responding to the payment operation instruction to complete the corresponding payment.
为解决上述技术问题,本申请采用的另一技术方案是提供一种终端设备,该终端设备包括处理器以及与处理器连接的摄像头模组以及存储器;存储器用于存储程序数据,处理器用于执行程序数据,以实现如上述的方法。In order to solve the above technical problems, another technical solution adopted in this application is to provide a terminal device, which includes a processor, a camera module connected to the processor, and a memory; the memory is used to store program data, and the processor is used to execute Program data to implement the method described above.
为解决上述技术问题,本申请采用的另一技术方案是提供一种计算机存储介质,该计算机存储介质用于存储程序数据,程序数据在被处理器执行时,用于实现如上述的方法。In order to solve the above technical problem, another technical solution adopted in this application is to provide a computer storage medium for storing program data, and the program data is used to implement the above-mentioned method when the program data is executed by a processor.
为解决上述技术问题,本申请采用的另一技术方案是提供一种终端设备,该终端设备包括:第一识别模块,用于在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别;第二识别模块,用于在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别;响应模块,用于在唇形识别通过后,响应设定操作指令。In order to solve the above technical problem, another technical solution adopted in this application is to provide a terminal device, the terminal device includes: a first identification module, when the setting operation instruction is obtained, the first image to be detected is collected, and Perform face recognition on the first image to be detected; the second recognition module is used to collect a plurality of consecutive second images to be detected after the face recognition is passed, and perform lip shape recognition on the second image to be detected; response module , Used to respond to setting operation instructions after lip shape recognition is passed.
通过上述方案,本申请的有益效果是:区别于现有技术中,本申请的一种身份验证方法,通过人脸识别与唇形识别相结合,实现验证方式难复制、抗遗忘、不接触等特点,提高身份验证的准确性、让终端设备的使用更加安全。Through the above solution, the beneficial effects of the present application are: different from the prior art, an identity verification method of the present application combines face recognition and lip shape recognition to achieve verification methods that are difficult to copy, anti-forgetting, non-contact, etc. Features to improve the accuracy of identity verification and make the use of terminal equipment more secure.
【附图说明】【Explanation of drawings】
为了更清楚地说明本申请实施例中的技术方案,下面将对实施例描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的一些实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图获得其他的附图。其中:In order to more clearly describe the technical solutions in the embodiments of the present application, the following will briefly introduce the drawings needed in the description of the embodiments. Obviously, the drawings in the following description are only some embodiments of the present application. For those of ordinary skill in the art, other drawings can be obtained from these drawings without creative work. among them:
图1是本申请提供的身份验证方法第一实施例的流程示意图;FIG. 1 is a schematic flowchart of the first embodiment of the identity verification method provided by the present application;
图2是本申请提供的身份验证方法第二实施例的流程示意图;2 is a schematic flowchart of a second embodiment of the identity verification method provided by the present application;
图3是本申请提供的身份验证方法第三实施例的流程示意图;FIG. 3 is a schematic flowchart of a third embodiment of the identity verification method provided by the present application;
图4是本申请提供的身份验证方法第四实施例的流程示意图;4 is a schematic flowchart of a fourth embodiment of the identity verification method provided by the present application;
图5是本申请提供的身份验证方法第五实施例的流程示意图;FIG. 5 is a schematic flowchart of a fifth embodiment of the identity verification method provided by the present application;
图6是本申请提供的身份验证方法第六实施例的流程示意图;FIG. 6 is a schematic flowchart of a sixth embodiment of the identity verification method provided by the present application;
图7是本申请提供的终端设备第一实施例的结构示意图;FIG. 7 is a schematic structural diagram of a first embodiment of a terminal device provided by the present application;
图8是本申请提供的终端设备第二实施例的结构示意图;FIG. 8 is a schematic structural diagram of a second embodiment of a terminal device provided by the present application;
图9是本申请提供的计算机存储介质一实施例的结构示意图。FIG. 9 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application.
【具体实施方式】【Detailed ways】
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述。可以理解的是,此处所描述的具体实施例仅用于解释本申请,而非对本申请的限定。另外还需要说明的是,为了便于描述,附图中仅示出了与本申请相关的部分而非全部结构。基于本申请中的实施例,本领域普通技术人员在没有作出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。The technical solutions in the embodiments of the present application will be clearly and completely described below in conjunction with the drawings in the embodiments of the present application. It can be understood that the specific embodiments described here are only used to explain the application, but not to limit the application. In addition, it should be noted that, for ease of description, the drawings only show a part of the structure related to the present application instead of all of the structure. Based on the embodiments in this application, all other embodiments obtained by those of ordinary skill in the art without creative work fall within the protection scope of this application.
本申请中的术语“第一”、“第二”等是用于区别不同对象,而不是用于描述特定顺序。此外,术语“包括”和“具有”以及它们任何变 形,意图在于覆盖不排他的包含。例如包含了一系列步骤或单元的过程、方法、系统、产品或设备没有限定于已列出的步骤或单元,而是可选地还包括没有列出的步骤或单元,或可选地还包括对于这些过程、方法、产品或设备固有的其它步骤或单元。The terms "first", "second", etc. in this application are used to distinguish different objects, rather than to describe a specific sequence. In addition, the terms "including" and "having" and any variations thereof are intended to cover non-exclusive inclusion. For example, a process, method, system, product, or device that includes a series of steps or units is not limited to the listed steps or units, but optionally includes unlisted steps or units, or optionally also includes Other steps or units inherent to these processes, methods, products or equipment.
在本文中提及“实施例”意味着,结合实施例描述的特定特征、结构或特性可以包含在本申请的至少一个实施例中。在说明书中的各个位置出现该短语并不一定均是指相同的实施例,也不是与其它实施例互斥的独立的或备选的实施例。本领域技术人员显式地和隐式地理解的是,本文所描述的实施例可以与其它实施例相结合。Reference to "embodiments" herein means that a specific feature, structure, or characteristic described in conjunction with the embodiments may be included in at least one embodiment of the present application. The appearance of the phrase in various places in the specification does not necessarily refer to the same embodiment, nor is it an independent or alternative embodiment mutually exclusive with other embodiments. Those skilled in the art clearly and implicitly understand that the embodiments described herein can be combined with other embodiments.
参阅图1,图1是本申请提供的身份验证方法第一实施例的流程示意图,该方法包括:Referring to Fig. 1, Fig. 1 is a schematic flowchart of a first embodiment of an identity verification method provided by the present application. The method includes:
步骤11:终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。Step 11: When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
终端设备在获取到设定操作指令时,打来摄像头模组,采集第一待检测图像,并对第一待检测图像进行人脸识别。在本实施例中,终端设备可以是移动终端,如智能手机、平板电脑、可穿戴设备等,设定操作指令可以是解锁屏幕指令,当设备终端获取到解锁屏幕指令时,打开摄像头,采集当前摄像头拍摄范围内的图像信息,对采集到的图像信息进行人脸识别。可以理解,当采集的图像信息中没有人脸信息时,终端设备停止识别,或提示重新采集并再次识别。When the terminal device obtains the setting operation instruction, it calls the camera module, collects the first image to be detected, and performs face recognition on the first image to be detected. In this embodiment, the terminal device may be a mobile terminal, such as a smart phone, a tablet computer, a wearable device, etc., and the setting operation instruction may be an unlock screen instruction. When the device terminal obtains the unlock screen instruction, it turns on the camera to collect the current The image information within the shooting range of the camera is used for face recognition on the collected image information. It can be understood that when there is no face information in the collected image information, the terminal device stops the recognition, or prompts to re-collect and recognize again.
人脸识别可以分为人脸图像采集及检测、人脸图像预处理、人脸图像特征提取以及匹配与识别。Face recognition can be divided into face image acquisition and detection, face image preprocessing, face image feature extraction, and matching and recognition.
人脸图像采集:不同的人脸图像都能通过摄像镜头采集下来,比如静态图像、动态图像、不同的位置、不同表情等方面都可以得到很好的采集。当用户在采集设备的拍摄范围内时,采集设备会自动搜索并拍摄用户的人脸图像。Face image collection: Different face images can be collected through the camera lens, such as static images, dynamic images, different positions, different expressions, etc. can be well collected. When the user is within the shooting range of the capture device, the capture device will automatically search for and shoot the user's face image.
人脸检测:人脸检测在实际中主要用于人脸识别的预处理,即在图像中准确标定出人脸的位置和大小。人脸图像中包含的模式特征十分丰富,如直方图特征、颜色特征、模板特征、结构特征及Haar特征等。人脸检测就是把这其中有用的信息挑出来,并利用这些特征实现人脸检测。人脸检测过程中使用Adaboost算法挑选出一些最能代表人脸的矩形特征(弱分类器),按照加权投票的方式将弱分类器构造为一个强分类器,再将训练得到的若干强分类器串联组成一个级联结构的层叠分类器,有效地提高分类器的检测速度。Face detection: In practice, face detection is mainly used for preprocessing of face recognition, that is, to accurately calibrate the position and size of the face in the image. The pattern features contained in face images are very rich, such as histogram features, color features, template features, structural features, and Haar features. Face detection is to pick out the useful information, and use these features to realize face detection. In the face detection process, the Adaboost algorithm is used to select some rectangular features (weak classifiers) that best represent the face, and the weak classifier is constructed into a strong classifier according to the weighted voting method, and then several strong classifiers obtained by training A cascade structure of stacked classifiers is formed in series, which effectively improves the detection speed of the classifier.
人脸图像预处理:对于人脸的图像预处理是基于人脸检测结果,对 图像进行处理并最终服务于特征提取的过程。系统获取的原始图像由于受到各种条件的限制和随机干扰,往往不能直接使用,必须在图像处理的早期阶段对它进行灰度校正、噪声过滤等图像预处理。对于人脸图像而言,其预处理过程主要包括人脸图像的光线补偿、灰度变换、直方图均衡化、归一化、几何校正、滤波以及锐化等。Face image preprocessing: The image preprocessing of the face is based on the face detection result, the image is processed and finally serves the process of feature extraction. Due to various conditions and random interference, the original image acquired by the system cannot be used directly. It must be preprocessed by grayscale correction and noise filtering in the early stage of image processing. For face images, the preprocessing process mainly includes light compensation, gray scale transformation, histogram equalization, normalization, geometric correction, filtering and sharpening of the face image.
人脸图像特征提取:人脸识别可使用的特征通常分为视觉特征、像素统计特征、人脸图像变换系数特征、人脸图像代数特征等。人脸特征提取就是针对人脸的某些特征进行的。人脸特征提取,也称人脸表征,它是对人脸进行特征建模的过程。人脸特征提取的方法归纳起来分为两大类:一种是基于知识的表征方法;另外一种是基于代数特征或统计学习的表征方法。Face image feature extraction: The features that can be used in face recognition are usually divided into visual features, pixel statistical features, face image transformation coefficient features, and face image algebraic features. Facial feature extraction is based on certain features of the human face. Face feature extraction, also known as face representation, is a process of feature modeling of human faces. Facial feature extraction methods can be summarized into two categories: one is knowledge-based representation methods; the other is based on algebraic features or statistical learning.
基于知识的表征方法主要是根据人脸器官的形状描述以及他们之间的距离特性来获得有助于人脸分类的特征数据,其特征分量通常包括特征点间的欧氏距离、曲率和角度等。人脸由眼睛、鼻子、嘴、下巴等局部构成,对这些局部和它们之间结构关系的几何描述,可作为识别人脸的重要特征,这些特征被称为几何特征。基于知识的人脸表征主要包括基于几何特征的方法和模板匹配法。The knowledge-based representation method is mainly based on the shape description of the face organs and the distance characteristics between them to obtain feature data that is helpful for face classification. Its feature components usually include the Euclidean distance, curvature, and angle between feature points. . The human face is composed of parts such as eyes, nose, mouth, and chin. The geometric description of these parts and the structural relationship between them can be used as important features to recognize the face. These features are called geometric features. Knowledge-based face representation mainly includes geometric feature-based methods and template matching methods.
人脸图像匹配与识别:提取的人脸图像的特征数据与数据库中存储的特征模板进行搜索匹配,通过设定一个阈值,当相似度超过这一阈值,则把匹配得到的结果输出。人脸识别就是将待识别的人脸特征与已得到的人脸特征模板进行比较,根据相似程度对人脸的身份信息进行判断。这一过程又分为两类:一类是确认,是一对一进行图像比较的过程,另一类是辨认,是一对多进行图像匹配对比的过程。Face image matching and recognition: The feature data of the extracted face image is searched and matched with the feature template stored in the database. By setting a threshold, when the similarity exceeds this threshold, the matching result is output. Face recognition is to compare the facial features to be recognized with the obtained facial feature template, and judge the identity information of the face based on the degree of similarity. This process is divided into two categories: one is confirmation, which is a process of one-to-one image comparison, and the other is identification, which is a process of one-to-many image matching and comparison.
步骤12:在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别。Step 12: After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected images.
在人脸识别通过后,采集连续的多个第二待检测图像的方式可以是用户说出一段文字,终端设备的摄像头模组采集连续的图像。After the face recognition is passed, the manner of collecting consecutive multiple second to-be-detected images may be that the user speaks a paragraph of text, and the camera module of the terminal device collects consecutive images.
首先识别连续图像中的人脸信息,然后提取人脸中连续的唇形变化特征,再进行唇形单元匹配,将唇形特征输入到唇形识别模型中,识别出对应的发音,再将识别出的发音与密码字符进行匹配,得到用户说出的文字信息。First recognize the face information in the continuous image, then extract the continuous lip shape change features in the face, then perform lip unit matching, input the lip shape feature into the lip shape recognition model, identify the corresponding pronunciation, and then recognize The pronunciation is matched with the characters of the password, and the text information spoken by the user is obtained.
唇形识别模型是通过复杂端到端深度神经网络技术进行唇语序列建模,建立词汇表。The lip shape recognition model uses complex end-to-end deep neural network technology to model the lip sequence and establish a vocabulary.
在本实施中,可以由终端设备语音播报一段文字,由用户重复此段文字,采集用户重复此段文字的连续图像,将连续图像进行唇形特征提 取,通过唇形识别模型识别出对应发音,将对应发音进行密码字符匹配,得出文字信息,将此文字信息与终端设备语音播报的文字信息进行匹配,匹配通过,则执行步骤13。In this implementation, the terminal device can broadcast a paragraph of text by voice, the user repeats this paragraph of text, collects continuous images of the user repeats this paragraph of text, extracts the lip features of the continuous images, and recognizes the corresponding pronunciation through the lip recognition model. The corresponding pronunciation is matched with password characters to obtain text information, and this text information is matched with the text information voiced by the terminal device. If the matching is passed, then step 13 is executed.
在本实施例中,可以由终端设备的显示屏显示一段文字,由用户读出此段文字,采集用户读此段文字的连续图像,将连续图像进行唇形特征提取,通过唇形识别模型识别出对应发音,将对应发音进行密码字符匹配,得出文字信息,将此文字信息与终端设备语音播报的文字信息进行匹配,匹配通过,则执行步骤13。可以理解,终端设备显示的文字可以是预设的文字信息,也可以是随机的文字信息。In this embodiment, a section of text can be displayed on the display screen of the terminal device, and the user can read this section of text, collect continuous images of the text read by the user, extract the lip features of the continuous images, and recognize them through the lip recognition model The corresponding pronunciation is generated, the corresponding pronunciation is matched with the password characters, and the text information is obtained, and the text information is matched with the text information voiced by the terminal device. If the matching is passed, then step 13 is executed. It can be understood that the text displayed by the terminal device may be preset text information or random text information.
步骤13:在唇形识别通过后,响应设定操作指令。Step 13: After the lip shape recognition is passed, respond to the setting operation instruction.
在唇形识别通过后,若设定操作指令为解锁屏幕,则终端设备解锁屏幕,显示屏幕内容;若设定操作指令为解锁终端设备私密相册,则终端设备解锁私密相册,展示私密相册的相片;若设定操作指令为支付指令,则终端设备完成相应的支付;若设定操作指令为查看私密信息,则终端设备展示私密信息。After the lip recognition is passed, if the operation instruction is set to unlock the screen, the terminal device will unlock the screen and display the screen content; if the operation instruction is set to unlock the terminal device’s private album, the terminal device will unlock the private album and display the photos in the private album ; If the set operation instruction is a payment instruction, the terminal device completes the corresponding payment; if the set operation instruction is to view private information, the terminal device displays the private information.
在一个应用场景中,用户需要使用终端设备进行账单支付时,终端设备获取到支付指令,提示用户进行人脸识别,通过人脸识别后,提示用户说一段话,在用户说话时同步采集连续的图像信息,将图像信息进行唇形特征提取,进行唇形识别,在唇形识别通过后,终端设备完成相应的支付。In an application scenario, when a user needs to use a terminal device to pay a bill, the terminal device obtains a payment instruction and prompts the user to perform face recognition. After face recognition, the user is prompted to say a paragraph, and the user is synchronized to collect continuous For image information, lip feature extraction is performed on the image information, and lip shape recognition is performed. After the lip shape recognition is passed, the terminal device completes the corresponding payment.
在另一个应用场景中,用户点击终端设备上的一应用程序,此时应用程序需要身份验证,终端设备获取到操作指令,提示用户进行人脸识别,通过人脸识别后,在显示屏上显示文字信息,提示用户读出文字信息,在用户读文字信息时同步采集连续的图像信息,将图像信息进行唇形提取,唇形识别,在唇形识别通过后,解锁应用程序。In another application scenario, the user clicks on an application on the terminal device. At this time, the application requires identity verification. The terminal device obtains the operation instruction and prompts the user to perform face recognition. After the face recognition is passed, it is displayed on the display The text information prompts the user to read the text information. When the user reads the text information, the continuous image information is synchronously collected, and the image information is subjected to lip shape extraction and lip shape recognition. After the lip shape recognition is passed, the application is unlocked.
在其他实施例中,终端设备在获取到设定操作指令后,对用户进行人脸识别,人脸识别通过后,获取用户说出文字信息的视频信息,终端设备将视频信息拆分为音频流和图像流,对音频流进行语音识别,识别出文字信息,对图像流进行连续的唇形特征提取,通过唇形识别模型进行识别,计算出图像流中唇形特征所包含的文字信息,音频流识别出的文字信息和图像流识别出的文字信息进行对比,如相同,则将识别出的文字信息与预设的文字信息进行匹配,匹配成功,则认为身份验证成功,终端设备响应设定操作指令。In other embodiments, after obtaining the setting operation instruction, the terminal device performs face recognition on the user. After the face recognition is passed, it obtains the video information of the text information spoken by the user, and the terminal device splits the video information into audio streams. And the image stream, perform voice recognition on the audio stream, recognize the text information, perform continuous lip feature extraction on the image stream, recognize through the lip recognition model, and calculate the text information contained in the lip feature in the image stream. Audio The text information recognized by the stream is compared with the text information recognized by the image stream. If the same, the recognized text information is matched with the preset text information. If the matching is successful, the identity verification is considered successful, and the terminal device responds to the setting Operating instructions.
区别于现有技术的情况,本申请的一种应用程序的操作方法,一种身份验证方法,该方法包括:终端设备在获取到设定操作指令时,采集 第一待检测图像,并对第一待检测图像进行人脸识别;在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别;在唇形识别通过后,响应设定操作指令。通过上述方式,实现验证方式难复制、抗遗忘、不接触等特点,提高身份验证的准确性、让终端设备的使用更加安全。Different from the situation in the prior art, an application operation method and an identity verification method of the present application include: when a terminal device obtains a setting operation instruction, collecting a first image to be detected, Perform face recognition on an image to be detected; after the face recognition is passed, collect consecutive multiple second to be detected images, and perform lip shape recognition on the second to be detected image; after the lip shape recognition is passed, respond to the setting operation instruction. Through the above method, the characteristics of the verification method such as difficult to copy, anti-forgetting, and non-contact are realized, the accuracy of identity verification is improved, and the use of terminal equipment is more secure.
参阅图2,图2是本申请提供的身份验证方法第二实施例的流程示意图,该方法包括:Refer to Figure 2. Figure 2 is a schematic flowchart of a second embodiment of the identity verification method provided by the present application, and the method includes:
步骤21:终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。Step 21: When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
步骤22:在人脸识别通过后,采集连续的多个第二待检测图像。Step 22: After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected.
可选的,在人脸识别通过后,终端设备提示用户进行第二待检测图像采集,如提示用户面部正对摄像头并说出文字信息。Optionally, after the face recognition is passed, the terminal device prompts the user to collect the second to-be-detected image, such as prompting the user to face the camera and speak text information.
可选的,当人脸识别通过后,一定程度上是认可当前用户的身份的,所以在提示用户进行第二待检测图像采集时,可以通过相关信息的引导,提示用户说出预设白名单中的文字信息。Optionally, when the face recognition is passed, the identity of the current user is recognized to a certain extent, so when the user is prompted to collect the second to-be-detected image, the user can be guided by relevant information to prompt the user to say the preset whitelist Text messages in.
步骤23:对多个第二待检测图像进行唇形识别,以得到识别文字信息。Step 23: Perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information.
唇形识别技术是一种依据说话时嘴唇的动作来解读说话内容的技术,在进行唇形自动识别时,需采集包含说话人嘴唇动作的多张图像或者采集包含说话人嘴唇动作的视频,然后结合图像处理技术、深度学习技术,识别多帧连续图像序列,通过识别多帧连续图像序列中的唇形,将唇形映射到发音,再根据连续时间段内的发音情况确定对应的自然语言词句,即说话内容。Lip shape recognition technology is a technology that interprets the content of speech based on the movement of the lips when speaking. When performing automatic lip shape recognition, it is necessary to collect multiple images containing the lips movement of the speaker or collect a video containing the lips movement of the speaker, and then Combining image processing technology and deep learning technology to identify multi-frame continuous image sequences, by identifying the lip shape in the multi-frame continuous image sequence, mapping the lip shape to the pronunciation, and then determining the corresponding natural language words and sentences based on the pronunciation in a continuous period of time , That is, the content of the speech.
可选的,对采集连续的多个第二待检测图像提取多个第二待检测图像的人脸信息;从多个人脸信息中提取多个连续变化的唇形特征;将多个连续变化的唇形特征输入至唇形识别模型,以使唇形识别模型出对应的发音信息,并基于发音信息,计算出对应的识别文字信息。Optionally, extract the face information of multiple second to-be-detected images from the collection of multiple consecutive second to-be-detected images; extract multiple continuously changing lip features from the multiple face information; combine multiple continuously changing lip features The lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain corresponding pronunciation information, and based on the pronunciation information, the corresponding recognized text information is calculated.
具体地,唇形识别模型可以是基于编码器-解码器架构融合时空卷积神经网络的特征提取器和词嵌入网络,并使用注意力机制的端到端的算法模型。其中特征提取器使用的是时空卷积神经网络(STCNN),编码器-解码器子单元采用的是长短时记忆网络(LSTM),词嵌入(Embedding)编码方式采用的是Word2vec。Specifically, the lip recognition model may be an end-to-end algorithm model based on the encoder-decoder architecture fusion spatiotemporal convolutional neural network feature extractor and word embedding network, and using the attention mechanism. Among them, the feature extractor uses a spatio-temporal convolutional neural network (STCNN), the encoder-decoder subunit uses a long short-term memory network (LSTM), and the word embedding (Embedding) encoding method uses Word2vec.
可选的,唇形识别模型可以使用汉语普通话的唇形识别数据集来训练模型,使用改进的多阶段卷积神经网络(MTCNN)提取静默视频中唇部区域,而后将提取的唇部区域送入时空卷积网络STCNN中,用于提取 唇部动作的视觉特征信息。基于LSTM的编码器-解码器用于将唇部视觉特征信息进行编码并在模型推断时,将其解码成为相关的文本信息。注意力机制可以使得模型解码器关注特定位置的编码器编码内容,而不用将整个编码内容都作为解码的依据,进而提高模型解码效果。使用优化后的THULAC(THU Lexical Analyzer for Chinese,中文词法分析工具包)来对汉字语句进行分词,分词后的结果送入Word2vec,该部分在网络中的作用本质上来说是充当字符编码的作用。编码器-解码器架构将可变长度序列编码为固定长度表示,并将给定的固定长度向量表示为可变长序列。从概率角度看,该模型是在一个可变长度序列的条件下,使用通用的方法来学习另一个可变长序列的条件概率分布。Optionally, the lip shape recognition model can use the lip shape recognition data set of Mandarin Chinese to train the model, use an improved multi-stage convolutional neural network (MTCNN) to extract the lip region in the silent video, and then send the extracted lip region In the spatio-temporal convolutional network STCNN, it is used to extract the visual feature information of the lip action. The encoder-decoder based on LSTM is used to encode lip visual feature information and decode it into relevant text information during model inference. The attention mechanism can make the model decoder pay attention to the coded content of the encoder at a specific location, instead of using the entire coded content as a basis for decoding, thereby improving the decoding effect of the model. Use optimized THULAC (THU Lexical Analyzer for Chinese, Chinese lexical analysis toolkit) to segment Chinese character sentences, and the result of segmentation is sent to Word2vec. The role of this part in the network is essentially to act as a character encoding. The encoder-decoder architecture encodes a variable-length sequence into a fixed-length representation, and represents a given fixed-length vector as a variable-length sequence. From a probabilistic point of view, the model uses a general method to learn the conditional probability distribution of another variable-length sequence under the condition of a variable-length sequence.
可以理解,唇形识别模型可以根据不同语言,采用上述方案或其他相关方案建立不同的数据库,以应用于不同语言地区。It can be understood that the lip shape recognition model can use the above scheme or other related schemes to establish different databases according to different languages, so as to be applied to different language regions.
步骤24:判断识别文字信息是否为预设白名单中的文字信息。Step 24: Determine whether the recognized text information is the text information in the preset whitelist.
可选的,用户在终端设备中录入一些文字信息,并将录入的文字信息加入白名单。Optionally, the user enters some text information in the terminal device, and adds the entered text information to the whitelist.
判断当通过唇形检测后识别出的文字信息与白名单中的文字信息相同时,则执行步骤25。When it is determined that the text information recognized after the lip shape detection is the same as the text information in the white list, step 25 is executed.
可选的,白名单中的文字可以只有一段,也可以是多段。当白名单中为多段文字时,只需要通过唇形检测后识别出的文字信息是其中任一段即可。Optionally, the text in the whitelist can have only one paragraph or multiple paragraphs. When there are multiple paragraphs of text in the whitelist, only any paragraph of the text information recognized after the lip shape detection is required.
步骤25:确定唇形识别通过。Step 25: Confirm that the lip shape recognition passes.
步骤26:在唇形识别通过后,响应设定操作指令。Step 26: After the lip shape recognition is passed, respond to the setting operation instruction.
在一应用场景中,终端设备包含有私密短信息,需要进行身份验证方可查看。当用户点击查看私密短信息,终端设备响应此操作指令,对用户进行人脸识别,通过人脸识别后,终端设备通过摄像头采集用户读出文字信息的连续的多个图像进行检测,获取图像中的唇形特征,通过唇形识别模型计算出对应的文字信息,将对应的文字信息与白名单中的预设文字信息进行匹配,匹配成功,则确定唇形识别通过,终端设备则响应设定操作指令,显示私密短信息供用户查看。In an application scenario, the terminal device contains private short messages, which need to be authenticated before they can be viewed. When the user clicks to view the private short message, the terminal device responds to this operation instruction to perform face recognition on the user. After face recognition, the terminal device uses the camera to collect multiple consecutive images of the text information read by the user for detection. According to the lip shape feature, the corresponding text information is calculated through the lip shape recognition model, and the corresponding text information is matched with the preset text information in the whitelist. If the matching is successful, the lip shape recognition is determined to pass, and the terminal device responds to the setting Operation instructions, display private short messages for users to view.
参阅图3,图3是本申请提供的身份验证方法第三实施例的流程示意图,该方法包括:Refer to Fig. 3, which is a schematic flowchart of a third embodiment of an identity verification method provided by the present application, and the method includes:
步骤31:终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。Step 31: When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
步骤32:在人脸识别通过后,采集连读的多个第二待检测图像。Step 32: After the face recognition is passed, a plurality of second to-be-detected images that are read consecutively are collected.
步骤31-32与上述实施例有相同或相似的技术方案,这里不做赘述。Steps 31-32 have the same or similar technical solutions as the foregoing embodiment, and will not be repeated here.
步骤33:对多个第二待检测图像进行唇形识别,以得到识别文字信息。Step 33: Perform lip shape recognition on a plurality of second to-be-detected images to obtain recognized text information.
可选的,对采集连续的多个第二待检测图像提取多个第二待检测图像的人脸信息;从多个人脸信息中提取多个连续变化的唇形特征;将多个连续变化的唇形特征输入至唇形识别模型,以使唇形识别模型出对应的发音信息,并基于发音信息,计算出对应的识别文字信息。Optionally, extract the face information of multiple second to-be-detected images from the collection of multiple consecutive second to-be-detected images; extract multiple continuously changing lip features from the multiple face information; combine multiple continuously changing lip features The lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain corresponding pronunciation information, and based on the pronunciation information, the corresponding recognized text information is calculated.
步骤34:判断识别文字信息是否为预设黑名单中的文字信息。Step 34: Determine whether the recognized text information is the text information in the preset blacklist.
可选的,用户在终端设备中录入一些文字信息,并将录入的文字信息加入黑名单。Optionally, the user enters some text information in the terminal device, and adds the entered text information to the blacklist.
判断当通过唇形检测后识别出的文字信息与黑名单中的文字信息相同时,则执行步骤35。When it is determined that the text information recognized after the lip shape detection is the same as the text information in the blacklist, step 35 is executed.
可选的,黑名单中的文字可以只有一段,也可以是多段。当黑名单中为多段文字时,只需要通过唇形检测后识别出的文字信息是其中任一段即可。Optionally, the text in the blacklist can have only one paragraph or multiple paragraphs. When there are multiple paragraphs of text in the blacklist, only any paragraph of the text information recognized after the lip shape detection is required.
可选的,将白名单中的至少一段文字信息加入黑名单中,并将文字信息在白名单中删除。Optionally, at least one piece of text information in the white list is added to the black list, and the text information is deleted from the white list.
步骤35:确定唇形识别不通过。Step 35: Determine that the lip shape recognition fails.
可选的,当识别出的文字信息与黑名单中的文字信息不同时,将识别出的文字信息与白名单中的文字信息进行匹配,若相同,则唇形识别通过,终端设备响应设定操作指令。Optionally, when the recognized text information is different from the text information in the blacklist, the recognized text information is matched with the text information in the white list. If they are the same, the lip recognition passes and the terminal device responds to the settings Operating instructions.
在一应用场景中,用户发现终端设备的白名单中部分文字信息有被盗的风险或已经被盗,则将这部分文字信息从白名单中删除并加入黑名单中,使这部分文字信息用于验证唇形识别出的文字信息是否安全。In an application scenario, if the user finds that part of the text information in the whitelist of the terminal device is at risk of being stolen or has been stolen, then this part of the text information is deleted from the whitelist and added to the blacklist, so that this part of the text information can be used To verify whether the text information recognized by the lip shape is safe.
在一应用场景中,终端设备的白名单的每段文字信息都是有时效性的。如每段文字信息用于身份验证有时间限制(时间限制可以是两小时、二十小时、四十八小时,具体限制的时间由系统设置或用户需求设置),当文字信息超过时间限制,终端设备会自动将其删除并加入黑名单中,并提示用户或提示用户该文字信息已超过时间限制或许有安全隐患,请用户自行处理。如每段文字信息用于身份验证有次数限制(次数限制可以是十次、二十次、五十次、一百次,具体限制次数由系统设置或用户需求设置),当文字信息超过次数限制,终端设备会自动将其删除并加入黑名单中并提示用户或提示用户该文字信息已超过使用次数或许有安全隐患,请用户自行处理。这样保证白名单中文字信息的迭代更新,易于保证信息安全,不易被盗,即使被盗,文字信息也已经在黑名单中,使用该文字信息不能通过身份验证In an application scenario, each piece of text information in the whitelist of the terminal device is time-sensitive. For example, there is a time limit for each piece of text information used for identity verification (the time limit can be two hours, twenty hours, forty-eight hours, and the specific time limit is set by the system or user requirements). When the text message exceeds the time limit, the terminal device will Automatically delete it and add it to the blacklist, and prompt the user or remind the user that the text message has exceeded the time limit may be a security risk, please handle it by yourself. For example, there is a limit on the number of times each piece of text information is used for identity verification (the limit can be ten, twenty, fifty, one hundred, and the specific limit is set by the system or user needs). When the text information exceeds the limit , The terminal device will automatically delete it and add it to the blacklist and prompt the user or remind the user that the text message has exceeded the number of uses may be a security risk, please handle it by yourself. This ensures the iterative update of the text information in the whitelist, which is easy to ensure information security and is not easy to be stolen. Even if the text information is stolen, the text information is already in the blacklist, and the use of the text information cannot pass identity verification
参阅图4,图4是本申请提供的身份验证方法第四实施例的流程示意图,该方法包括:Refer to FIG. 4, which is a schematic flowchart of a fourth embodiment of an identity verification method provided by the present application, and the method includes:
步骤41:终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。Step 41: When the terminal device obtains the setting operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
步骤42:在人脸识别通过后,显示标准文字信息,并采集连续的多个第二待检测图像。Step 42: After the face recognition is passed, the standard text information is displayed, and a plurality of consecutive second to-be-detected images are collected.
终端设备在用户人脸识别通过后,显示屏上显示标准文字,提示用户面部正对摄像头读出显示的标准文字,同时通过摄像头采集用户读标准文字时的连续的多个第二待检测图像。After the user's face recognition is passed, the terminal device displays standard text on the display screen, prompting the user's face to face the camera to read the displayed standard text, and at the same time collect multiple consecutive second to-be-detected images when the user reads the standard text through the camera.
可选的,标准文字信息可以是用户提前录入的多个文字信息中的一个。Optionally, the standard text information may be one of multiple text information entered in advance by the user.
可选的,标准文字信息可以是从数据库中多个文字信息中随机选择一个文字信息作为标准文字信息。Optionally, the standard text information may be randomly selected from a plurality of text information in the database as the standard text information.
可选的,标准文字信息可以是从云服务器中随机选择一个文字信息作为标准文字信息。Optionally, the standard text information may be randomly selected from the cloud server as the standard text information.
步骤43:对多个第二待检测图像进行唇形识别,以得到识别文字信息。Step 43: Perform lip recognition on multiple second to-be-detected images to obtain recognized text information.
可选的,步骤43具体为对采集连续的多个第二待检测图像提取多个第二待检测图像的人脸信息;从多个人脸信息中提取多个连续变化的唇形特征;将多个连续变化的唇形特征输入至唇形识别模型,以使唇形识别模型出对应的发音信息,并基于发音信息,计算出对应的识别文字信息。Optionally, step 43 is specifically extracting the face information of multiple second to-be-detected images from the continuous collection of multiple second to-be-detected images; extracting multiple continuously changing lip features from the multiple face information; A continuously changing lip shape feature is input to the lip shape recognition model, so that the lip shape recognition model can obtain the corresponding pronunciation information, and based on the pronunciation information, the corresponding recognition text information is calculated.
步骤44:判断识别文字信息与标准文字信息是否相同。Step 44: Determine whether the recognized text information is the same as the standard text information.
判断识别文字信息与标准文字信息是否相同,若相同执行步骤45。It is judged whether the recognized text information is the same as the standard text information, and if the same, step 45 is executed.
步骤45:确定唇形识别通过。Step 45: Confirm that the lip shape recognition passes.
确定唇形识别通过后,终端设备响应设定操作指令,完成相应操作。After confirming that the lip shape recognition is passed, the terminal device responds to the setting operation instruction to complete the corresponding operation.
参阅图5,图5是本申请提供的身份验证方法第五实施例的流程示意图,该方法包括:Referring to Fig. 5, Fig. 5 is a schematic flowchart of a fifth embodiment of an identity verification method provided by the present application. The method includes:
步骤51:移动终端在获取到设定操作指令时,采集第一待检测图像。Step 51: The mobile terminal collects the first image to be detected when acquiring the setting operation instruction.
步骤52:提取第一待检测图像中的人脸图像。Step 52: Extract the face image in the first image to be detected.
可选的,如未提取到第一待检测图像的人脸图像,终端设备将会重新采集第一待检测图像并提示用户正面面对摄像头,以便于采集的第一待检测图像包含人脸图像。Optionally, if the face image of the first image to be detected is not extracted, the terminal device will re-acquire the first image to be detected and prompt the user to face the camera, so that the collected first image to be detected contains the face image .
步骤53:从人脸图像中提取人脸特征信息。Step 53: Extract face feature information from the face image.
可选的,对于人脸图像中提取人脸特征信息的方法可以采用局部特 征提取法。Optionally, a local feature extraction method can be used for the method for extracting facial feature information from the facial image.
具体地,可以采用基于面部器官的特征提取、基于模板的特征提取、基于弹性图匹配法的特征提取。Specifically, feature extraction based on facial organs, feature extraction based on templates, and feature extraction based on elastic map matching methods can be used.
可选的,对于人脸图像中提取人脸特征信息的方法可以采用整体特征提取法。Optionally, the method for extracting facial feature information from the facial image may adopt an overall feature extraction method.
具体地,可以采用基于代数方法的特征提取、基于神经网络的特征提取、基于小波多分辨率的特征提取。Specifically, feature extraction based on algebraic method, feature extraction based on neural network, feature extraction based on wavelet multi-resolution can be used.
步骤54:将人脸特征信息与预存的标准人脸特征信息进行相似度比对。Step 54: Perform a similarity comparison between the facial feature information and the pre-stored standard facial feature information.
可选的,预存的标准人脸特征信息为用户提前采集的人脸图像信息所提取出的人脸特征信息,预存的标准人脸特征信息可以以组为单位,每组里的人脸特征信息构成一个人脸图像,这样终端设备里可以预存多组标准人脸特征信息。Optionally, the pre-stored standard facial feature information is the facial feature information extracted from the facial image information collected in advance by the user. The pre-stored standard facial feature information can be grouped as a unit, and the facial feature information in each group A face image is formed so that multiple sets of standard facial feature information can be pre-stored in the terminal device.
步骤55:在相似度比对的结果满足预设要求时,确定人脸识别通过。Step 55: When the result of the similarity comparison meets the preset requirements, it is determined that the face recognition passes.
可选的,相似度的比对方式可以是以单个特征与单个标准人脸特征进行比对,然后将多个单个特征的比对结果相乘,相乘的结果大于一个预设值时,确定人脸识别通过。以单个特征为鼻子、眼睛、嘴巴为例,鼻子的相似度比对值为0.95、眼睛的相似度比对值为0.85、嘴巴的相似度比对值为0.99,将三个比对值相乘为0.95*0.85*0.99≈0.8,预设值为0.75,0.8>0.75,所以相似度比对结果大于预设要求,确定人脸识别通过。Optionally, the similarity comparison method can be to compare a single feature with a single standard facial feature, and then multiply the comparison results of multiple single features. When the multiplied result is greater than a preset value, it is determined Face recognition passed. Taking a single feature as an example of nose, eyes, and mouth, the similarity comparison value of nose is 0.95, the similarity comparison value of eyes is 0.85, and the similarity comparison value of mouth is 0.99. Multiply the three comparison values. It is 0.95*0.85*0.99≈0.8, the preset value is 0.75, 0.8>0.75, so the similarity comparison result is greater than the preset requirement, and the face recognition is determined to pass.
可选的,相似度的比对方式可以是以整体特征与整体标准人脸特征进行比对,比对结果大于预设值时,确定人脸识别通过。Optionally, the similarity comparison method may be to compare the overall feature with the overall standard face feature, and when the comparison result is greater than a preset value, it is determined that the face recognition passes.
步骤56:在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别。Step 56: After the face recognition is passed, a plurality of consecutive second to-be-detected images are collected, and lip shape recognition is performed on the second to-be-detected images.
步骤57:在唇形识别通过后,响应设定操作指令。Step 57: After the lip shape recognition is passed, respond to the setting operation instruction.
参阅图6,图6是本申请提供的身份验证方法第六实施例的流程示意图,该方法包括:Referring to FIG. 6, FIG. 6 is a schematic flowchart of a sixth embodiment of an identity verification method provided by the present application. The method includes:
步骤61:终端设备在获取到支付操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。Step 61: When the terminal device obtains the payment operation instruction, it collects the first image to be detected, and performs face recognition on the first image to be detected.
步骤62:在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别。Step 62: After the face recognition is passed, collect a plurality of consecutive second to-be-detected images, and perform lip shape recognition on the second to-be-detected images.
步骤63:在唇形识别通过后,响应支付操作指令,以完成相应的支付。Step 63: After the lip shape recognition is passed, respond to the payment operation instruction to complete the corresponding payment.
在一应用场景中,当终端设备获取到支付操作指令时,如支付金额 属于小额支付,则可以跳过步骤61中的人脸识别,只需要当前用户正确读出设置于白名单中的任一段文字信息,当唇形识别通过后,终端设备响应支付操作指令,完成相应支付。In an application scenario, when the terminal device obtains a payment operation instruction, if the payment amount is a small payment, the face recognition in step 61 can be skipped, and the current user only needs to correctly read out the tasks set in the whitelist. A piece of text information. After the lip shape recognition is passed, the terminal device responds to the payment operation instruction to complete the corresponding payment.
在另一应用场景中,当终端设备获取到支付操作指令时,需要当前用户通过人脸识别,确定当前用户在终端设备拥有权限,然后正确读出设置于白名单中的任一段文字信息,当唇形识别通过后,终端设备响应支付操作指令,完成相应支付。In another application scenario, when the terminal device obtains the payment operation instruction, the current user needs to confirm that the current user has the authority in the terminal device through face recognition, and then correctly read any piece of text information set in the whitelist. After the lip shape recognition is passed, the terminal device responds to the payment operation instruction to complete the corresponding payment.
在另一应用场景中,当终端设备获取到支付操作指令时,需要当前用户通过人脸识别,确定当前用户在终端设备拥有权限,然后正确读出终端设备上随机的文字信息,当唇形识别通过后,终端设备响应支付操作指令,完成相应支付。In another application scenario, when the terminal device obtains the payment operation instruction, the current user needs to confirm that the current user has the authority in the terminal device through face recognition, and then correctly read the random text information on the terminal device. After passing, the terminal device responds to the payment operation instruction to complete the corresponding payment.
在另一应用场景中,用户发现终端设备中白名单中的文字信息出现安全隐患,如被他人盗取,则将有安全隐患的文字信息删除并加入黑名单。这样,即使通过了人脸识别,当唇形识别出文字信息为黑名单中的文字信息时,终端设备立即锁死,结束一切操作指令。In another application scenario, the user finds that the text information in the whitelist in the terminal device has hidden security risks. If it is stolen by others, the text information with hidden security risks will be deleted and added to the blacklist. In this way, even if the face recognition is passed, when the lip shape recognizes that the text information is the text information in the blacklist, the terminal device is immediately locked and all operation instructions are terminated.
在另一应用场景中,终端设备的用户分为不同权限的身份,最高权限用户可以快速更新白名单,以及确定支付权限用户名单,并且可以随时更改支付权限用户名单。In another application scenario, users of terminal devices are divided into identities with different permissions, and users with the highest permissions can quickly update the white list, determine the list of users with payment permissions, and can change the list of users with payment permissions at any time.
参阅图7,图7是本申请提供的终端设备第一实施例的结构示意图,该终端设备70包括处理器71以及与处理器71连接的摄像头模组72以及存储器73;存储器73用于存储程序数据,处理器71用于执行程序数据,以实现以下方法:Referring to FIG. 7, FIG. 7 is a schematic structural diagram of a first embodiment of a terminal device provided by the present application. The terminal device 70 includes a processor 71, a camera module 72 connected to the processor 71, and a memory 73; the memory 73 is used to store programs Data, the processor 71 is used to execute program data to implement the following methods:
终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别;在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别;在唇形识别通过后,响应设定操作指令。When the terminal device obtains the setting operation instruction, it collects the first image to be detected and performs face recognition on the first image to be detected; after the face recognition is passed, it collects a plurality of consecutive second images to be detected, and The second image to be detected performs lip shape recognition; after the lip shape recognition passes, responds to the setting operation instruction.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:在人脸识别通过后,采集连续的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识别文字信息;判断识别文字信息是否为预设白名单中的文字信息;若是,则确定唇形识别通过。Optionally, the processor 71 is used to execute the program data to implement the following method: after the face recognition is passed, a plurality of consecutive second to-be-detected images are collected; Recognize to obtain recognized text information; determine whether the recognized text information is the text information in the preset whitelist; if so, determine that the lip shape recognition passes.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:获取用户录入的文字信息;将录入的文字信息加入白名单中。Optionally, the processor 71 used to execute the program data is also used to implement the following methods: acquiring text information entered by the user; adding the entered text information to the white list.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:在人脸识别通过后,采集连读的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识别文字信息;判断识别文字信息是否为 预设黑名单中的文字信息;若是,则确定唇形识别不通过。Optionally, the processor 71 is configured to execute the program data to implement the following method: after the face recognition is passed, collect a plurality of second to-be-detected images that are read consecutively; Shape recognition to obtain recognized text information; determine whether the recognized text information is the text information in the preset blacklist; if so, determine that the lip shape recognition fails.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:将白名单中的至少一段文字信息加入黑名单中,并将文字信息在白名单中删除。。Optionally, the processor 71 used to execute the program data is also used to implement the following method: adding at least one piece of text information in the white list to the black list, and deleting the text information from the white list. .
可选地,处理器71用于执行该程序数据还用以实现以下的方法:在人脸识别通过后,显示标准文字信息,并采集连续的多个第二待检测图像;对多个第二待检测图像进行唇形识别,以得到识别文字信息;判断识别文字信息与标准文字信息是否相同;若是,则确定唇形识别通过。Optionally, the processor 71 is used to execute the program data to implement the following method: after the face recognition is passed, display standard text information, and collect consecutive multiple second to-be-detected images; Perform lip shape recognition on the image to be detected to obtain recognized text information; determine whether the recognized text information is the same as the standard text information; if yes, determine that the lip shape recognition passes.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:从数据库中的多个文字信息中随机选择一个文字信息作为标准文字信息,并显示标准文字信息。Optionally, the processor 71 used to execute the program data is also used to implement the following method: randomly select one text information from a plurality of text information in the database as the standard text information, and display the standard text information.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:提取多个第二待检测图像的人脸信息;从多个人脸信息中提取多个连续变化的唇形特征;基于多个连续变化的唇形特征,得到识别文字信息。Optionally, the processor 71 is used to execute the program data to implement the following method: extracting face information of multiple second images to be detected; extracting multiple continuously changing lip features from multiple face information; Based on multiple continuously changing lip features, the recognized text information is obtained.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:将多个连续变化的唇形特征输入至唇形识别模型,以使唇形识别模型出对应的发音信息,并基于发音信息,计算出对应的识别文字信息。Optionally, the processor 71 used to execute the program data is also used to implement the following method: input a plurality of continuously changing lip features into the lip recognition model, so that the lip recognition model can output corresponding pronunciation information, and Based on the pronunciation information, the corresponding recognized text information is calculated.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:移动终端在获取到设定操作指令时,采集第一待检测图像;提取第一待检测图像中的人脸图像;对人脸图像进行人脸识别。Optionally, the processor 71 is configured to execute the program data to implement the following method: when the mobile terminal obtains the setting operation instruction, collect the first image to be detected; extract the face image in the first image to be detected ; Perform face recognition on face images.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:从人脸图像中提取人脸特征信息;将人脸特征信息与预存的标准人脸特征信息进行相似度比对;在相似度比对的结果满足预设要求时,确定人脸识别通过。Optionally, the processor 71 is used to execute the program data to implement the following method: extracting facial feature information from a face image; comparing the facial feature information with pre-stored standard facial feature information for similarity ; When the result of the similarity comparison meets the preset requirements, the face recognition is determined to pass.
可选地,处理器71用于执行该程序数据还用以实现以下的方法:在唇形识别通过后,响应支付操作指令,以完成相应的支付。Optionally, the processor 71 used to execute the program data is also used to implement the following method: after the lip shape recognition is passed, respond to the payment operation instruction to complete the corresponding payment.
参阅图8,图8是本申请提供的终端设备第二实施例的结构示意图,该终端设备80包括:第一识别模块81、第二识别模块82和响应模块83。Referring to FIG. 8, FIG. 8 is a schematic structural diagram of a second embodiment of a terminal device provided by the present application. The terminal device 80 includes a first identification module 81, a second identification module 82 and a response module 83.
第一识别模块81用于在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别。The first recognition module 81 is configured to collect the first image to be detected and perform face recognition on the first image to be detected when the setting operation instruction is acquired.
第二识别模块82用于在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别。The second recognition module 82 is configured to collect a plurality of consecutive second to-be-detected images after the face recognition is passed, and perform lip-shape recognition on the second to-be-detected images.
响应模块83用于在唇形识别通过后,响应设定操作指令。The response module 83 is used to respond to the setting operation instruction after the lip shape recognition is passed.
参阅图9,图9是本申请提供的计算机存储介质一实施例的结构示 意图,该计算机存储介质90用于存储程序数据91,程序数据91在被处理器执行时,用于实现以下方法:Referring to Fig. 9, Fig. 9 is a schematic structural diagram of an embodiment of a computer storage medium provided by the present application. The computer storage medium 90 is used to store program data 91. When the program data 91 is executed by a processor, it is used to implement the following methods:
终端设备在获取到设定操作指令时,采集第一待检测图像,并对第一待检测图像进行人脸识别;在人脸识别通过后,采集连续的多个第二待检测图像,并对第二待检测图像进行唇形识别;在唇形识别通过后,响应设定操作指令。When the terminal device obtains the setting operation instruction, it collects the first image to be detected and performs face recognition on the first image to be detected; after the face recognition is passed, it collects a plurality of consecutive second images to be detected, and The second image to be detected performs lip shape recognition; after the lip shape recognition passes, responds to the setting operation instruction.
可以理解,程序数据91在被处理器执行时,还用于实现上述任一实施例方法。It can be understood that, when the program data 91 is executed by the processor, it is also used to implement the method in any of the foregoing embodiments.
在本申请所提供的几个实施方式中,应该理解到,所揭露的方法以及设备,可以通过其它的方式实现。例如,以上所描述的设备实施方式仅仅是示意性的,例如,所述模块或单元的划分,仅仅为一种逻辑功能划分,实际实现时可以有另外的划分方式,例如多个单元或组件可以结合或者可以集成到另一个系统,或一些特征可以忽略,或不执行。In the several implementation manners provided in this application, it should be understood that the disclosed method and device may be implemented in other ways. For example, the device implementation described above is merely illustrative. For example, the division of the modules or units is only a logical function division, and there may be other divisions in actual implementation, for example, multiple units or components may be Combined or can be integrated into another system, or some features can be ignored or not implemented.
所述作为分离部件说明的单元可以是或者也可以不是物理上分开的,作为单元显示的部件可以是或者也可以不是物理单元,即可以位于一个地方,或者也可以分布到多个网络单元上。可以根据实际的需要选择其中的部分或者全部单元来实现本实施方式方案的目的。The units described as separate components may or may not be physically separated, and the components displayed as units may or may not be physical units, that is, they may be located in one place, or they may be distributed on multiple network units. Some or all of the units may be selected according to actual needs to achieve the objectives of the solutions of this embodiment.
另外,在本申请各个实施方式中的各功能单元可以集成在一个处理单元中,也可以是各个单元单独物理存在,也可以两个或两个以上单元集成在一个单元中。上述集成的单元既可以采用硬件的形式实现,也可以采用软件功能单元的形式实现。In addition, the functional units in the various embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units may be integrated into one unit. The above-mentioned integrated unit can be implemented in the form of hardware or software functional unit.
上述其他实施方式中的集成的单元如果以软件功能单元的形式实现并作为独立的产品销售或使用时,可以存储在一个计算机可读取存储介质中。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分或者该技术方案的全部或部分可以以软件产品的形式体现出来,该计算机软件产品存储在一个存储介质中,包括若干指令用以使得一台计算机设备(可以是个人计算机,服务器,或者网络设备等)或处理器(processor)执行本申请各个实施方式所述方法的全部或部分步骤。而前述的存储介质包括:U盘、移动硬盘、只读存储器(ROM,Read-Only Memory)、随机存取存储器(RAM,Random Access Memory)、磁碟或者光盘等各种可以存储程序代码的介质。If the integrated units in the other embodiments described above are implemented in the form of software functional units and sold or used as independent products, they can be stored in a computer readable storage medium. Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology or all or part of the technical solution can be embodied in the form of a software product, and the computer software product is stored in a storage medium , Including several instructions to enable a computer device (which may be a personal computer, a server, or a network device, etc.) or a processor to execute all or part of the steps of the methods described in the various embodiments of the present application. The aforementioned storage media include: U disk, mobile hard disk, read-only memory (ROM, Read-Only Memory), random access memory (RAM, Random Access Memory), magnetic disk or optical disk and other media that can store program code .
以上仅为本申请的实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only examples of this application, and do not limit the scope of this application. Any equivalent structure or equivalent process transformation made by using the description and drawings of this application, or directly or indirectly applied to other related technical fields, The same reasoning is included in the scope of patent protection of this application.

Claims (15)

  1. 一种身份验证方法,其特征在于,所述方法包括:An identity verification method, characterized in that the method includes:
    终端设备在获取到设定操作指令时,采集第一待检测图像,并对所述第一待检测图像进行人脸识别;When the terminal device obtains the setting operation instruction, collects the first image to be detected, and performs face recognition on the first image to be detected;
    在所述人脸识别通过后,采集连续的多个第二待检测图像,并对所述第二待检测图像进行唇形识别;After the face recognition is passed, collecting a plurality of consecutive second to-be-detected images, and performing lip-shape recognition on the second to-be-detected images;
    在所述唇形识别通过后,响应所述设定操作指令。After the lip shape recognition is passed, respond to the setting operation instruction.
  2. 根据权利要求1所述的方法,其特征在于,The method according to claim 1, wherein:
    所述在所述人脸识别通过后,采集连续的多个第二待检测图像,并对所述第二待检测图像进行唇形识别,包括:After the face recognition is passed, collecting a plurality of consecutive second to-be-detected images and performing lip-shape recognition on the second to-be-detected images includes:
    在所述人脸识别通过后,采集连续的多个第二待检测图像;After the face recognition is passed, collecting a plurality of consecutive second to-be-detected images;
    对所述多个第二待检测图像进行唇形识别,以得到识别文字信息;Performing lip shape recognition on the plurality of second to-be-detected images to obtain recognized text information;
    判断所述识别文字信息是否为预设白名单中的文字信息;Determine whether the recognized text information is text information in a preset whitelist;
    若是,则确定所述唇形识别通过。If yes, it is determined that the lip shape recognition passes.
  3. 根据权利要求2所述的方法,其特征在于,The method according to claim 2, wherein:
    所述方法还包括:The method also includes:
    获取用户录入的文字信息;Obtain the text information entered by the user;
    将录入的所述文字信息加入所述白名单中。Add the entered text information to the white list.
  4. 根据权利要求1所述的方法,其特征在于,The method according to claim 1, wherein:
    所述在所述人脸识别通过后,采集连续的多个第二待检测图像,并对所述第二待检测图像进行唇形识别,包括:After the face recognition is passed, collecting a plurality of consecutive second to-be-detected images and performing lip-shape recognition on the second to-be-detected images includes:
    在所述人脸识别通过后,采集连读的多个第二待检测图像;After the face recognition is passed, collecting a plurality of second to-be-detected images read continuously;
    对所述多个第二待检测图像进行唇形识别,以得到识别文字信息;Performing lip shape recognition on the plurality of second to-be-detected images to obtain recognized text information;
    判断所述识别文字信息是否为预设黑名单中的文字信息;Determine whether the recognized text information is text information in a preset blacklist;
    若是,则确定所述唇形识别不通过。If yes, it is determined that the lip shape recognition fails.
  5. 根据权利要求4所述的方法,其特征在于,The method according to claim 4, wherein:
    所述方法还包括:The method also includes:
    将白名单中的至少一段文字信息加入所述黑名单中,并将所述文字信息在白名单中删除。At least one piece of text information in the white list is added to the black list, and the text information is deleted from the white list.
  6. 根据权利要求1所述的方法,其特征在于,The method according to claim 1, wherein:
    所述在所述人脸识别通过后,采集连续的多个第二待检测图像,并对所述第二待检测图像进行唇形识别,包括:After the face recognition is passed, collecting a plurality of consecutive second to-be-detected images and performing lip-shape recognition on the second to-be-detected images includes:
    在所述人脸识别通过后,显示标准文字信息,并采集连续的多个第二待检测图像;After the face recognition is passed, display standard text information, and collect consecutive multiple second to-be-detected images;
    对所述多个第二待检测图像进行唇形识别,以得到识别文字信息;Performing lip shape recognition on the plurality of second to-be-detected images to obtain recognized text information;
    判断所述识别文字信息与所述标准文字信息是否相同;Determine whether the recognized text information is the same as the standard text information;
    若是,则确定所述唇形识别通过。If yes, it is determined that the lip shape recognition passes.
  7. 根据权利要求6所述的方法,其特征在于,The method according to claim 6, wherein:
    所述显示标准文字信息,包括:The display standard text information includes:
    从数据库中的多个文字信息中随机选择一个文字信息作为标准文字信息,并显示所述标准文字信息。One text information is randomly selected from a plurality of text information in the database as the standard text information, and the standard text information is displayed.
  8. 根据权利要求2、4、6任一项所述的方法,其特征在于,The method according to any one of claims 2, 4, 6, wherein:
    所述对所述多个第二待检测图像进行唇形识别,以得到识别文字信息,包括:The performing lip recognition on the plurality of second to-be-detected images to obtain recognized text information includes:
    提取所述多个第二待检测图像的人脸信息;Extracting face information of the plurality of second images to be detected;
    从多个所述人脸信息中提取多个连续变化的所述唇形特征;Extracting a plurality of continuously changing lip features from a plurality of facial information;
    基于多个连续变化的所述唇形特征,得到识别文字信息。Based on the plurality of continuously changing lip features, the recognized text information is obtained.
  9. 根据权利要求8所述的方法,其特征在于,The method according to claim 8, wherein:
    所述基于多个连续变化的所述唇形特征,得到识别文字信息,包括:The obtaining the recognized text information based on the plurality of continuously changing lip features includes:
    将多个连续变化的所述唇形特征输入至唇形识别模型,以使所述唇形识别模型出对应的发音信息,并基于所述发音信息,计算出对应的识别文字信息。A plurality of continuously changing lip shape features are input to a lip shape recognition model, so that the lip shape recognition model can obtain corresponding pronunciation information, and based on the pronunciation information, corresponding recognized text information is calculated.
  10. 根据权利要求1所述的方法,其特征在于,The method according to claim 1, wherein:
    所述移动终端在获取到设定操作指令时,采集第一待检测图像,并对所述第一待检测图像进行人脸识别,包括:When the mobile terminal acquires the setting operation instruction, collecting the first image to be detected and performing face recognition on the first image to be detected includes:
    移动终端在获取到设定操作指令时,采集第一待检测图像;When the mobile terminal obtains the setting operation instruction, collect the first image to be detected;
    提取所述第一待检测图像中的人脸图像;Extracting a face image in the first image to be detected;
    对所述人脸图像进行人脸识别。Perform face recognition on the face image.
  11. 根据权利要求10所述的方法,其特征在于,The method of claim 10, wherein:
    所述对所述人脸图像进行人脸识别,包括:The performing face recognition on the face image includes:
    从所述人脸图像中提取人脸特征信息;Extracting facial feature information from the facial image;
    将所述人脸特征信息与预存的标准人脸特征信息进行相似度比对;Comparing the facial feature information with pre-stored standard facial feature information for similarity;
    在所述相似度比对的结果满足预设要求时,确定所述人脸识别通过。When the result of the similarity comparison meets a preset requirement, it is determined that the face recognition passes.
  12. 根据权利要求1所述的方法,其特征在于,The method according to claim 1, wherein:
    所述设定操作指令为支付操作指令;The setting operation instruction is a payment operation instruction;
    所述在所述唇形识别通过后,响应所述设定操作指令,包括:The responding to the setting operation instruction after the lip shape recognition is passed includes:
    在所述唇形识别通过后,响应所述支付操作指令,以完成相应的支付。After the lip shape recognition is passed, respond to the payment operation instruction to complete the corresponding payment.
  13. 一种终端设备,其特征在于,所述终端设备包括处理器以及与所述处理器连接的摄像头模组以及存储器;A terminal device, characterized in that the terminal device includes a processor, a camera module connected with the processor, and a memory;
    所述存储器用于存储程序数据,所述处理器用于执行所述程序数据,以实现如权利要求1-12任一项所述的方法。The memory is used to store program data, and the processor is used to execute the program data to implement the method according to any one of claims 1-12.
  14. 一种计算机存储介质,其特征在于,所述计算机存储介质用于存储程序数据,所述程序数据在被处理器执行时,用于实现如权利要求1-12任一项所述的方法。A computer storage medium, wherein the computer storage medium is used to store program data, and the program data is used to implement the method according to any one of claims 1-12 when executed by a processor.
  15. 一种终端设备,其特征在于,所述终端设备包括:A terminal device, characterized in that the terminal device includes:
    第一识别模块,用于在获取到设定操作指令时,采集第一待检测图像,并对所述第一待检测图像进行人脸识别;The first recognition module is configured to collect the first image to be detected and perform face recognition on the first image to be detected when the setting operation instruction is acquired;
    第二识别模块,用于在所述人脸识别通过后,采集连续的多个第二待检测图像,并对所述第二待检测图像进行唇形识别;The second recognition module is configured to collect a plurality of consecutive second to-be-detected images after the face recognition is passed, and perform lip-shape recognition on the second to-be-detected images;
    响应模块,用于在所述唇形识别通过后,响应所述设定操作指令。The response module is used to respond to the setting operation instruction after the lip shape recognition is passed.
PCT/CN2019/096579 2019-07-18 2019-07-18 Identity authentication method, terminal device, and storage medium WO2021007857A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201980010284.3A CN111684459A (en) 2019-07-18 2019-07-18 Identity authentication method, terminal equipment and storage medium
PCT/CN2019/096579 WO2021007857A1 (en) 2019-07-18 2019-07-18 Identity authentication method, terminal device, and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2019/096579 WO2021007857A1 (en) 2019-07-18 2019-07-18 Identity authentication method, terminal device, and storage medium

Publications (1)

Publication Number Publication Date
WO2021007857A1 true WO2021007857A1 (en) 2021-01-21

Family

ID=72451431

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2019/096579 WO2021007857A1 (en) 2019-07-18 2019-07-18 Identity authentication method, terminal device, and storage medium

Country Status (2)

Country Link
CN (1) CN111684459A (en)
WO (1) WO2021007857A1 (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112966550A (en) * 2021-01-28 2021-06-15 广东技术师范大学 Method and device for preventing and treating huanglongbing based on plant identification model and computer equipment
CN113486712A (en) * 2021-05-31 2021-10-08 内蒙古大学 Multi-face recognition method, system and medium based on deep learning
CN115240281A (en) * 2022-09-23 2022-10-25 平安银行股份有限公司 Private information display method and device, storage medium and mobile terminal
CN116152529A (en) * 2023-04-20 2023-05-23 吉林省信息技术研究所 Authority identification system for guaranteeing information security

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113420663B (en) * 2021-06-23 2022-02-22 深圳市海清视讯科技有限公司 Child face recognition method and system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110071830A1 (en) * 2009-09-22 2011-03-24 Hyundai Motor Company Combined lip reading and voice recognition multimodal interface system
CN102932212A (en) * 2012-10-12 2013-02-13 华南理工大学 Intelligent household control system based on multichannel interaction manner
CN104376250A (en) * 2014-12-03 2015-02-25 优化科技(苏州)有限公司 Real person living body identity verification method based on sound-type image feature
US20160134785A1 (en) * 2014-11-10 2016-05-12 Echostar Technologies L.L.C. Video and audio processing based multimedia synchronization system and method of creating the same
CN106709402A (en) * 2015-11-16 2017-05-24 优化科技(苏州)有限公司 Living person identity authentication method based on voice pattern and image features

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2011072717A1 (en) * 2009-12-15 2011-06-23 Novelty Group Limited Authorizing system, comparing device, and method for authorizing a subject
CN103324918B (en) * 2013-06-25 2016-04-27 浙江中烟工业有限责任公司 The identity identifying method that a kind of recognition of face matches with lipreading recognition
JP2015169977A (en) * 2014-03-05 2015-09-28 日本電気通信システム株式会社 Personal authentication device, personal authentication method, personal authentication program, and automatic transaction system
CN104361276B (en) * 2014-11-18 2017-07-18 新开普电子股份有限公司 A kind of multi-modal biological characteristic identity identifying method and system
CN107404381A (en) * 2016-05-19 2017-11-28 阿里巴巴集团控股有限公司 A kind of identity identifying method and device
CN107483394A (en) * 2017-05-26 2017-12-15 上海斐讯数据通信技术有限公司 A kind of router identification verification method and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110071830A1 (en) * 2009-09-22 2011-03-24 Hyundai Motor Company Combined lip reading and voice recognition multimodal interface system
CN102932212A (en) * 2012-10-12 2013-02-13 华南理工大学 Intelligent household control system based on multichannel interaction manner
US20160134785A1 (en) * 2014-11-10 2016-05-12 Echostar Technologies L.L.C. Video and audio processing based multimedia synchronization system and method of creating the same
CN104376250A (en) * 2014-12-03 2015-02-25 优化科技(苏州)有限公司 Real person living body identity verification method based on sound-type image feature
CN106709402A (en) * 2015-11-16 2017-05-24 优化科技(苏州)有限公司 Living person identity authentication method based on voice pattern and image features

Cited By (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112966550A (en) * 2021-01-28 2021-06-15 广东技术师范大学 Method and device for preventing and treating huanglongbing based on plant identification model and computer equipment
CN112966550B (en) * 2021-01-28 2022-03-11 广东技术师范大学 Method and device for preventing and treating huanglongbing based on plant identification model and computer equipment
CN113486712A (en) * 2021-05-31 2021-10-08 内蒙古大学 Multi-face recognition method, system and medium based on deep learning
CN113486712B (en) * 2021-05-31 2022-09-06 内蒙古大学 Multi-face recognition method, system and medium based on deep learning
CN115240281A (en) * 2022-09-23 2022-10-25 平安银行股份有限公司 Private information display method and device, storage medium and mobile terminal
CN116152529A (en) * 2023-04-20 2023-05-23 吉林省信息技术研究所 Authority identification system for guaranteeing information security

Also Published As

Publication number Publication date
CN111684459A (en) 2020-09-18

Similar Documents

Publication Publication Date Title
WO2021007857A1 (en) Identity authentication method, terminal device, and storage medium
US11843901B2 (en) Secure nonscheduled video visitation system
Yang et al. Preventing deepfake attacks on speaker authentication by dynamic lip movement analysis
US10275672B2 (en) Method and apparatus for authenticating liveness face, and computer program product thereof
CN108446601B (en) Face recognition method based on dynamic and static feature fusion
CN106850648B (en) Identity verification method, client and service platform
CN112997186A (en) Detection system for' viability
Raj et al. Face recognition based smart attendance system
US20130226587A1 (en) Lip-password Based Speaker Verification System
WO2021135685A1 (en) Identity authentication method and device
CN104361276A (en) Multi-mode biometric authentication method and multi-mode biometric authentication system
KR20010039771A (en) Methods and apparatus for audio-visual speaker recognition and utterance verification
WO2021007856A1 (en) Identity verification method, terminal device, and storage medium
Liu et al. Learning multi-boosted HMMs for lip-password based speaker verification
CN111104852B (en) Face recognition technology based on heuristic Gaussian cloud transformation
CN104965589A (en) Human living body detection method and device based on human brain intelligence and man-machine interaction
CN112507311A (en) High-security identity verification method based on multi-mode feature fusion
Huang et al. CALLip: Lipreading using contrastive and attribute learning
Syambas et al. Image processing and face detection analysis on face verification based on the age stages
Uzan et al. I know that voice: Identifying the voice actor behind the voice
CN116883900A (en) Video authenticity identification method and system based on multidimensional biological characteristics
CN103700151A (en) Morning run check-in method
Chen et al. An Identity Authentication Method Based on Multi-modal Feature Fusion
Zhang et al. Single Biometric Recognition Research: A Summary
Wright Lip-based biometric authentication

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 19937581

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 19937581

Country of ref document: EP

Kind code of ref document: A1