CN114329399A - Face video verification method, device, equipment and storage medium - Google Patents

Face video verification method, device, equipment and storage medium Download PDF

Info

Publication number
CN114329399A
CN114329399A CN202111574536.0A CN202111574536A CN114329399A CN 114329399 A CN114329399 A CN 114329399A CN 202111574536 A CN202111574536 A CN 202111574536A CN 114329399 A CN114329399 A CN 114329399A
Authority
CN
China
Prior art keywords
recognized
picture
face video
video
face
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111574536.0A
Other languages
Chinese (zh)
Inventor
刘冠廷
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Dajia Internet Information Technology Co Ltd
Original Assignee
Beijing Dajia Internet Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Dajia Internet Information Technology Co Ltd filed Critical Beijing Dajia Internet Information Technology Co Ltd
Priority to CN202111574536.0A priority Critical patent/CN114329399A/en
Publication of CN114329399A publication Critical patent/CN114329399A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Image Analysis (AREA)

Abstract

The disclosure relates to a method, a device, equipment and a storage medium for checking a face video, and relates to the technical field of internet. The method comprises the following steps: receiving a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, which are sent by a terminal; based on the generated information, performing frame extraction processing on the face video to be recognized to obtain a target picture; and when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold value, determining that the face video to be recognized is the face video corresponding to the target account. In the disclosure, the server can improve the checking accuracy of the face video. Furthermore, when the face video to be recognized is the face video corresponding to the target account, the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.

Description

Face video verification method, device, equipment and storage medium
Technical Field
The present disclosure relates to the field of internet technologies, and in particular, to a method, an apparatus, a device, and a storage medium for checking a face video.
Background
At present, a terminal can send a face video of an account and a face picture of the account to a server, so that the server can perform face verification on the account, specifically, verify the validity of the account.
However, the face image may be a tampered image, for example, the face image of the target person is processed to obtain the face image of the account. Therefore, the server cannot accurately verify the account which sends the face picture at present, and the accuracy of account verification is reduced.
Disclosure of Invention
The invention provides a verification method, a verification device, verification equipment and a storage medium of a face video, and solves the technical problems that in the prior art, a server cannot accurately verify an account which sends a face picture at present, and the accuracy of account verification is reduced.
The technical scheme of the embodiment of the disclosure is as follows:
according to a first aspect of the embodiments of the present disclosure, a method for checking a face video is provided. The method can comprise the following steps: receiving a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, wherein the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, the generation information is used for representing a target video frame corresponding to the to-be-recognized picture, and the to-be-recognized picture includes a face area corresponding to a target account; based on the generated information, performing frame extraction processing on the face video to be recognized to obtain a target picture; and when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold value, determining that the face video to be recognized is the face video corresponding to the target account.
Optionally, the to-be-recognized face video including the to-be-recognized picture is obtained by adding, by the terminal, target auxiliary information to the to-be-recognized face video, where the target auxiliary information is one of a plurality of auxiliary information corresponding to the to-be-recognized face video, and the method for checking the face video further includes: and when the target auxiliary information does not exist in the face video to be recognized comprising the picture to be recognized, determining that the face video to be recognized is not the face video corresponding to the target account.
Optionally, the method for checking a face video further includes: determining whether the target video frame belongs to the plurality of video frames; and when the target video frame does not belong to the plurality of video frames, determining that the generation information corresponding to the picture to be identified is abnormal.
Optionally, the method for checking a face video further includes: and when the similarity between the target picture and the picture to be recognized is smaller than the similarity threshold value, determining that the face video to be recognized is not the face video corresponding to the target account.
According to a second aspect of the embodiments of the present disclosure, a method for checking a face video is provided. The method can comprise the following steps: acquiring a face video to be recognized; determining generation information corresponding to a picture to be recognized, wherein the picture to be recognized is one of a plurality of pictures, the plurality of pictures are pictures in a plurality of video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the picture to be recognized, and the picture to be recognized comprises a face area corresponding to a target account; adding the picture to be recognized into the face video to be recognized to obtain the face video to be recognized comprising the picture to be recognized; and sending the to-be-recognized face video including the to-be-recognized picture and the generation information to a server.
Optionally, the adding the picture to be recognized to the face video to be recognized to obtain the face video to be recognized including the picture to be recognized specifically includes: adding the picture to be recognized into target auxiliary information, wherein the target auxiliary information is one of a plurality of auxiliary information corresponding to the face video to be recognized; and adding the target auxiliary information into the face video to be recognized to obtain the face video to be recognized comprising the picture to be recognized.
According to a third aspect of the embodiments of the present disclosure, a verification apparatus for a face video is provided. The apparatus may include a receiving module, a processing module, and a determining module; the receiving module is configured to receive a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, wherein the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, the generation information is used for representing a target video frame corresponding to the to-be-recognized picture, and the to-be-recognized picture includes a face area corresponding to a target account; the processing module is configured to perform frame extraction processing on the face video to be recognized based on the generation information to obtain a target picture; the determining module is configured to determine that the face video to be recognized is the face video corresponding to the target account when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold.
Optionally, the to-be-recognized face video including the to-be-recognized picture is obtained by adding target auxiliary information to the to-be-recognized face video by the terminal, where the target auxiliary information is one of a plurality of auxiliary information corresponding to the to-be-recognized face video.
The determining module is further configured to determine that the face video to be recognized is not the face video corresponding to the target account when the target auxiliary information does not exist in the face video to be recognized including the picture to be recognized.
Optionally, the determining module is further configured to determine whether the target video frame belongs to the plurality of video frames; the determining module is further configured to determine that the generation information corresponding to the picture to be recognized is abnormal when the target video frame does not belong to the plurality of video frames.
Optionally, the determining module is further configured to determine that the face video to be recognized is not the face video corresponding to the target account when the similarity between the target picture and the picture to be recognized is smaller than the similarity threshold.
According to a fourth aspect of the embodiments of the present disclosure, a device for checking a face video is provided. The apparatus may include: the device comprises an acquisition module, a determination module, a processing module and a sending module; the acquisition module is configured to acquire a face video to be recognized; the determining module is configured to determine generation information corresponding to a picture to be recognized, wherein the picture to be recognized is one of a plurality of pictures, the plurality of pictures are pictures in a plurality of video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the picture to be recognized, and the picture to be recognized comprises a face area corresponding to a target account; the processing module is configured to add the picture to be recognized to the face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized; the sending module is configured to send the to-be-recognized face video including the to-be-recognized picture and the generation information to a server.
Optionally, the processing module is specifically configured to add the picture to be recognized to target auxiliary information, where the target auxiliary information is one of a plurality of auxiliary information corresponding to the face video to be recognized; the processing module is specifically configured to add the target auxiliary information to a face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized.
According to a fifth aspect of embodiments of the present disclosure, there is provided a server, which may include: a processor and a memory configured to store processor-executable instructions; wherein the processor is configured to execute the instructions to implement any one of the above-mentioned first aspect optionally human face video verification methods.
According to a sixth aspect of the embodiments of the present disclosure, there is provided a terminal, which may include: a processor and a memory configured to store processor-executable instructions; wherein the processor is configured to execute the instructions to implement any one of the above-mentioned second aspect optionally human face video verification methods.
According to a seventh aspect of the embodiments of the present disclosure, there is provided a computer-readable storage medium, on which instructions are stored, and when the instructions in the computer-readable storage medium are executed by an apparatus, the apparatus is enabled to execute any one of the above-mentioned first aspect optional face video verification methods, or execute any one of the above-mentioned second aspect optional face video verification methods.
According to an eighth aspect of embodiments of the present disclosure, there is provided a computer program product comprising computer instructions which, when run on a device, cause the device to perform the method of checking optionally face video according to any of the first aspects, or to perform the method of checking optionally face video according to any of the second aspects.
The technical scheme provided by the embodiment of the disclosure at least brings the following beneficial effects:
based on any one of the above aspects, in the present disclosure, the server may receive the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture, which are sent by the terminal, and perform frame extraction processing on the to-be-recognized face video based on the generation information to obtain the target picture. Because the generation information corresponding to the picture to be recognized is used for representing the target video frame corresponding to the picture to be recognized, the server performs frame extraction processing on the face video to be recognized so as to obtain a picture (namely a target picture) corresponding to the video frame which is the same as the target video frame from a plurality of video frames corresponding to the face video to be recognized, and determines the similarity between the target picture and the picture to be recognized. When the similarity is greater than or equal to the similarity threshold, it is indicated that the similarity between the target picture and the picture to be recognized is large, and the picture to be recognized is free from the risk of being tampered. Therefore, the server can determine that the face video to be recognized is the face video corresponding to the target account, and the verification accuracy of the face video can be improved. Furthermore, when the face video to be recognized is the face video corresponding to the target account, it is indicated that the current account (specifically, the account for instructing the terminal to send the face video to be recognized to the server) is not a real account (i.e., the target account), so that the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.
It is to be understood that both the foregoing general description and the following detailed description are exemplary and explanatory only and are not restrictive of the disclosure.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the present disclosure and, together with the description, serve to explain the principles of the disclosure and are not to be construed as limiting the disclosure.
Fig. 1 is a schematic diagram illustrating a system for checking a face video according to an embodiment of the present disclosure;
fig. 2 is a schematic flowchart illustrating a method for verifying a face video according to an embodiment of the present disclosure;
fig. 3 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 4 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 5 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 6 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 7 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 8 is a schematic flowchart illustrating a further method for verifying a face video according to an embodiment of the present disclosure;
fig. 9 is a schematic structural diagram illustrating a verification apparatus for a face video according to an embodiment of the present disclosure;
fig. 10 is a schematic structural diagram illustrating a further apparatus for verifying a face video according to an embodiment of the present disclosure;
fig. 11 is a schematic structural diagram illustrating an apparatus for verifying a face video according to an embodiment of the present disclosure;
fig. 12 shows a schematic structural diagram of another apparatus for verifying a face video according to an embodiment of the present disclosure.
Detailed Description
In order to make the technical solutions of the present disclosure better understood by those of ordinary skill in the art, the technical solutions in the embodiments of the present disclosure will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present disclosure and in the above-described drawings are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the disclosure described herein are capable of operation in sequences other than those illustrated or otherwise described herein. The implementations described in the exemplary embodiments below are not intended to represent all implementations consistent with the present disclosure. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the present disclosure, as detailed in the appended claims.
It will be further understood that the terms "comprises" and/or "comprising," when used in this specification, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, and/or components.
The data to which the present disclosure relates may be data that is authorized by a user or sufficiently authorized by parties.
As described in the background art, since the face picture sent by the terminal to the server in the prior art may be a picture obtained after tampering, the server cannot accurately verify the account currently sending the face picture, and the accuracy of account verification is reduced (specifically, it may not be accurately verified whether the current account is a legal account). Based on this, the embodiment of the disclosure provides a verification method for a face video, which when the similarity is greater than or equal to the similarity threshold, indicates that the similarity between the target picture and the picture to be recognized is relatively large, and the picture to be recognized has no risk of being tampered. Therefore, the server can determine that the face video to be recognized is the face video corresponding to the target account, and the verification accuracy of the face video can be improved. Furthermore, when the face video to be recognized is the face video corresponding to the target account, it is indicated that the current account (specifically, the account for instructing the terminal to send the face video to be recognized to the server) is not a real account (i.e., the target account), so that the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.
The verification method, the verification device, the verification equipment and the verification storage medium of the face video, which are provided by the embodiment of the disclosure, are applied to a scene of verifying the legality of a certain account. When the server receives the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture sent by the terminal, whether the to-be-recognized face video is the face video corresponding to the target account or not can be determined according to the method provided by the embodiment of the disclosure, and whether the account of the to-be-recognized face video is legal or not can be further determined.
The verification method of the face video provided by the embodiment of the present disclosure is exemplarily described below with reference to the accompanying drawings:
fig. 1 is a schematic view of a face video verification system provided in an embodiment of the present disclosure, as shown in fig. 1, the face video verification system may include a terminal 101 and a server 102, and the terminal 101 may establish a connection with the server 102 through a wired network or a wireless network.
The terminal 101 may be a mobile phone, a tablet computer, a desktop, a laptop, a handheld computer, a notebook, an ultra-mobile personal computer (UMPC), a netbook, a cellular phone, a Personal Digital Assistant (PDA), an Augmented Reality (AR) device, a Virtual Reality (VR) device, or other devices that can be installed and used for content community applications, and the disclosure does not particularly limit the specific form of the terminal. The system can be used for man-machine interaction with a user through one or more modes of a keyboard, a touch pad, a touch screen, a remote controller, voice interaction or handwriting equipment and the like.
Specifically, the terminal 101 may obtain a face video to be recognized. In the embodiment of the present disclosure, the terminal 101 may further send the target file and the generation information corresponding to the picture to be recognized to the server 102.
The server 102 may receive the target file sent by the terminal 101 and the generation information corresponding to the picture to be recognized.
As shown in fig. 2, when the verification method of the face video is applied to the terminal 101, the method may include S101-S104.
S101, the terminal obtains a face video to be recognized.
It should be understood that the face video to be recognized is a video including a face region of the target account. The terminal acquires the face video to be recognized, and can also be understood as collecting (or recording) the face video to be recognized.
In an implementation manner of the embodiment of the present disclosure, the face video to be recognized is used for representing a process of live face detection of a target account. Specifically, when the terminal needs to perform face live detection on the target account, some verification instructions (for example, "open mouth", "blink", and the like) are issued, and the terminal records the face live detection process, so that the face video to be recognized can be obtained.
S102, the terminal determines the generation information corresponding to the picture to be identified.
The image to be recognized is one of a plurality of images, the plurality of images are images in a plurality of video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the image to be recognized, and the image to be recognized comprises a face area corresponding to a target account.
It can be understood that the terminal may perform frame extraction processing on the face video to be recognized to obtain each of the multiple pictures (or may also be understood as obtaining a picture of each of the multiple video frames). And the video frame corresponding to each picture may be recorded (which may also be understood as determining that each picture corresponds to the several frames in the plurality of video frames), that is, the generation information corresponding to each picture in the plurality of pictures may be obtained.
It should be noted that the picture to be recognized is a picture with higher definition in the plurality of pictures.
S103, adding the picture to be recognized to the face video to be recognized by the terminal to obtain the face video to be recognized comprising the picture to be recognized.
It should be understood that after the terminal obtains the face video to be recognized including the picture to be recognized, the face video to be recognized including the picture to be recognized may be sent to the server.
It should be noted that the disclosed embodiment does not limit the execution sequence of S102 and S103. For example, S102 may be performed first and then S103, or S103 may be performed first and then S102, or S102 and S103 may be performed simultaneously. For convenience of example, in fig. 2, S102 is executed first, and then S103 is executed.
And S104, the terminal sends the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture to the server.
It should be understood that the terminal sends the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture to the server, that is, sends the to-be-recognized picture, the to-be-recognized face video and the generation information corresponding to the to-be-recognized picture to the server. So that the server can acquire the three items of content, specifically, the picture to be recognized is included in the face video to be recognized.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: S101-S104 show that the terminal can acquire a face video to be recognized and determine the corresponding generation information of the picture to be recognized; and then the terminal adds the picture to be recognized to the face video to be recognized to obtain the face video to be recognized comprising the picture to be recognized, and sends the face video to be recognized comprising the picture to be recognized and the generation information to the server. Therefore, the terminal can ensure that the server can acquire the three contents of the face video to be recognized, the picture to be recognized and the generation information of the picture to be recognized.
Furthermore, the generation information is used for representing a target video frame corresponding to the picture to be recognized, the server can perform frame extraction processing on the face video to be recognized to obtain a target picture based on the generation information after receiving the face video to be recognized including the picture to be recognized, the similarity between the target picture and the picture to be recognized is determined, and then whether the face video to be recognized is the face video corresponding to the target account is determined based on the similarity, so that the verification accuracy of the face video can be improved.
With reference to fig. 2, as shown in fig. 3, in an implementation manner of the embodiment of the present disclosure, the adding the picture to be recognized to the face video to be recognized to obtain the face video to be recognized including the picture to be recognized includes S1031 to S1032.
And S1031, adding the picture to be identified to the target auxiliary information by the terminal.
The target auxiliary information is one of a plurality of auxiliary information corresponding to the face video to be recognized.
It is understood that for a video (e.g., a video of a face to be recognized), the video may correspond to a plurality of auxiliary information. Specifically, one piece of attached information may represent the author of the video, the album to which the video belongs, or the cover of the video, etc. In the embodiment of the present disclosure, the target auxiliary information may represent a cover of the face video to be recognized, that is, the terminal may add the picture to be recognized to the cover of the face video to be recognized.
S1032, the terminal adds the target auxiliary information to the face video to be recognized to obtain the face video to be recognized including the picture to be recognized.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: from S1031 to S1032, the terminal may add the picture to be recognized to the target auxiliary information (i.e., one of the plurality of auxiliary information corresponding to the face video to be recognized), and then add the target auxiliary information to the face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized. In the embodiment of the disclosure, the terminal can add the picture to be recognized to a certain auxiliary information corresponding to the face video to be recognized and add the auxiliary information to the face video to be recognized, so that the face video to be recognized including the picture to be recognized can be conveniently and quickly obtained, and the sending efficiency of the face video to be recognized can be further improved.
As shown in fig. 4, when the verification method of the face video is applied to the server 102, the verification method of the face video may include S201 to S203.
S201, a server receives a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, which are sent by a terminal.
With reference to the description of the above embodiment, it should be understood that the picture to be recognized is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the face video to be recognized, the generation information is used to represent a target video frame corresponding to the picture to be recognized, and the picture to be recognized includes a face area corresponding to a target account.
S202, the server performs frame extraction processing on the face video to be recognized based on the generation information corresponding to the picture to be recognized to obtain a target picture.
It should be understood that the server performs frame extraction processing on the face video to be recognized to obtain a plurality of pictures (which may also be understood as pictures in a plurality of video frames). And the server may also record (or determine) generation information corresponding to each picture (or each frame of pictures) in the plurality of pictures. In the embodiment of the disclosure, the server obtains the target picture based on the generation information corresponding to the picture to be recognized, that is, the picture corresponding to the generation information that is the same as the generation information corresponding to the picture to be recognized in the plurality of pictures obtained by the server is determined as the target picture.
For example, assuming that a video frame represented by generation information corresponding to the picture to be recognized is the 10 th frame of the plurality of video frames, after performing frame extraction processing on the face video to be recognized, the server determines the picture in the 10 th frame as the target picture.
S203, when the similarity between the target picture and the picture to be recognized is larger than or equal to the similarity threshold, the server determines that the face video to be recognized is the face video corresponding to the target account.
It should be understood that the server may determine the similarity between the target picture and the picture to be recognized. When the similarity between the target picture and the picture to be recognized is greater than or equal to the similarity threshold, it is indicated that the similarity between the target picture and the picture to be recognized is large, and the picture to be recognized is free from the risk of being tampered. Thus, the server may determine that the face video to be recognized is the face video corresponding to the target account (it may also be understood that the face video to be recognized including the picture to be recognized has no abnormality). Furthermore, the server may determine that the current account (specifically, the account indicating the terminal to send the face video to be recognized to the server) is a real account (i.e., a target account), and issue a corresponding access or operation right to the terminal (or the current account).
The technical scheme provided by the embodiment can at least bring the following beneficial effects: as can be seen from S201 to S203, the server may receive the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture, which are sent by the terminal, and perform frame extraction processing on the to-be-recognized face video based on the generation information to obtain the target picture. Because the generation information corresponding to the picture to be recognized is used for representing the target video frame corresponding to the picture to be recognized, the server performs frame extraction processing on the face video to be recognized so as to obtain a picture (namely a target picture) corresponding to the video frame which is the same as the target video frame from a plurality of video frames corresponding to the face video to be recognized, and determines the similarity between the target picture and the picture to be recognized. When the similarity is greater than or equal to the similarity threshold, it is indicated that the similarity between the target picture and the picture to be recognized is large, and the picture to be recognized is free from the risk of being tampered. Therefore, the server can determine that the face video to be recognized is the face video corresponding to the target account, and the verification accuracy of the face video can be improved. Furthermore, when the face video to be recognized is the face video corresponding to the target account, it is indicated that the current account (specifically, the account for instructing the terminal to send the face video to be recognized to the server) is not a real account (i.e., the target account), so that the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.
With reference to fig. 4 and as shown in fig. 5, in an implementation manner of the embodiment of the present disclosure, the to-be-recognized face video including the to-be-recognized picture is obtained by adding, by the terminal, target attached information to the to-be-recognized face video, where the target attached information is one of a plurality of attached information corresponding to the to-be-recognized face video. The verification method for the face video provided by the embodiment of the present disclosure further includes S204.
S204, when the target auxiliary information does not exist in the face video to be recognized including the picture to be recognized, the server determines that the face video to be recognized is not the face video corresponding to the target account.
It can be understood that the face video to be recognized sent by the terminal to the server includes the target accessory information (specifically, the picture to be recognized) and the face video to be recognized. When the target auxiliary information does not exist in the to-be-recognized face video including the to-be-recognized picture sent by the server, it is indicated that the to-be-recognized face video is incomplete or the to-be-recognized face video may have been tampered, so that the server can determine that the to-be-recognized face video is not the face video corresponding to the target account, and can also understand that the to-be-recognized face video is abnormal.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: as can be seen from S204, when the target auxiliary information does not exist in the to-be-recognized face video including the to-be-recognized picture, it is indicated that the to-be-recognized face video is incomplete or the to-be-recognized face video may have been tampered, so that the server may determine that the to-be-recognized face video is not a face video corresponding to the target account. The verification of the account validity can be carried out on the basis of the complete face video to be recognized by the server, and the verification efficiency of the face video is improved.
With reference to fig. 4, as shown in fig. 6, the verification method for a face video according to the embodiment of the present disclosure further includes S205-S206.
S205, the server determines whether the target video frame belongs to a plurality of video frames corresponding to the face video to be recognized.
It should be understood that, after receiving a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture sent by a terminal, a server may determine whether a video frame represented by the generation information (i.e., a target video frame corresponding to the to-be-recognized picture) belongs to the plurality of video frames (i.e., video frames corresponding to the to-be-recognized face video).
S206, when the target video frame does not belong to the plurality of video frames, the server determines that the generation information corresponding to the picture to be identified is abnormal.
It should be understood that, when the target video frame does not belong to the plurality of video frames, it is indicated that the generation information is not generation information corresponding to each of a plurality of pictures (specifically, pictures in the plurality of video frames) corresponding to the face video to be recognized, and at this time, the server may determine that the generation information corresponding to the picture to be recognized is abnormal.
For example, assuming that the video frame represented by the generation information is the 30 th frame, and the above-mentioned face video to be recognized corresponds to 25 video frames, the server determines that the target video frame does not belong to the plurality of video frames.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: as known from S205-S206, the server may determine whether the target video frame belongs to a plurality of video frames corresponding to the face video to be recognized; when the target video frame does not belong to the plurality of video frames, it is indicated that the generation information corresponding to the picture to be recognized is not the generation information corresponding to the picture (i.e. the plurality of pictures corresponding to the face video to be recognized) in the plurality of video frames corresponding to the face video to be recognized, so that the server can determine that the generation information corresponding to the picture to be recognized is abnormal, can accurately determine whether the generation information corresponding to a certain picture received by the server is abnormal, and further improves the verification accuracy of the face video.
With reference to fig. 4, as shown in fig. 7, the method for checking a face video according to the embodiment of the present disclosure further includes S207.
And S207, when the similarity between the target picture and the picture to be recognized is smaller than a similarity threshold value, the server determines that the face video to be recognized is not the face video corresponding to the target account.
It should be understood that when the similarity between the target picture and the picture to be recognized is smaller than the similarity threshold, it indicates that the similarity between the target picture and the picture to be recognized is smaller. In the embodiment of the disclosure, when the similarity between the target picture and the picture to be recognized is small, it is indicated that the picture to be recognized is at risk of being tampered, and the server may determine that the face video to be recognized is not the face video corresponding to the target account.
Optionally, when the face video to be recognized is not the face video corresponding to the target account, the server may determine that the face video to be recognized including the picture to be recognized is abnormal, specifically, the picture to be recognized is abnormal.
It can be understood that, when the to-be-recognized face video including the to-be-recognized picture is abnormal (specifically, the to-be-recognized picture is abnormal), the server may determine that the current account is not a real account, so that the server may effectively verify that the current account has a behavior of using the real account abnormally, and may improve accuracy of account verification.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: s207 may determine that, when the similarity between the target picture and the picture to be recognized is smaller than the similarity threshold, it indicates that the similarity between the target picture and the picture to be recognized is smaller, and the picture to be recognized is at risk of being tampered. Therefore, the server can determine that the face video to be recognized is not the face video corresponding to the target account, and the verification accuracy of the face video can be improved. Furthermore, when the face video to be recognized is not the face video corresponding to the target account, it is indicated that the current account is not the real account, so that the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.
As shown in fig. 8, when the verification method of the face video is based on the above-mentioned interaction process between the terminal 101 and the server 102, the verification method of the face video may include S301 to S307.
S301, the terminal obtains a face video to be recognized.
S302, the terminal determines the generation information corresponding to the picture to be identified.
The image to be recognized is one of a plurality of images, the plurality of images are images in a plurality of video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the image to be recognized, and the image to be recognized comprises a face area corresponding to a target account.
And S303, adding the picture to be recognized into the face video to be recognized by the terminal to obtain the face video to be recognized comprising the picture to be recognized.
S304, the terminal sends the face video to be recognized including the picture to be recognized and the generation information corresponding to the picture to be recognized to the server.
S305, the server receives the to-be-recognized face video including the to-be-recognized picture and the generation information corresponding to the to-be-recognized picture, which are sent by the terminal.
S306, the server performs frame extraction processing on the face video to be recognized based on the generation information corresponding to the picture to be recognized to obtain a target picture.
S307, when the similarity between the target picture and the picture to be recognized is larger than or equal to the similarity threshold, the server determines that the face video to be recognized is the face video corresponding to the target account.
It should be understood that the explanation of S301-S307 can refer to the description in the above embodiments, and the description is omitted here.
The technical scheme provided by the embodiment can at least bring the following beneficial effects: S301-S307 show that the terminal can acquire a face video to be recognized and determine the generation information corresponding to the picture to be recognized; and then the terminal adds the picture to be recognized into the face video to be recognized to obtain the face video to be recognized comprising the picture to be recognized, and sends the face video to be recognized comprising the picture to be recognized and the generation information to the server. Therefore, the terminal can ensure that the server can acquire the three contents of the face video to be recognized, the picture to be recognized and the generation information of the generation time of the picture to be recognized.
Further, the server may receive a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, which are sent by the terminal, and perform frame extraction processing on the to-be-recognized face video based on the generation information to obtain a target picture. Because the generation information corresponding to the picture to be recognized is used for representing the target video frame corresponding to the picture to be recognized, the server performs frame extraction processing on the face video to be recognized so as to obtain a picture (namely a target picture) corresponding to the video frame which is the same as the target video frame from a plurality of video frames corresponding to the face video to be recognized, and determines the similarity between the target picture and the picture to be recognized. When the similarity is greater than or equal to the similarity threshold, it is indicated that the similarity between the target picture and the picture to be recognized is large, and the picture to be recognized is free from the risk of being tampered. Therefore, the server can determine that the face video to be recognized is the face video corresponding to the target account, and the verification efficiency of the face video can be improved. Furthermore, when the face video to be recognized is the face video corresponding to the target account, it is indicated that the current account (specifically, the account for instructing the terminal to send the face video to be recognized to the server) is not the real account (i.e., the target account), so that the server can effectively verify whether the current account has a behavior of using the real account abnormally, and the accuracy of account verification can be improved.
It is understood that, in practical implementation, the terminal/server according to the embodiments of the present disclosure may include one or more hardware structures and/or software modules for implementing the foregoing verification method for the face video, and these hardware structures and/or software modules may constitute a terminal/server. Those of skill in the art will readily appreciate that the present disclosure can be implemented in hardware or a combination of hardware and computer software for implementing the exemplary algorithm steps described in connection with the embodiments disclosed herein. Whether a function is performed as hardware or computer software drives hardware depends upon the particular application and design constraints imposed on the solution. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.
Based on such understanding, the embodiment of the present disclosure further provides a verification apparatus for a face video correspondingly, and fig. 9 shows a schematic structural diagram of the verification apparatus for a face video provided by the embodiment of the present disclosure. As shown in fig. 9, the verification apparatus 40 for human face video may include: a receiving module 401, a processing module 402 and a determining module 403.
The receiving module 401 is configured to receive a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, where the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, the generation information is used to represent a target video frame corresponding to the to-be-recognized picture, and the to-be-recognized picture includes a face region corresponding to a target account.
And the processing module 402 is configured to perform frame extraction processing on the face video to be recognized based on the generation information to obtain a target picture.
The determining module 403 is configured to determine that the face video to be recognized is the face video corresponding to the target account when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold.
Optionally, the to-be-recognized face video including the to-be-recognized picture is obtained by adding target auxiliary information to the to-be-recognized face video by the terminal, where the target auxiliary information is one of a plurality of auxiliary information corresponding to the to-be-recognized face video.
The determining module 403 is further configured to determine that the face video to be recognized is not the face video corresponding to the target account when the target auxiliary information does not exist in the face video to be recognized including the picture to be recognized.
Optionally, the determining module 403 is further configured to determine whether the target video frame belongs to the plurality of video frames.
The determining module 403 is further configured to determine that the generation information corresponding to the picture to be recognized is abnormal when the target video frame does not belong to the plurality of video frames.
Optionally, the determining module 403 is further configured to determine that the face video to be recognized is not the face video corresponding to the target account when the similarity between the target picture and the picture to be recognized is smaller than the similarity threshold.
As described above, the embodiment of the present disclosure may perform functional module division on the verification apparatus for a face video according to the above method example. The integrated module can be realized in a hardware form, and can also be realized in a software functional module form. In addition, it should be further noted that the division of the modules in the embodiments of the present disclosure is schematic, and is only a logic function division, and there may be another division manner in actual implementation. For example, the functional blocks may be divided for the respective functions, or two or more functions may be integrated into one processing block.
Regarding the verification apparatus for a face video in the foregoing embodiment, the specific manner in which each module executes operations and the beneficial effects thereof have been described in detail in the foregoing method embodiment, and are not described herein again.
Fig. 10 is a schematic structural diagram of another verification apparatus for face video provided by the present disclosure. As shown in fig. 10, the apparatus 50 for verifying a human face video may include at least one processor 501 and a memory 503 for storing processor-executable instructions. The processor 501 is configured to execute the instructions in the memory 503 to implement the verification method of the face video in the above embodiment.
In addition, the verification device 50 for human face video may further include a communication bus 502 and at least one communication interface 504.
The processor 501 may be a Central Processing Unit (CPU), a micro-processing unit, an ASIC, or one or more integrated circuits for controlling the execution of programs according to the present disclosure.
The communication bus 502 may include a path that conveys information between the aforementioned components.
The communication interface 504 may be any device, such as a transceiver, for communicating with other devices or communication networks, such as an ethernet, a Radio Access Network (RAN), a Wireless Local Area Network (WLAN), etc.
The memory 503 may be, but is not limited to, a read-only memory (ROM) or other type of static storage device that may store static information and instructions, a Random Access Memory (RAM) or other type of dynamic storage device that may store information and instructions, an electrically erasable programmable read-only memory (EEPROM), a compact disc read-only memory (CD-ROM) or other optical disk storage, optical disk storage (including compact disc, laser disc, optical disc, digital versatile disc, blu-ray disc, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. The memory may be self-contained and connected to the processing unit by a bus. The memory may also be integrated with the processing unit.
The memory 503 is used for storing instructions for executing the disclosed solution, and is controlled by the processor 501. The processor 501 is configured to execute instructions stored in the memory 503 to implement the functions of the disclosed method.
In particular implementations, processor 501 may include one or more CPUs such as CPU0 and CPU1 in fig. 10, for example, as an example.
In one embodiment, the apparatus 50 for verifying a face video may include a plurality of processors, such as the processor 501 and the processor 507 in fig. 10. Each of these processors may be a single-core (single-CPU) processor or a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In a specific implementation, the apparatus for verifying a human face video 50 may further include an output device 505 and an input device 506. An output device 505, which is in communication with the processor 501, may display information in a variety of ways. For example, the output device 505 may be a Liquid Crystal Display (LCD), a Light Emitting Diode (LED) display device, a Cathode Ray Tube (CRT) display device, a projector (projector), or the like. The input device 506 is in communication with the processor 501 and can accept user input in a variety of ways. For example, the input device 506 may be a mouse, a keyboard, a touch screen device, or a sensing device, among others.
Fig. 11 is a diagram illustrating a structure of a face video verification apparatus according to the present disclosure. As shown in fig. 11, the verification apparatus 60 for human face video may include: an obtaining module 601, a determining module 602, a processing module 603 and a sending module 604.
The acquiring module 601 is configured to acquire a face video to be recognized.
The determining module 602 is configured to determine generation information corresponding to a to-be-recognized picture, where the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, and the generation information is used to represent a target video frame corresponding to the to-be-recognized picture, where the to-be-recognized picture includes a face region corresponding to a target account.
The processing module 603 is configured to add the picture to be recognized to the face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized.
A sending module 604 configured to send the to-be-recognized face video including the to-be-recognized picture and the generation information to a server.
Optionally, the processing module 603 is specifically configured to add the picture to be recognized to target auxiliary information, where the target auxiliary information is one of a plurality of auxiliary information corresponding to the face video to be recognized.
The processing module 603 is specifically configured to add the target auxiliary information to the face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized.
Fig. 12 is a schematic structural diagram of another verification apparatus for face video according to the present disclosure. As shown in fig. 12, the verification apparatus 70 for human face video may include at least one processor 701 and a memory 703 for storing processor executable instructions. The processor 701 is configured to execute instructions in the memory 703 to implement the verification method of the face video in the above embodiment.
In addition, the verification device 70 for human face video may further include a communication bus 702 and at least one communication interface 704.
The processor 701 may be a CPU, micro-processing unit, ASIC, or one or more integrated circuits for controlling the execution of programs in accordance with the disclosed aspects.
The communication bus 702 may include a path that conveys information between the aforementioned components.
The communication interface 704 may be any device, such as a transceiver, for communicating with other devices or a communication network, such as an ethernet, RAN, WLAN, etc.
The memory 703 may be, but is not limited to, ROM or other type of static storage device that can store static information and instructions, RAM or other type of dynamic storage device that can store information and instructions, EEPROM, CD-ROM or other optical disk storage, optical disk storage (including compact disk, laser disk, optical disk, digital versatile disk, blu-ray disk, etc.), magnetic disk storage media or other magnetic storage devices, or any other medium that can be used to carry or store desired program code in the form of instructions or data structures and that can be accessed by a computer. The memory may be self-contained and connected to the processing unit by a bus. The memory may also be integrated with the processing unit.
The memory 703 is used for storing instructions for executing the disclosed solution, and is controlled by the processor 701. The processor 701 is configured to execute instructions stored in the memory 703 to implement the functions of the disclosed method.
In particular implementations, processor 701 may include one or more CPUs such as CPU0 and CPU1 in fig. 12 for one embodiment.
In one embodiment, the apparatus for verifying a face video 70 may include a plurality of processors, such as the processor 701 and the processor 707 in fig. 12. Each of these processors may be a single-core (single-CPU) processor or a multi-core (multi-CPU) processor. A processor herein may refer to one or more devices, circuits, and/or processing cores for processing data (e.g., computer program instructions).
In a specific implementation, as an embodiment, the apparatus for verifying a face video 70 may further include an output device 705 and an input device 706. An output device 705 is in communication with the processor 701 and may display information in a variety of ways. For example, the output device 705 may be an LCD, LED display device, CRT display device, projector, or the like. The input device 706 communicates with the processor 701 and may accept input from a user in a variety of ways. For example, the input device 706 may be a mouse, a keyboard, a touch screen device, or a sensing device, among others.
It will be appreciated by those skilled in the art that the configurations shown in fig. 10 or 12 above do not constitute a limitation of the verification means for the face video, and may include more or fewer components than those shown, or some components may be combined, or a different arrangement of components may be employed.
In addition, the present disclosure also provides a computer-readable storage medium, which includes instructions, when executed by a device, cause the device to execute the verification method of the face video provided in the above embodiment.
In addition, the present disclosure also provides a computer program product including instructions, which when executed by a device, cause the device to execute the verification method of the face video provided in the above embodiment.
Other embodiments of the disclosure will be apparent to those skilled in the art from consideration of the specification and practice of the disclosure disclosed herein. This disclosure is intended to cover any variations, uses, or adaptations of the disclosure following, in general, the principles of the disclosure and including such departures from the present disclosure as come within known or customary practice within the art to which the disclosure pertains. It is intended that the specification and examples be considered as exemplary only, with a true scope and spirit of the disclosure being indicated by the following claims.

Claims (10)

1. A verification method of a face video is applied to a server and is characterized by comprising the following steps:
receiving a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, wherein the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, the generation information is used for representing a target video frame corresponding to the to-be-recognized picture, and the to-be-recognized picture includes a face area corresponding to a target account;
based on the generated information, performing frame extraction processing on the face video to be recognized to obtain a target picture;
and when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold value, determining that the face video to be recognized is the face video corresponding to the target account.
2. The method for verifying the face video according to claim 1, wherein the face video to be recognized including the picture to be recognized is obtained by adding target auxiliary information to the face video to be recognized by the terminal, and the target auxiliary information is one of a plurality of auxiliary information corresponding to the face video to be recognized, and the method further comprises:
when the target auxiliary information does not exist in the face video to be recognized comprising the picture to be recognized, determining that the face video to be recognized is not the face video corresponding to the target account.
3. A face video verification method is applied to a terminal and is characterized by comprising the following steps:
acquiring a face video to be recognized;
determining generation information corresponding to a picture to be recognized, wherein the picture to be recognized is one of a plurality of pictures, the plurality of pictures are pictures in a plurality of video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the picture to be recognized, and the picture to be recognized comprises a face area corresponding to a target account;
adding the picture to be recognized into the face video to be recognized to obtain the face video to be recognized comprising the picture to be recognized;
and sending the face video to be recognized including the picture to be recognized and the generated information to a server.
4. A checking device for face video is characterized by comprising: the device comprises a receiving module, a processing module and a determining module;
the receiving module is configured to receive a to-be-recognized face video including a to-be-recognized picture and generation information corresponding to the to-be-recognized picture, wherein the to-be-recognized picture is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the to-be-recognized face video, the generation information is used for representing a target video frame corresponding to the to-be-recognized picture, and the to-be-recognized picture includes a face area corresponding to a target account;
the processing module is configured to perform frame extraction processing on the face video to be recognized based on the generation information to obtain a target picture;
the determining module is configured to determine that the face video to be recognized is the face video corresponding to the target account when the similarity between the target picture and the picture to be recognized is greater than or equal to a similarity threshold value.
5. A checking device for face video is characterized by comprising: the device comprises an acquisition module, a determination module, a processing module and a sending module;
the acquisition module is configured to acquire a face video to be recognized;
the determining module is configured to determine generation information corresponding to a picture to be recognized, where the picture to be recognized is one of multiple pictures, the multiple pictures are pictures in multiple video frames corresponding to the face video to be recognized, the generation information is used for representing a target video frame corresponding to the picture to be recognized, and the picture to be recognized includes a face region corresponding to a target account;
the processing module is configured to add the picture to be recognized to the face video to be recognized, so as to obtain the face video to be recognized including the picture to be recognized;
the sending module is configured to send the to-be-recognized face video including the to-be-recognized picture and the generation information to a server.
6. The system for checking the face video is characterized by comprising a server and a terminal;
the server is used for executing the verification method of the face video in any one of the claims 1-4;
the terminal is used for executing the verification method of the face video in the claim 5 or 6.
7. A server, characterized in that the server comprises:
a processor;
a memory configured to store the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the verification method of the face video according to claim 1 or 2.
8. A terminal, characterized in that the terminal comprises:
a processor;
a memory configured to store the processor-executable instructions;
wherein the processor is configured to execute the instructions to implement the verification method of the face video according to claim 3.
9. A computer-readable storage medium having instructions stored thereon, wherein the instructions in the computer-readable storage medium, when executed by a device, enable the device to perform the verification method of the face video according to claim 1 or 2 or perform the verification method of the face video according to claim 3.
10. A computer program product, characterized in that it comprises computer instructions which, when run on a device, cause the device to carry out the method of verification of face video according to claim 1 or 2, or to carry out the method of verification of face video according to claim 3.
CN202111574536.0A 2021-12-21 2021-12-21 Face video verification method, device, equipment and storage medium Pending CN114329399A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111574536.0A CN114329399A (en) 2021-12-21 2021-12-21 Face video verification method, device, equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111574536.0A CN114329399A (en) 2021-12-21 2021-12-21 Face video verification method, device, equipment and storage medium

Publications (1)

Publication Number Publication Date
CN114329399A true CN114329399A (en) 2022-04-12

Family

ID=81054227

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111574536.0A Pending CN114329399A (en) 2021-12-21 2021-12-21 Face video verification method, device, equipment and storage medium

Country Status (1)

Country Link
CN (1) CN114329399A (en)

Similar Documents

Publication Publication Date Title
CN107273280B (en) Log processing method and device, electronic equipment and storage medium
CN110505141B (en) Instant messaging message processing method and device, readable medium and electronic equipment
CN110059623B (en) Method and apparatus for generating information
CN111026493B (en) Interface rendering processing method and device
WO2023103390A1 (en) Task processing method, task processing apparatus, electronic device and storage medium
US20220414971A1 (en) Method of processing information, method of rendering image, and electronic device
CN110222775A (en) Image processing method, device, electronic equipment and computer readable storage medium
WO2019019356A1 (en) Application program test method and apparatus, computer device and storage medium
CN115205925A (en) Expression coefficient determining method and device, electronic equipment and storage medium
US10115012B1 (en) Capture object boundary jitter reduction
CN111881740A (en) Face recognition method, face recognition device, electronic equipment and medium
CN112379967B (en) Simulator detection method, device, equipment and medium
CN114329399A (en) Face video verification method, device, equipment and storage medium
CN112632422B (en) Intelligent graph cutting method and device, electronic equipment and storage medium
CN107315970B (en) Sensitive data interaction method and device
CN111859985B (en) AI customer service model test method and device, electronic equipment and storage medium
CN114302207A (en) Bullet screen display method, device, system, equipment and storage medium
CN110492935B (en) Data processing system, method and device and terminal equipment
CN110532304B (en) Data processing method and device, computer readable storage medium and electronic device
CN114218166A (en) Data processing method and device, electronic equipment and readable storage medium
CN113486749A (en) Image data collection method, device, electronic equipment and computer readable medium
CN113467823A (en) Configuration information acquisition method, device, system and storage medium
US20230087081A1 (en) Method of detecting event, electronic device, and storage medium
CN111510370B (en) Content processing method and device, computer medium and electronic equipment
CN115051799B (en) Digital information processing system based on block chain

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination