WO2022022493A1 - Image authenticity determination method and system - Google Patents

Image authenticity determination method and system Download PDF

Info

Publication number
WO2022022493A1
WO2022022493A1 PCT/CN2021/108607 CN2021108607W WO2022022493A1 WO 2022022493 A1 WO2022022493 A1 WO 2022022493A1 CN 2021108607 W CN2021108607 W CN 2021108607W WO 2022022493 A1 WO2022022493 A1 WO 2022022493A1
Authority
WO
WIPO (PCT)
Prior art keywords
sequence
shooting
image
parameters
original image
Prior art date
Application number
PCT/CN2021/108607
Other languages
French (fr)
Chinese (zh)
Inventor
王智恒
张天明
张春卫
薛韬略
周士奇
汪昊
井海鹏
赵宁宁
张明文
程博
冯懋
孟辉
张远游
Original Assignee
北京嘀嘀无限科技发展有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 北京嘀嘀无限科技发展有限公司 filed Critical 北京嘀嘀无限科技发展有限公司
Publication of WO2022022493A1 publication Critical patent/WO2022022493A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/48Matching video sequences
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/40Scenes; Scene-specific elements in video content
    • G06V20/46Extracting features or characteristics from the video content, e.g. video fingerprints, representative shots or key frames
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/32Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials
    • H04L9/3226Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols including means for verifying the identity or authority of a user of the system or for message authentication, e.g. authorization, entity authentication, data integrity or data verification, non-repudiation, key authentication or verification of credentials using a predetermined code, e.g. password, passphrase or PIN
    • H04L9/3231Biological data, e.g. fingerprint, voice or retina

Definitions

  • the embodiments of this specification relate to the technical field of image processing, and in particular, to a method and system for judging the authenticity of an image.
  • the embodiments of this specification propose a method and system for judging the authenticity of an image, so as to improve the accuracy of identity authentication.
  • An aspect of the embodiments of the present specification provides a method for judging the authenticity of an image, which is applied to a server, the method includes: acquiring an original image from a client; extracting multiple images in the original image according to a preset extraction rule Or the image part is used as multiple extracted images; based on the multiple extracted images, the degree of matching between the multiple extracted images and the preset sequence is determined by the trained machine learning model; based on the matching degree, the original image is judged
  • the preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging the authenticity of the image from the shooting device.
  • An aspect of the embodiments of this specification provides a method for judging the authenticity of an image, which is applied to a client, the method includes: acquiring a shooting parameter sequence generated and delivered by a server; generating the original image based on the shooting parameter sequence ; send the original image to the server; obtain the information sent by the server that includes the result of judging the authenticity of the original image.
  • An aspect of the embodiments of the present specification provides a system for judging the authenticity of an image, which is applied to a server side.
  • the system includes: a first acquisition module for acquiring an original image from a client; an extraction module for acquiring an original image according to a preset The extraction rule extracts multiple images or image parts in the original image as multiple extracted images; the determining module is configured to determine, based on the multiple extracted images, through the trained machine learning model, the multiple extracted images and the preset images.
  • the matching degree of the sequence is set; the judgment module is used to judge the authenticity of the original image based on the matching degree; the preset sequence corresponds to the shooting device of the client, and the judgment of the authenticity of the original image is specifically: : Determine the authenticity of the image from the photographing device.
  • An aspect of the embodiments of the present specification provides a system for judging the authenticity of an image, which is applied to a client, and the system includes: a second acquisition module for acquiring a shooting parameter sequence generated and delivered by the server; a generation module for using to generate the original image based on the shooting parameter sequence; a sending module is used to send the original image to the server; a third acquisition module is used to obtain the original image sent by the server, including Information about the results of the judgment of authenticity.
  • One aspect of the embodiments of this specification provides an apparatus for judging the authenticity of an image
  • the apparatus includes a processor and a memory; the memory is used for storing instructions, and the processor is used for executing the instructions, so as to implement any of the above Operations corresponding to the method for judging the authenticity of an image.
  • One aspect of the embodiments of this specification provides a computer-readable storage medium, where the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the method for determining the authenticity of an image as described in any of the above corresponding operation.
  • FIG. 1 is a schematic diagram of an exemplary application scenario of an image authenticity judgment system according to some embodiments of the present specification
  • FIG. 2 is an exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification
  • FIG. 3 is an exemplary schematic diagram of extracting a plurality of extracted images according to some embodiments of the present specification
  • FIG. 4 is a schematic diagram of an exemplary structure of a machine learning model according to some embodiments of the present specification.
  • FIG. 5 is an exemplary flowchart of training a machine learning model according to some embodiments of the present specification
  • FIG. 6 is another exemplary structural schematic diagram of a machine learning model according to some embodiments of the present specification.
  • FIG. 7 is another exemplary flowchart of a method for judging authenticity of an image according to some embodiments of the present specification.
  • FIG. 8 is a schematic diagram of interaction between a server and a client according to some embodiments of the present specification.
  • system means for distinguishing different components, elements, parts, parts or assemblies at different levels.
  • device means for converting signals into signals.
  • unit means for converting signals into signals.
  • module means for converting signals into signals.
  • FIG. 1 is a schematic diagram of an exemplary application scenario of an image authenticity determination system according to some embodiments of the present specification.
  • the identity authentication scenarios may include scenarios in which face recognition is applied, such as face-swiping payment, face-swiping access control, and face-swiping attendance.
  • the identity authentication scenario may also include a credential identification scenario. For example, a user needs to register as a driver on an online car-hailing platform, and the online car-hailing platform will identify the driver's license, driving license and other relevant documents provided by the user, and determine whether the certificate information is true and whether it complies with relevant regulations.
  • this specification proposes a method and system for judging the authenticity of an image, which is used to effectively verify whether a video or image is a video or image shot by a shooting device on-site, that is, a legal video or image, thereby improving the accuracy of identity authentication.
  • an application scenario of the image authenticity determination system 100 shown in the embodiment of this specification may include a first computing system 140 , a second computing system 170 and a client 110 .
  • the first computing system 140 may be used to determine the authenticity of the original image.
  • the first computing system 140 may be used to determine whether the original image is a real image captured by a shooting device on site. For example, it can automatically determine whether the original images such as faces, fingerprints, palm prints, and certificates are real images captured by the shooting equipment on the spot, so as to avoid the camera being hijacked to complete false verification and improve the accuracy of identity authentication.
  • the first computing system 140 may acquire the extraction image 130 .
  • the extracted image 130 may be obtained from the original image 120 , and the original image 120 may be obtained by the client 110 .
  • the client 110 may be a photographing device, such as a camera, video recorder, camera, or the like.
  • the client 110 may be various types of devices having a camera function or including a camera device, such as a mobile phone 110-1, a tablet computer 110-2, a computer 110-3, and the like.
  • the extracted image 130 may enter the first computing system 140 through various common means (eg, a network). Through the model 141 in the first computing system 140, the degree of matching 150 can be output. The first computing system 140 further obtains a judgment result of the authenticity of the original image based on the matching degree 150 .
  • the parameters of the model 141 can be obtained through training.
  • the second computing system 170 may acquire multiple sets of training samples 160 , and each set of training samples includes sample image frames and corresponding labels.
  • the second computing system 170 updates the parameters of the initial model 171 through multiple sets of training samples 160 to obtain a trained model.
  • the parameters of the model 141 come from the trained model 171 . Among them, parameters can be passed in any common way.
  • a model may refer to a collection of several methods performed based on a processing device. These methods can include a large number of parameters.
  • the parameters used can be preset or can be dynamically adjusted. Some parameters can be obtained through training, and some parameters can be obtained during execution.
  • FIG. 4, FIG. 5 and their related descriptions For the specific description of the models involved in this specification, please refer to the relevant parts of this specification (FIG. 4, FIG. 5 and their related descriptions).
  • the first computing system 140 and the second computing system 170 may be the same or different.
  • the first computing system 140 and the second computing system 170 refer to systems with computing capabilities, which may include various computers, such as servers, personal computers, or computing platforms composed of multiple computers connected in various structures.
  • Processing devices may be included in the first computing system 140 and the second computing system 170, and the processing devices may execute program instructions.
  • the processing device may include various common general-purpose central processing units (CPUs), graphics processing units (Graphics Processing Units, GPUs), microprocessors, application-specific integrated circuits (ASICs), or other types of integrated circuits.
  • the first computing system 140 and the second computing system 170 may include storage media, and the storage media may store instructions and may also store data.
  • the storage medium may include mass memory, removable memory, volatile read-write memory, read-only memory (ROM), and the like, or any combination thereof.
  • the embodiments of this specification provide an image authenticity judgment system, the system is applied to the server side, and the image authenticity judgment system may include a first acquisition module, an extraction module, a determination module, and a judgment module.
  • the fetch module can be used to fetch raw images from the client.
  • the first obtaining module is further configured to: generate a shooting parameter sequence; send the shooting parameter sequence to the client; obtain the original image from the client, the original image Generated by the client based on the shooting parameter sequence.
  • the first obtaining module is further configured to: determine identification information of the shooting device; determine a shooting parameter set of the shooting device based on the identification information; generate the shooting parameter set based on the shooting parameter set Describe the shooting parameter sequence.
  • the first obtaining module is further configured to: randomly select a preset number of shooting parameters from the shooting parameter set, and generate the shooting parameter sequence based on the shooting parameters.
  • the sequence of shooting parameters includes a sequence of color temperature parameters.
  • the extraction module may be configured to extract multiple images or image parts in the original image as multiple extracted images according to preset extraction rules.
  • the determining module may be configured to determine, based on the multiple extracted images, the degree of matching between the multiple extracted images and the preset sequence by using a trained machine learning model.
  • the machine learning model includes at least a plurality of convolutional neural network units and a sequence-to-sequence unit; the determining module is further configured to: pass through each of the plurality of convolutional neural network units, respectively processing each of the multiple extracted images to obtain an image representation vector corresponding to each extracted image; processing the image representation vector through the sequence-to-sequence unit to obtain predictions of the multiple extracted images A change sequence of shooting parameters; the matching degree is determined based on the predicted change sequence of shooting parameters and the preset sequence.
  • the machine learning model may be trained by the following method: acquiring a plurality of training samples carrying labels, the training samples including a plurality of sample image frames obtained based on sample shooting parameters, the labels including the The variation relationship of the sample shooting parameters between the multiple sample image frames; the initial machine learning model is trained based on the multiple training samples carrying the labels, and the machine learning model is obtained.
  • the judging module can be used to judge the authenticity of the original image based on the matching degree; the preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging that the image comes from The authenticity of the photographing equipment.
  • the embodiments of this specification provide an image authenticity judgment system, which is applied to a client, and the image authenticity judgment system may include a second acquisition module, a generation module, a transmission module, a third acquisition module, and an upload module.
  • the second acquiring module may be configured to acquire the shooting parameter sequence generated and delivered by the server.
  • a generating module configured to generate the original image based on the shooting parameter sequence.
  • the shooting parameter sequence is randomly generated by the server.
  • the shooting parameter sequence is randomly generated by the server based on shooting parameters of the shooting device; the shooting parameters correspond to identification information of the shooting device.
  • the sequence of shooting parameters includes a sequence of color temperature parameters.
  • the sending module can be used to send the original image to the server.
  • the third obtaining module may be configured to obtain the information sent by the server and including the judgment result of the authenticity of the original image.
  • the uploading module may be configured to upload the identification information of the photographing device of the client to the server.
  • system and its modules can be implemented in various ways.
  • the system and its modules may be implemented in hardware, software, or a combination of software and hardware.
  • the hardware part can be realized by using dedicated logic;
  • the software part can be stored in a memory and executed by a suitable instruction execution system, such as a microprocessor or specially designed hardware.
  • a suitable instruction execution system such as a microprocessor or specially designed hardware.
  • the methods and systems described above may be implemented using computer-executable instructions and/or embodied in processor control code, for example on a carrier medium such as a disk, CD or DVD-ROM, such as a read-only memory (firmware) ) or a data carrier such as an optical or electronic signal carrier.
  • the system and its modules of this specification can be implemented not only by hardware circuits such as very large scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, etc., or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc. , can also be implemented by, for example, software executed by various types of processors, and can also be implemented by a combination of the above-mentioned hardware circuits and software (eg, firmware).
  • FIG. 2 is an exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification, and the method is applied to the server side.
  • the process 200 may be implemented by the first computing system 140 shown in FIG. 1 . As shown in FIG. 2, the process 200 may include the following steps:
  • Step 210 obtain the original image from the client.
  • this step 210 may be performed by the first acquisition module.
  • the client can be any terminal that includes a camera device. Such as mobile phones, tablets or laptops, etc.
  • the original image may be an image of the object to be detected.
  • the image of the object to be detected obtained by the application platform.
  • the object to be detected may be any object that needs to be judged whether it is collected by the photographing device on-site, that is, it is judged whether it is a real object collected by the photographing device or a false object forged in advance before the photographing device collects.
  • the object to be detected may be a face, a palm print, a fingerprint, etc. to be detected, or a certificate to be detected, such as an ID card, a driver's license, and the like.
  • the original image may be an image included in a video of the object to be detected recorded by the photographing device.
  • the video of the object to be detected recorded by the photographing device may be a video of a preset duration.
  • a 3-second or 5-second video a video of a preset duration.
  • the original image may be an image included in a 3-second or 5-second video.
  • the original image may be one or more images of the object to be detected captured by the capturing device.
  • the multiple images may be multiple images generated by the photographing device continuously photographing the object to be detected, or may be multiple images generated by the photographing device photographing the object to be detected at preset time intervals.
  • the embodiments of this specification do not specifically limit how the photographing device obtains multiple images of the object to be detected.
  • the original image may include still one or more photographs, may also include video, or a mixture thereof.
  • the first acquisition module may acquire the original image in various ways.
  • the acquisition module may acquire the original image from the storage device.
  • the original image is pre-generated and stored in the storage device.
  • the photographing device of the client 110 collects the video and/or image of the object to be detected and sends it to the storage device for storage. At this time, the acquisition module can directly obtain the original image from the storage device.
  • the acquisition module may acquire in real time the video and/or image of the object to be detected acquired by the photographing device of the client 110 .
  • the original image may be generated by the client based on a sequence of shooting parameters.
  • acquiring the original image from the client may include: generating a shooting parameter sequence, and sending the shooting parameter sequence to the client; acquiring an original image from the client, where the original image is generated by the client based on the shooting parameter sequence.
  • the sequence of shooting parameters may be a sequence of multiple shooting parameters.
  • the shooting parameters may refer to parameters used by the shooting device when recording or shooting.
  • the shooting parameters may include a color temperature parameter, a sharpening degree parameter, a color saturation parameter, a brightness parameter, a contrast parameter, a shutter parameter, an aperture parameter, and the like.
  • the shooting parameter sequence may be a color temperature parameter sequence, that is, a sequence composed of multiple color temperature parameters.
  • the above capture parameters may be mixed or combined to generate a sequence of capture parameters.
  • the embodiments of this specification take the shooting parameter sequence as the color temperature parameter sequence as an example for description.
  • the shooting parameter sequence is not limited to the color temperature parameter sequence, for example, it can also be a color saturation parameter sequence or an aperture.
  • the shooting parameter sequence may further include time information and/or image number information.
  • the time information is used to reflect the information of the corresponding time period when the photographing device uses the corresponding photographing parameters to record the video of the object to be detected. For example, taking the above-mentioned original image as an image included in a 3s video and the shooting parameter sequence as a color temperature parameter sequence as an example, if the 3s is evenly divided into 3 time periods, namely 0-1s, 1-2s and 2-3s; Then the time information can reflect the information of the videos recorded in the time periods of 0-1s, 1-2s and 2-3s respectively using the corresponding color temperature parameters.
  • the information on the number of images is used to reflect the information on the number of corresponding images of the object to be detected that the photographing device uses with the corresponding photographing parameters.
  • 10 images and the 10th to 15th images; the number of images can reflect the information of the 1st to 5th images, the 5th to 10th images, and the 10th to 15th images, respectively, using the corresponding color temperature parameters.
  • the server side may generate the shooting parameter sequence based on the shooting parameter set contained in the client terminal.
  • generating a shooting parameter sequence may include: determining identification information of the shooting device; determining a shooting parameter set of the shooting device based on the identification information; generating the shooting parameter sequence based on the shooting parameter set.
  • the identification information may include the model or performance parameters of the photographing device.
  • the first obtaining module may determine the identification information of the photographing device based on an operating system included in the client. For example, if the client is a mobile phone, the model or performance parameter corresponding to the shooting device may be determined based on the IOS or Android operating system included in the mobile phone.
  • the set of shooting parameters may be all or part of the settable shooting parameters contained in the shooting device.
  • different models of photographing devices may have different sets of photographing parameters. Still taking the above example as an example, the IOS mobile phone and the Android mobile phone have different shooting parameter sets.
  • the first computing system 140 may randomly select a preset number of shooting parameters from the shooting parameter set, and generate a shooting parameter sequence based on the shooting parameters. For example, still taking the color temperature parameter sequence as an example, if the selectable color temperature parameter ranges of the photographing device include 400-420nm, 460-470nm, 568-572nm, 6000-6500k, 10000-12000k, and 601-606nm, then the first computing system 140 You can select 5 color temperature parameter points from the above 6 color temperature parameter points, such as 460-470nm (hereinafter referred to as color temperature parameter 1), 568-572nm (hereinafter referred to as color temperature parameter 2), 6000-6500k (hereinafter referred to as color temperature parameter 3) , 10000-12000k (hereinafter referred to as color temperature parameter 4) and 601-606 nm (hereinafter referred to as color temperature parameter 5) to generate a color temperature parameter sequence, for example, the color 460-470nm (hereinafter referred to as
  • the shooting parameter sequence may include time information and/or image number information
  • the color temperature parameter sequence may include time information and/or image number information.
  • the first computing system 140 may deliver the sequence of shooting parameters to the client through the network.
  • the raw image can be generated by the client based on the sequence of shooting parameters.
  • the original image may be generated by the client's photographing device by photographing or recording a corresponding image or video based on the photographing parameter sequence.
  • Step 220 Extract multiple images or image parts in the original image as multiple extracted images according to a preset extraction rule. In some embodiments, this step 220 may be performed by an extraction module.
  • the plurality of extracted images may be a plurality of images or portions of images in the original image.
  • the plurality of extracted images may be a video of the object to be detected recorded by the photographing device or a plurality of images included in a plurality of captured images of the object to be detected.
  • the plurality of extracted images may be image portions of each of the plurality of images. For example, an image of a certain area is extracted from each image as an image part.
  • the preset extraction rules may be specifically set according to actual requirements. In some embodiments, the preset extraction rules may match how the original images were acquired. In some embodiments, preset extraction rules may be matched to sequences of shooting parameters. In some embodiments, the preset extraction rule may be matched with time information and/or image number information reflected by the shooting parameter sequence.
  • the original image is 0-1s of video recorded with color temperature parameter point 1, 1-2s of video recorded with color temperature parameter point 2, and 3s video of 2-3s recorded with color temperature parameter point 3, then
  • the preset extraction rule may be to extract any image from the video segments of 0-1s, 1-2s, and 2-3s, respectively, to generate multiple extracted images.
  • FIG. 3 is an exemplary schematic diagram illustrating extracting a plurality of extracted images according to this example.
  • the original image 310 is a video with a duration of 3s, then the last frame of image (shown in gray) can be extracted from the video segments of 0-1s, 1-2s, and 2-3s respectively, as multiple Image 320 is extracted.
  • different regions of an image may be extracted according to preset rules, and each region may be used as an extracted image, or multiple regions may be extracted from each of multiple images as multiple extracted images.
  • Step 230 Based on the multiple extracted images, determine the degree of matching between the multiple extracted images and the preset sequence by using the trained machine learning model. In some embodiments, this step 230 may be performed by a determination module.
  • the machine learning model may be a pre-trained model.
  • a trained machine learning model can determine how well multiple extracted images match a preset sequence.
  • a convolutional neural network model can be used to determine the features of each extracted image, and the output features can be compared to a preset sequence.
  • the convolutional neural network model used can be obtained by training. During training, images can be obtained by using the shooting parameters corresponding to the preset sequence, and further extracted as training data, with the corresponding preset sequence as the label, through the iterative method of optimizing the loss function. to train.
  • a combination of a convolutional neural network unit and a sequence-to-sequence (Seq2Seq) unit may be used as a machine learning model. For details, see FIG. 4 and related descriptions, which will not be repeated here.
  • a preset sequence is a conditional value used for feature recognition of an image. It can be a sequence composed of multiple values, multiple vectors, or other data, or a value or other representation can be uniformly referred to as a sequence.
  • the preset sequence corresponds to the client's camera device.
  • the preset sequence may be a sequence corresponding to changes in parameters in the sequence of shooting parameters.
  • the preset sequence may be the change of color temperature parameters 1 to 2 and the change of color temperature parameters 2 to 3. A sequence of changes made up of changes.
  • the changed values of the parameters in the preset sequence may be determined by preset encoding information.
  • the preset sequence may be a code value used to distinguish photographing equipment.
  • the preset sequence may have various representation forms, and its functions are not substantially different, which is not limited in this specification.
  • the matching degree may reflect the similarity between the plurality of extracted images and the preset sequence. In some embodiments, the matching degree may reflect the similarity between the sequence of changes of the predicted shooting parameters of the multiple extracted images and the preset sequence, that is, the difference between the sequence of changes of the predicted shooting parameters of the multiple extracted images and the parameters in the sequence of shooting parameters Similarity between sequences of changes. It can be understood that the greater the value of the matching degree, the greater the similarity between the two, and the greater the possibility that the original image is real.
  • an attacker can effectively prevent an attacker from bypassing the photographing device and directly uploading a prefabricated image.
  • an attacker can effectively prevent an attacker from bypassing the photographing device and directly uploading a prefabricated image.
  • an attacker can effectively prevent an attacker from bypassing the photographing device and directly uploading a prefabricated image.
  • the shooting device of the client by generating and delivering the shooting parameter sequence to the client, the shooting device of the client generates the corresponding original image. Therefore, the shooting parameters of the original image actually shot by the shooting device must be the same as the shooting parameter sequence. The shooting parameters are the same. Since the pre-prepared fake images cannot contain exactly the same shooting parameters, the attacker can prevent the attacker from hijacking the shooting equipment and complete the identity authentication through the fake images. Since the shooting parameter sequence in some embodiments of this specification is randomly issued and ready to use, an attacker cannot make false images through the shooting device in advance, which greatly improves the reliability of on-site verification.
  • the change sequence of the predicted shooting parameters is compared with the change sequence of the parameters in the shooting parameter sequence to determine the matching degree, that is, the matching degree is determined by comparing the changes of the parameters, which can eliminate the influence of the environment on the specific parameter value.
  • the environment such as light
  • the determination of the color temperature parameter which improves the accuracy of subsequent determination of authenticity through matching, thereby improving the accuracy of identity authentication.
  • Step 240 based on the matching degree, determine the authenticity of the original image. This step 240 may be performed by a judgment module.
  • judging the authenticity of the original image is specifically: judging the authenticity of the image from a photographing device. According to the description of the above step 210, it can be known that the original image from the photographing device is a real image, and on the contrary, it is a fake image.
  • the judgment module may judge the authenticity of the original image based on the matching degree. For example, when the matching degree is greater than a preset threshold, the original image is a real image.
  • the first computing system 140 may send a relevant instruction to the client 110 to terminate further operations of the client 110 (such as registering the application platform).
  • This step may be performed by other executive bodies, or performed in other transforming manners, and has no substantial impact on the technical solutions of this specification.
  • FIG. 4 is a schematic diagram of an exemplary structure of a machine learning model according to some embodiments of the present specification.
  • the machine learning model 400 may include at least a plurality of convolutional neural network units 410 and a sequence-to-sequence unit 420 .
  • Each of the plurality of convolutional neural network units 410 may be configured to process each of the plurality of extracted images to obtain an image representation vector corresponding to each of the extracted images.
  • each of the plurality of convolutional neural network units 410 may process each of the plurality of extracted images (eg, extracted image 1 to extracted image n) extracted in the above step 240 to obtain the extracted image.
  • the convolutional neural network unit 410 may employ a conventional convolutional neural network including a base convolutional layer 4101 and a fully connected layer 4102, eg, LeNet, AlexNet, GoogLeNet, and the like.
  • the sequence-to-sequence unit 420 may process the image representation vector to obtain a sequence of changes of predicted shooting parameters of a plurality of extracted images. Specifically, the sequence-to-sequence unit 420 may process the image representation vector output by each fully connected layer 4102 in the multiple convolutional neural network units 410 to obtain a sequence of changes of predicted shooting parameters of multiple extracted images.
  • the sequence of changes in the predicted shooting parameters of the plurality of extracted images may refer to a sequence formed by changes of the predicted shooting parameters between each of the multiple extracted images.
  • the predicted shot parameters match shot parameters included in the sequence of shot parameters. For example, if the shooting parameter sequence is a color temperature parameter sequence, the predicted shooting parameter is a color temperature parameter.
  • the shooting parameter sequence is a color temperature parameter sequence
  • the predicted shooting parameter change sequence can be the predicted color temperature parameter change between the extracted image 1 and the extracted image 2, and the predicted change between the extracted image 2 and the extracted image 3.
  • the extracted image input by the convolutional neural network unit 410 may be extracted from the original image collected by the photographing device, the environmental factors (such as light and shade) of the location where the photographing device is collected may affect the color distribution of the original image, due to some Shooting parameters (such as color temperature parameters) reflect the color distribution of the image and therefore may reduce the prediction accuracy of the machine learning model.
  • the parameters of the machine learning model can be adjusted by using the comparison image frames, so that the machine learning model can accurately obtain the change sequence of the predicted shooting parameters of multiple extracted images, thereby reducing the environmental factors affecting the machine learning model. 400 impact.
  • the comparison image frame can be a frame of image captured under the specified shooting parameters, and the image representation vector is obtained by inputting the frame image into the convolutional neural network unit 410, and then the image representation vector is obtained based on the relationship between the image representation vector and the vector of the specified shooting parameters. The difference between the parameters of the convolutional neural network unit 410 is adjusted until the obtained image representation vector is the same as the specified shooting parameter value.
  • FIG. 5 is an exemplary flowchart of training a machine learning model according to some embodiments of the present specification.
  • the machine learning model 400 may be a model constructed by the base convolutional layer 4101 , the fully connected layer 4102 , and the sequence-to-sequence unit 420 .
  • the process 500 may include the following steps:
  • Step 510 Acquire a plurality of training samples carrying labels, where the training samples include a plurality of sample image frames obtained based on the sample shooting parameters, and the labels include a change relationship of the sample shooting parameters among the plurality of sample image frames.
  • the training samples may be data input into the initial machine learning model for training the machine learning model.
  • the training samples may include a plurality of sample image frames acquired based on sample capture parameters. For example, still taking the sample shooting parameters as the color temperature parameters, and the color temperature parameters including color temperature parameters 1 to 6 as an example, one of the training samples may be the sample image frame 1 obtained based on the color temperature parameter 1, and the sample image frame obtained based on the color temperature parameter 2. 2. The sample image frame 3 obtained based on the color temperature parameter 5.
  • the label may include the variation relationship of the sample capture parameters among the plurality of sample image frames.
  • the label may be a sequence of changes in sample capture parameters between multiple sample image frames.
  • Step 520 Train an initial machine learning model based on the plurality of labeled training samples to obtain the machine learning model.
  • some shooting parameters reflect the color distribution of the image. Its impact on the image is global, for example, the color distribution of the same extracted image under different receptive fields should be similar or the same.
  • the convolutional neural network unit 410 in the machine learning model 400 shown in FIG. 4 will focus on the contour information in the image. Therefore, in some embodiments, the machine learning model 400 shown in FIG. 4 may be improved, so that It identifies global features.
  • the constructed machine learning model 600 may further include a first sampling convolution layer 4103 and a second sampling convolution layer 4104.
  • a first sampled convolutional layer 4103 and a second sampled convolutional layer 4104 may be added to the machine learning model 400 during training.
  • the basic convolution layer 4101 of the convolutional neural network unit 410 is connected to the first sampling convolution layer 4103 and the second sampling convolution layer 4104 respectively.
  • the parameters of the machine learning model 600 can be adjusted during the training process of the machine learning model 600, thereby ensuring that the convolutional neural network unit 410 for the same extracted image in different
  • the color distribution under the receptive field is similar or the same. Avoid the convolutional neural network unit 410 focusing on extracting contour information in the image (for example, focusing on extracting the outline of objects in the image), strengthen the convolutional neural network unit 410's ability to recognize global features, and then improve the convolutional neural network unit 410. Recognition ability to extract images.
  • the first sampled convolutional layer 4103 and the second sampled convolutional layer 4104 may be atrous convolutional layers.
  • the convolution kernels of the first sampled convolutional layer 4103 and the second sampled convolutional layer 4104 have the same size.
  • the size of the convolution kernel of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 is both 3*3.
  • the convolution sampling points of the convolution kernels of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 have different spacings.
  • the sampling interval of the first sampling convolution layer 4103 is 0, and the sampling interval of the second sampling convolution layer 4104 is 2.
  • the sampling interval of the first sampling convolution layer 4103 is 2, and the sampling interval of the second sampling convolution layer 4104 is 0.
  • the sampling intervals of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 are different, correspondingly, the first sampling convolution layer 4103 and the second sampling convolution layer 4104 have different receptive fields for the same extracted image .
  • the sampling interval of the first sampling convolutional layer 4103 is 0 and the sampling interval of the second sampling convolutional layer 4104 as 2 as an example, since the sampling interval of the second sampling convolutional layer 4104 is larger, the sampling interval of the second sampling convolutional layer 4104 is larger.
  • the receptive field of the two-sampling convolutional layer 4103 is larger than that of the second sampling convolutional layer 4104, and a wider receptive field is obtained through the second sampling convolutional layer 4104, which makes better use of the global features of the image.
  • an initial machine learning model may be trained end-to-end based on a plurality of labeled training samples to obtain a trained machine learning model (eg, machine learning models 400 and 600 ).
  • the parameters of the initial machine learning model can be continuously adjusted to reduce the loss function value corresponding to each training sample, so that the loss function value satisfies the preset condition. For example, the loss function value converges, or the loss function value is smaller than a preset value.
  • the model training is completed, and the trained machine learning model is obtained.
  • the loss function value corresponding to each training sample may be determined through the following process: processing a plurality of sample image frames through an initial machine learning model, obtaining the change relationship of the predicted sample shooting parameters among the plurality of sample image frames, The difference between the change relationship of the sample shooting parameters and the change relationship of the sample shooting parameters in the label determines the loss function value corresponding to the training sample.
  • the machine learning model 600 may be a model constructed by a base convolutional layer 4101 , a fully connected layer 4102 , a first sampled convolutional layer 4103 , a second sampled convolutional layer 4104 , and a sequence-to-sequence unit 420 .
  • the model 600 can also determine the color distribution difference of each extracted image under different receptive fields based on the first sampling convolution layer 4103 and the second sampling convolution layer 4104, The loss function corresponding to the training sample is determined based on the difference.
  • the KL divergence of the first sampling convolutional layer 4103 and the second sampling convolutional layer 4104 can be calculated based on the feature vectors output by the two to determine the color distribution difference of each extracted image under different receptive fields, It is taken as the constraint parameter of the convolutional neural network unit 410, and then the convolutional neural network unit 410 can ensure that the color distribution of the same extracted image under different receptive fields is approximate by adjusting the constraint parameter. For example, adjust the constraint parameter to a minimum value of 0.
  • the constraint parameter D KL of the convolutional neural network unit 410 can be determined by the following formula (1):
  • P conv2 (x)) represents the KL divergence calculation for P conv1 (x) and P conv2 (x)
  • P conv1 (x) represents the output of the first sampling convolutional layer
  • Feature vector, P conv2 (x) represents the feature vector output by the second sampling convolution layer
  • argmin represents the minimum value of the KL divergence calculation.
  • the constraint parameter is 0.
  • the difference between the feature vectors output by the first sampling convolution layer 4103 and the second sampling convolution layer 4104 is the smallest, and the feature distribution is the most similar. Therefore, the same extracted image has different receptive fields.
  • the color distribution below is the same.
  • the model training is completed, and the trained machine learning model 400 can be obtained, or when the loss functions constructed based on labels and constraint parameters satisfy the preset conditions, the model After the training is completed, the trained machine learning model 600 is obtained.
  • the convolutional neural network unit 410 and the machine learning model 600 may perform joint training to optimize the parameters of the basic convolutional layer 4101, so that the feature vector generated by the basic convolutional layer 4101 can better reflect the in-image and the shooting parameter-related overall features, thereby improving the recognition effect of the machine learning model 600 .
  • FIG. 7 is another exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification, and the method is applied to a client.
  • the process 700 may be implemented by the client 110 shown in FIG. 1 . As shown in FIG. 7, the process 700 may include the following steps:
  • Step 710 Obtain the shooting parameter sequence generated and delivered by the server.
  • this step 710 may be performed by a second acquisition module.
  • the shooting parameter sequence may be randomly generated on the server side. In some embodiments, the shooting parameter sequence may be randomly generated by the server based on a shooting parameter set of a shooting device; the shooting parameter set corresponds to the identification information of the shooting device. In some embodiments, the sequence of shooting parameters may include a sequence of color temperature parameters. For the specific details of step 710, reference may be made to the above-mentioned step 210 and its related description.
  • Step 720 Generate the original image based on the shooting parameter sequence.
  • this step 720 may be performed by a generation module.
  • step 720 For the specific details of step 720, reference may be made to the above-mentioned step 210 and its related description.
  • Step 730 Send the original image to the server.
  • this step 730 may be performed by a sending module.
  • the client 110 may send the original image to the server (eg, the first computing system 140 ) through the network.
  • the server eg, the first computing system 140
  • the client 110 may send the original image to the server (eg, the first computing system 140 ) through the network.
  • Step 740 Obtain the information sent by the server that includes the result of the judgment on the authenticity of the original image.
  • the client 110 may obtain the information sent by the server (eg, the first computing system 140 ) through the network, and including the judgment result of the authenticity of the original image.
  • the information of the judging result of the authenticity of the original image may include the judging result of whether the original image is authentic.
  • the client may also obtain verification information of the original image sent based on the judgment result. For example, whether the face or documents meet the requirements.
  • the client may also obtain relevant instructions sent by the server based on the judgment result. For example, when it is determined that the original image is a fake image, the server may send a termination instruction to terminate further operations of the client (for example, registering the application platform).
  • the embodiments of this specification describe the method for judging the authenticity of an image from the perspectives of the server side and the client side.
  • the embodiments of the present specification describe the method for judging the authenticity of an image from the overall perspective of the server side and the client side.
  • FIG. 8 is a schematic diagram of interaction between a server and a client according to some embodiments of the present specification.
  • the interaction between the server and the client in the schematic interaction diagram 800 includes but is not limited to: the server obtains the device model from the client, and then the server can determine the shooting parameter set included in the shooting device based on the device model of the client, A shooting parameter sequence is generated based on the shooting parameter set.
  • the server sends the shooting parameter sequence to the client, and the shooting device of the client generates the original image based on the shooting parameter sequence.
  • the server side obtains the original image from the client side, judges the authenticity of the original image through the methods of steps 220 to 240 above, and sends information including the judgment result to the client side.
  • An embodiment of the present specification further provides an apparatus for judging the authenticity of an image, the apparatus includes a processor and a memory; the memory is used for storing instructions, and the processor is used for executing the instructions, so as to achieve the above-mentioned items The operation corresponding to the method of judging the authenticity of the image.
  • Embodiments of this specification also provide a computer-readable storage medium, where the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the method corresponding to the method for judging the authenticity of an image described in any preceding item. operate.
  • the possible beneficial effects of the embodiments of this specification include, but are not limited to: (1) By generating and delivering a shooting parameter sequence to the client, the shooting device of the client generates a corresponding original image. Therefore, the original image actually shot by the shooting device The shooting parameters of the image must be consistent with the shooting parameters in the shooting parameter sequence.
  • the attacker can prevent the attacker from hijacking the shooting equipment and complete the identity authentication through the fake image; (2)
  • the shooting parameter sequence is randomly issued, and it can be used immediately, and the attacker can also It is impossible to make false images through the shooting equipment in advance, which greatly improves the reliability of on-site verification; (3)
  • the recognition ability of the convolutional neural network unit to the global features of the image is strengthened, and the recognition of the extracted image by the convolutional neural network unit is improved. ability to improve the prediction accuracy of machine learning models. It should be noted that different embodiments may have different beneficial effects, and in different embodiments, the possible beneficial effects may be any one or a combination of the above, or any other possible beneficial effects.
  • aspects of this specification may be illustrated and described in several patentable categories or situations, including any new and useful process, machine, product, or combination of matter, or combinations of them. of any new and useful improvements. Accordingly, various aspects of this specification may be performed entirely in hardware, entirely in software (including firmware, resident software, microcode, etc.), or in a combination of hardware and software.
  • the above hardware or software may be referred to as a "data block”, “module”, “engine”, “unit”, “component” or “system”.
  • aspects of this specification may be embodied as a computer product comprising computer readable program code embodied in one or more computer readable media.
  • a computer storage medium may contain a propagated data signal with the computer program code embodied therein, for example, on baseband or as part of a carrier wave.
  • the propagating signal may take a variety of manifestations, including electromagnetic, optical, etc., or a suitable combination.
  • Computer storage media can be any computer-readable media other than computer-readable storage media that can communicate, propagate, or transmit a program for use by coupling to an instruction execution system, apparatus, or device.
  • Program code on a computer storage medium may be transmitted over any suitable medium, including radio, cable, fiber optic cable, RF, or the like, or a combination of any of the foregoing.
  • the computer program coding required for the operation of the various parts of this manual may be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python etc., conventional procedural programming languages such as C language, Visual Basic, Fortran2003, Perl, COBOL2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages.
  • the program code may run entirely on the user's computer, or as a stand-alone software package on the user's computer, or partly on the user's computer and partly on a remote computer, or entirely on the remote computer or processing device.
  • the remote computer can be connected to the user's computer through any network, such as a local area network (LAN) or wide area network (WAN), or to an external computer (eg, through the Internet), or in a cloud computing environment, or as a service Use eg software as a service (SaaS).
  • LAN local area network
  • WAN wide area network
  • SaaS software as a service

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Multimedia (AREA)
  • General Engineering & Computer Science (AREA)
  • Computer Security & Cryptography (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biodiversity & Conservation Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Image Analysis (AREA)

Abstract

Disclosed in embodiments of the present description is an image authenticity determination method and system. The method is applied to a server, and the method comprises: obtaining an original image from a client; extracting, according to a preset extraction rule, multiple images or image parts from the original image as multiple extracted images; determining, by means of a trained machine learning model on the basis of the multiple extracted images, the degree of matching between the multiple extracted images and a preset sequence; and determining the authenticity of the original image on the basis of the degree of matching, the preset sequence corresponding to a camera device of the client, and determining the authenticity of the original image specifically comprising: determining the authenticity of an image from the camera device.

Description

一种判断图像真实性的方法及系统A method and system for judging the authenticity of an image
交叉引用cross reference
本申请要求于2020年7月30日提交的申请号为202010750632.5的中国申请的优先权,其全部内容通过引用结合于此。This application claims priority to Chinese Application No. 202010750632.5 filed on July 30, 2020, the entire contents of which are incorporated herein by reference.
技术领域technical field
本说明书实施例涉及图像处理技术领域,特别涉及一种判断图像真实性的方法及系统。The embodiments of this specification relate to the technical field of image processing, and in particular, to a method and system for judging the authenticity of an image.
背景技术Background technique
随着科学技术的快速发展,越来越多的应用场景(例如安防、金融以及应用平台的用户注册等)需要通过证件识别、人脸识别对用户身份进行验证。为防止不法分子伪造、冒用他人身份进行违法犯罪,对待识别的证件图像或人脸图像的真伪进行鉴别是身份认证的关键环节。With the rapid development of science and technology, more and more application scenarios (such as security, finance, and user registration of application platforms, etc.) need to verify user identity through certificate recognition and face recognition. In order to prevent criminals from forging and using other people's identities to commit crimes, it is a key link in identity authentication to identify the authenticity of the identified document images or face images.
为此,本说明书实施例提出一种判断图像真实性的方法及系统,提高身份认证的准确性。To this end, the embodiments of this specification propose a method and system for judging the authenticity of an image, so as to improve the accuracy of identity authentication.
发明内容SUMMARY OF THE INVENTION
本说明书实施例的一个方面提供一种判断图像真实性的方法,应用于服务器端,所述方法包括:从客户端获取原始图像;根据预设的提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像;基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度;基于所述匹配度,判断所述原始图像的真实性;所述预设序列对应于客户端的拍摄设备,判断所述原始图像的真实性具体为:判断所述图像来自所述拍摄设备的真实性。An aspect of the embodiments of the present specification provides a method for judging the authenticity of an image, which is applied to a server, the method includes: acquiring an original image from a client; extracting multiple images in the original image according to a preset extraction rule Or the image part is used as multiple extracted images; based on the multiple extracted images, the degree of matching between the multiple extracted images and the preset sequence is determined by the trained machine learning model; based on the matching degree, the original image is judged The preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging the authenticity of the image from the shooting device.
本说明书实施例的一个方面提供一种判断图像真实性的方法,应用于客户端,所述方法包括:获取服务器端生成并下发的拍摄参数序列;基于所述拍摄参数序列生成所述原始图像;将所述原始图像发送给服务器端;获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。An aspect of the embodiments of this specification provides a method for judging the authenticity of an image, which is applied to a client, the method includes: acquiring a shooting parameter sequence generated and delivered by a server; generating the original image based on the shooting parameter sequence ; send the original image to the server; obtain the information sent by the server that includes the result of judging the authenticity of the original image.
本说明书实施例的一个方面提供一种判断图像真实性的系统,应用于服务器端,所述系统包括:第一获取模块,用于从客户端获取原始图像;提取模块,用于根据预设的提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像;确定模块,用于基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设 序列的匹配度;判断模块,用于基于所述匹配度,判断所述原始图像的真实性;所述预设序列对应于所述客户端的拍摄设备,判断所述原始图像的真实性具体为:判断所述图像来自所述拍摄设备的真实性。An aspect of the embodiments of the present specification provides a system for judging the authenticity of an image, which is applied to a server side. The system includes: a first acquisition module for acquiring an original image from a client; an extraction module for acquiring an original image according to a preset The extraction rule extracts multiple images or image parts in the original image as multiple extracted images; the determining module is configured to determine, based on the multiple extracted images, through the trained machine learning model, the multiple extracted images and the preset images. The matching degree of the sequence is set; the judgment module is used to judge the authenticity of the original image based on the matching degree; the preset sequence corresponds to the shooting device of the client, and the judgment of the authenticity of the original image is specifically: : Determine the authenticity of the image from the photographing device.
本说明书实施例的一个方面提供一种判断图像真实性的系统,应用于客户端,所述系统包括:第二获取模块,用于获取服务器端生成并下发的拍摄参数序列;生成模块,用于基于所述拍摄参数序列生成所述原始图像;发送模块,用于将所述原始图像发送给服务器端;第三获取模块,用于获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。An aspect of the embodiments of the present specification provides a system for judging the authenticity of an image, which is applied to a client, and the system includes: a second acquisition module for acquiring a shooting parameter sequence generated and delivered by the server; a generation module for using to generate the original image based on the shooting parameter sequence; a sending module is used to send the original image to the server; a third acquisition module is used to obtain the original image sent by the server, including Information about the results of the judgment of authenticity.
本说明书实施例的一个方面提供一种判断图像真实性的装置,所述装置包括处理器以及存储器;所述存储器用于存储指令,所述处理器用于执行所述指令,以实现如上任一项所述判断图像真实性的方法对应的操作。One aspect of the embodiments of this specification provides an apparatus for judging the authenticity of an image, the apparatus includes a processor and a memory; the memory is used for storing instructions, and the processor is used for executing the instructions, so as to implement any of the above Operations corresponding to the method for judging the authenticity of an image.
本说明书实施例的一个方面提供一种计算机可读存储介质,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行如上任一项所述判断图像真实性的方法对应的操作。One aspect of the embodiments of this specification provides a computer-readable storage medium, where the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the method for determining the authenticity of an image as described in any of the above corresponding operation.
附图说明Description of drawings
本说明书将以示例性实施例的方式进一步描述,这些示例性实施例将通过附图进行详细描述。这些实施例并非限制性的,在这些实施例中,相同的编号表示相同的结构,其中:This specification will be further described by way of example embodiments, which will be described in detail with reference to the accompanying drawings. These examples are not limiting, and in these examples, the same numbers refer to the same structures, wherein:
图1是根据本说明书的一些实施例所示的图像真实性判断系统的示例性应用场景示意图;1 is a schematic diagram of an exemplary application scenario of an image authenticity judgment system according to some embodiments of the present specification;
图2是根据本说明书的一些实施例所示的判断图像真实性的方法的示例性流程图;FIG. 2 is an exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification;
图3是根据本说明书的一些实施例所示的提取多个提取图像的示例性示意图;FIG. 3 is an exemplary schematic diagram of extracting a plurality of extracted images according to some embodiments of the present specification;
图4是根据本说明书的一些实施例所示的机器学习模型的示例性结构示意图;4 is a schematic diagram of an exemplary structure of a machine learning model according to some embodiments of the present specification;
图5是根据本说明书的一些实施例所示的训练机器学习模型的示例性流程图;FIG. 5 is an exemplary flowchart of training a machine learning model according to some embodiments of the present specification;
图6是根据本说明书的一些实施例所示的机器学习模型的另一示例性结构示意图;FIG. 6 is another exemplary structural schematic diagram of a machine learning model according to some embodiments of the present specification;
图7是根据本说明书的一些实施例所示的判断图像真实性的方法的另一示例性流程图;FIG. 7 is another exemplary flowchart of a method for judging authenticity of an image according to some embodiments of the present specification;
图8是根据本说明书的一些实施例所示的服务器端和客户端的交互示意图。FIG. 8 is a schematic diagram of interaction between a server and a client according to some embodiments of the present specification.
具体实施方式detailed description
为了更清楚地说明本说明书实施例的技术方案,下面将对实施例描述中所需要使用的附图作简单的介绍。显而易见地,下面描述中的附图仅仅是本说明书的一些示例或实施例,对于本领域的普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据这些附图将本说明书应用于其它类似情景。除非从语言环境中显而易见或另做说明,图中相同标号代表相同结构或操作。In order to illustrate the technical solutions of the embodiments of the present specification more clearly, the following briefly introduces the accompanying drawings used in the description of the embodiments. Obviously, the accompanying drawings in the following description are only some examples or embodiments of the present specification. For those of ordinary skill in the art, without creative efforts, the present specification can also be applied to the present specification according to these drawings. other similar situations. Unless obvious from the locale or otherwise specified, the same reference numbers in the figures represent the same structure or operation.
应当理解,本说明书中所使用的“系统”、“装置”、“单元”和/或“模组”是用于区分不同级别的不同组件、元件、部件、部分或装配的一种方法。然而,如果其他词语可实现相同的目的,则可通过其他表达来替换所述词语。It should be understood that "system", "device", "unit" and/or "module" as used in this specification is a method used to distinguish different components, elements, parts, parts or assemblies at different levels. However, other words may be replaced by other expressions if they serve the same purpose.
如本说明书和权利要求书中所示,除非上下文明确提示例外情形,“一”、“一个”、“一种”和/或“该”等词并非特指单数,也可包括复数。一般说来,术语“包括”与“包含”仅提示包括已明确标识的步骤和元素,而这些步骤和元素不构成一个排它性的罗列,方法或者设备也可能包含其它的步骤或元素。As shown in the specification and claims, unless the context clearly dictates otherwise, the words "a", "an", "an" and/or "the" are not intended to be specific in the singular and may include the plural. Generally speaking, the terms "comprising" and "comprising" only imply that the clearly identified steps and elements are included, and these steps and elements do not constitute an exclusive list, and the method or apparatus may also include other steps or elements.
本说明书中使用了流程图用来说明根据本说明书的实施例的系统所执行的操作。应当理解的是,前面或后面操作不一定按照顺序来精确地执行。相反,可以按照倒序或同时处理各个步骤。同时,也可以将其他操作添加到这些过程中,或从这些过程移除某一步或数步操作。Flowcharts are used in this specification to illustrate operations performed by a system according to an embodiment of this specification. It should be understood that the preceding or following operations are not necessarily performed in the exact order. Instead, the various steps can be processed in reverse order or simultaneously. At the same time, other actions can be added to these procedures, or a step or steps can be removed from these procedures.
图1是根据本说明书的一些实施例所示的图像真实性判断系统的示例性应用场景示意图。FIG. 1 is a schematic diagram of an exemplary application scenario of an image authenticity determination system according to some embodiments of the present specification.
人们的生产生活领域,常常会遇到需要进行身份认证的场景。在一些实施例中,身份认证场景可以包括刷脸支付、刷脸门禁、刷脸考勤等应用人脸识别的场景。在一些实施例中,身份认证场景还可以包括证件识别场景。例如,用户需要在网约车平台注册成为司机,网约车平台会对用户提供的驾驶证、行驶证等相关证件进行识别,判断证件信息是否真实以及是否符合相关规定。In the field of production and life, people often encounter scenarios that require identity authentication. In some embodiments, the identity authentication scenarios may include scenarios in which face recognition is applied, such as face-swiping payment, face-swiping access control, and face-swiping attendance. In some embodiments, the identity authentication scenario may also include a credential identification scenario. For example, a user needs to register as a driver on an online car-hailing platform, and the online car-hailing platform will identify the driver's license, driving license and other relevant documents provided by the user, and determine whether the certificate information is true and whether it complies with relevant regulations.
然而,一些用户会使用提前录制的人脸动作视频进行人脸识别或者使用作弊的证件图片进行证件识别,完成虚假验证。例如,在证件识别中,不法分子可以通过黑产劫持应用平台摄像头,直接绕过现场拍摄实物证件的流程,把作弊图片上传到应用平台进行注册,从而完成虚假验证。又例如,在人脸识别中,不法分子事前录制在人脸识别 中常用的点头、眨眼、张嘴等动作,并在验证时现场劫持摄像头并输入事先录制的相应动作视频完成虚假验证。这些虚假验证降低了身份认证的准确性,同时也会产生一些安全隐患。为此,本说明书提出一种判断图像真实性的方法及系统,用于有效验证视频或图像是否为拍摄设备现场拍摄的视频或图像,即合法的视频或图像,从而提高身份认证的准确性。However, some users will use pre-recorded face action videos for face recognition or use fraudulent ID pictures for ID identification to complete false verification. For example, in the identification of documents, criminals can hijack the camera of the application platform through black production, directly bypass the process of shooting physical documents on the spot, and upload the cheating picture to the application platform for registration, thereby completing false verification. For another example, in face recognition, criminals record in advance actions such as nodding, blinking, and mouth opening, which are commonly used in face recognition, and hijack the camera on the spot during verification and enter the corresponding action video recorded in advance to complete false verification. These false verifications reduce the accuracy of identity authentication and also create some security risks. To this end, this specification proposes a method and system for judging the authenticity of an image, which is used to effectively verify whether a video or image is a video or image shot by a shooting device on-site, that is, a legal video or image, thereby improving the accuracy of identity authentication.
如图1所示,本说明书实施例所示的图像真实性判断系统100的应用场景中可以包括第一计算系统140、第二计算系统170和客户端110。As shown in FIG. 1 , an application scenario of the image authenticity determination system 100 shown in the embodiment of this specification may include a first computing system 140 , a second computing system 170 and a client 110 .
第一计算系统140可以用于判断原始图像的真实性。在一些实施例中,第一计算系统140可以用于判断原始图像是否为拍摄设备现场拍摄的真实图像。例如,自动判断人脸、指纹、掌纹、证件等原始图像是否为拍摄设备现场拍摄的真实图像,避免摄像头被劫持而完成虚假验证,提高身份认证的准确性。The first computing system 140 may be used to determine the authenticity of the original image. In some embodiments, the first computing system 140 may be used to determine whether the original image is a real image captured by a shooting device on site. For example, it can automatically determine whether the original images such as faces, fingerprints, palm prints, and certificates are real images captured by the shooting equipment on the spot, so as to avoid the camera being hijacked to complete false verification and improve the accuracy of identity authentication.
第一计算系统140可以获取提取图像130。提取图像130可以从原始图像120中获取,原始图像120可以通过客户端110获取。在一些实施例中,客户端110可以是拍摄设备,例如相机、录像机、摄像头等。在一些实施例中,客户端110可以是各类具有摄像功能或包括拍摄设备的设备,例如手机110-1、平板电脑110-2、计算机110-3等。The first computing system 140 may acquire the extraction image 130 . The extracted image 130 may be obtained from the original image 120 , and the original image 120 may be obtained by the client 110 . In some embodiments, the client 110 may be a photographing device, such as a camera, video recorder, camera, or the like. In some embodiments, the client 110 may be various types of devices having a camera function or including a camera device, such as a mobile phone 110-1, a tablet computer 110-2, a computer 110-3, and the like.
提取图像130可以通过各种常见的方式(例如,网络)进入第一计算系统140。通过第一计算系统140中的模型141,可以输出匹配度150。第一计算系统140进一步基于匹配度150得到原始图像的真实性的判断结果。The extracted image 130 may enter the first computing system 140 through various common means (eg, a network). Through the model 141 in the first computing system 140, the degree of matching 150 can be output. The first computing system 140 further obtains a judgment result of the authenticity of the original image based on the matching degree 150 .
模型141的参数可以通过训练得到。第二计算系统170可以获取多组训练样本160,每组训练样本包含样本图像帧及对应的标签。第二计算系统170通过多组训练样本160更新初始模型171的参数,得到训练好的模型。模型141的参数来自于训练好的模型171。其中,参数可以以任何常见的方式传递。The parameters of the model 141 can be obtained through training. The second computing system 170 may acquire multiple sets of training samples 160 , and each set of training samples includes sample image frames and corresponding labels. The second computing system 170 updates the parameters of the initial model 171 through multiple sets of training samples 160 to obtain a trained model. The parameters of the model 141 come from the trained model 171 . Among them, parameters can be passed in any common way.
模型(例如,模型141或/和模型171)可以指基于处理设备而进行的若干方法的集合。这些方法可以包括大量的参数。在执行模型时,所使用的参数可以是被预先设置好的,也可以是可以动态调整的。一些参数可以通过训练的方法获得,一些参数可以在执行的过程中获得。关于本说明书中涉及模型的具体说明,可参见本说明书的相关部分(图4、图5以及其相关描述)。A model (eg, model 141 or/and model 171 ) may refer to a collection of several methods performed based on a processing device. These methods can include a large number of parameters. When executing the model, the parameters used can be preset or can be dynamically adjusted. Some parameters can be obtained through training, and some parameters can be obtained during execution. For the specific description of the models involved in this specification, please refer to the relevant parts of this specification (FIG. 4, FIG. 5 and their related descriptions).
第一计算系统140和第二计算系统170可以相同也可以不同。第一计算系统140和第二计算系统170是指具有计算能力的系统,可以包括各种计算机,比如服务器、个人计算机,也可以是由多台计算机以各种结构连接组成的计算平台。The first computing system 140 and the second computing system 170 may be the same or different. The first computing system 140 and the second computing system 170 refer to systems with computing capabilities, which may include various computers, such as servers, personal computers, or computing platforms composed of multiple computers connected in various structures.
第一计算系统140和第二计算系统170中可以包括处理设备,处理设备可以执行程序指令。处理设备可以包括各种常见的通用中央处理器(central processing unit,CPU),图形处理器(Graphics Processing Unit,GPU)、微处理器、特殊应用集成电路(application-specific integrated circuit,ASIC)、或其他类型的集成电路。Processing devices may be included in the first computing system 140 and the second computing system 170, and the processing devices may execute program instructions. The processing device may include various common general-purpose central processing units (CPUs), graphics processing units (Graphics Processing Units, GPUs), microprocessors, application-specific integrated circuits (ASICs), or other types of integrated circuits.
第一计算系统140和第二计算系统170中可以包括存储介质,存储介质可以存储指令,也可以存储数据。存储介质可包括大容量存储器、可移动存储器、易失性读写存储器、只读存储器(ROM)等或其任意组合。The first computing system 140 and the second computing system 170 may include storage media, and the storage media may store instructions and may also store data. The storage medium may include mass memory, removable memory, volatile read-write memory, read-only memory (ROM), and the like, or any combination thereof.
本说明书实施例提供一种图像真实性判断系统,该系统应用于服务器端,所述图像真实性判断系统可以包括第一获取模块、提取模块、确定模块、以及判断模块。The embodiments of this specification provide an image authenticity judgment system, the system is applied to the server side, and the image authenticity judgment system may include a first acquisition module, an extraction module, a determination module, and a judgment module.
获取模块可以用于从客户端获取原始图像。在一些实施例中,所述第一获取模块进一步用于:生成拍摄参数序列;将所述拍摄参数序列下发至所述客户端;从所述客户端获取所述原始图像,所述原始图像由所述客户端基于所述拍摄参数序列生成。The fetch module can be used to fetch raw images from the client. In some embodiments, the first obtaining module is further configured to: generate a shooting parameter sequence; send the shooting parameter sequence to the client; obtain the original image from the client, the original image Generated by the client based on the shooting parameter sequence.
在一些实施例中,所述第一获取模块进一步用于:确定所述拍摄设备的识别信息;基于所述识别信息,确定所述拍摄设备的拍摄参数集;基于所述拍摄参数集,生成所述拍摄参数序列。In some embodiments, the first obtaining module is further configured to: determine identification information of the shooting device; determine a shooting parameter set of the shooting device based on the identification information; generate the shooting parameter set based on the shooting parameter set Describe the shooting parameter sequence.
在一些实施例中,所述第一获取模块进一步用于:从所述拍摄参数集中随机选取预设数量的拍摄参数,并基于该拍摄参数生成所述拍摄参数序列。在一些实施例中,所述拍摄参数序列包括色温参数序列。In some embodiments, the first obtaining module is further configured to: randomly select a preset number of shooting parameters from the shooting parameter set, and generate the shooting parameter sequence based on the shooting parameters. In some embodiments, the sequence of shooting parameters includes a sequence of color temperature parameters.
提取模块可以用于根据预设的提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像。The extraction module may be configured to extract multiple images or image parts in the original image as multiple extracted images according to preset extraction rules.
确定模块可以用于基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度。在一些实施例中,所述机器学习模型至少包括多个卷积神经网络单元和一个序列到序列单元;所述确定模块进一步用于:通过所述多个卷积神经网络单元中的每个分别对所述多个提取图像中的每个进行处理,获得每个提取图像对应的图像表示向量;通过所述序列到序列单元对所述图像表示向量进行处理,获得所述多个提取图像的预测拍摄参数的变化序列;基于所述预测拍摄参数的变化序列和所述预设序列确定所述匹配度。The determining module may be configured to determine, based on the multiple extracted images, the degree of matching between the multiple extracted images and the preset sequence by using a trained machine learning model. In some embodiments, the machine learning model includes at least a plurality of convolutional neural network units and a sequence-to-sequence unit; the determining module is further configured to: pass through each of the plurality of convolutional neural network units, respectively processing each of the multiple extracted images to obtain an image representation vector corresponding to each extracted image; processing the image representation vector through the sequence-to-sequence unit to obtain predictions of the multiple extracted images A change sequence of shooting parameters; the matching degree is determined based on the predicted change sequence of shooting parameters and the preset sequence.
在一些实施例中,所述机器学习模型可以通过如下方法训练得到:获取多个携带标签的训练样本,所述训练样本包括基于样本拍摄参数获取的多个样本图像帧,所述标签包括所述多个样本图像帧之间样本拍摄参数的变化关系;基于所述多个携带有标签 的训练样本训练初始机器学习模型,获得所述机器学习模型。In some embodiments, the machine learning model may be trained by the following method: acquiring a plurality of training samples carrying labels, the training samples including a plurality of sample image frames obtained based on sample shooting parameters, the labels including the The variation relationship of the sample shooting parameters between the multiple sample image frames; the initial machine learning model is trained based on the multiple training samples carrying the labels, and the machine learning model is obtained.
判断模块可以用于基于所述匹配度,判断所述原始图像的真实性;所述预设序列对应于所述客户端的拍摄设备,判断所述原始图像的真实性具体为:判断所述图像来自所述拍摄设备的真实性。The judging module can be used to judge the authenticity of the original image based on the matching degree; the preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging that the image comes from The authenticity of the photographing equipment.
本说明书实施例提供一种图像真实性判断系统,该系统应用于客户端,所述图像真实性判断系统可以包括第二获取模块、生成模块、发送模块、第三获取模块、以及上传模块。The embodiments of this specification provide an image authenticity judgment system, which is applied to a client, and the image authenticity judgment system may include a second acquisition module, a generation module, a transmission module, a third acquisition module, and an upload module.
第二获取模块可以用于获取服务器端生成并下发的拍摄参数序列。The second acquiring module may be configured to acquire the shooting parameter sequence generated and delivered by the server.
生成模块,用于基于所述拍摄参数序列生成所述原始图像。在一些实施例中,所述拍摄参数序列是所述服务器端随机生成的。在一些实施例中,所述拍摄参数序列是所述服务器端基于所述拍摄设备的拍摄参数随机生成的;所述拍摄参数对应于所述拍摄设备的识别信息。在一些实施例中,所述拍摄参数序列包括色温参数序列。A generating module, configured to generate the original image based on the shooting parameter sequence. In some embodiments, the shooting parameter sequence is randomly generated by the server. In some embodiments, the shooting parameter sequence is randomly generated by the server based on shooting parameters of the shooting device; the shooting parameters correspond to identification information of the shooting device. In some embodiments, the sequence of shooting parameters includes a sequence of color temperature parameters.
发送模块可以用于将所述原始图像发送给服务器端。The sending module can be used to send the original image to the server.
第三获取模块可以用于获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。The third obtaining module may be configured to obtain the information sent by the server and including the judgment result of the authenticity of the original image.
上传模块可以用于向服务器端上传所述客户端的拍摄设备的识别信息。The uploading module may be configured to upload the identification information of the photographing device of the client to the server.
应当理解,上述系统及其模块可以利用各种方式来实现。例如,在一些实施例中,系统及其模块可以通过硬件、软件或者软件和硬件的结合来实现。其中,硬件部分可以利用专用逻辑来实现;软件部分则可以存储在存储器中,由适当的指令执行系统,例如微处理器或者专用设计硬件来执行。本领域技术人员可以理解上述的方法和系统可以使用计算机可执行指令和/或包含在处理器控制代码中来实现,例如在诸如磁盘、CD或DVD-ROM的载体介质、诸如只读存储器(固件)的可编程的存储器或者诸如光学或电子信号载体的数据载体上提供了这样的代码。本说明书的系统及其模块不仅可以有诸如超大规模集成电路或门阵列、诸如逻辑芯片、晶体管等的半导体、或者诸如现场可编程门阵列、可编程逻辑设备等的可编程硬件设备的硬件电路实现,也可以用例如由各种类型的处理器所执行的软件实现,还可以由上述硬件电路和软件的结合(例如,固件)来实现。It should be understood that the above-described system and its modules can be implemented in various ways. For example, in some embodiments, the system and its modules may be implemented in hardware, software, or a combination of software and hardware. Wherein, the hardware part can be realized by using dedicated logic; the software part can be stored in a memory and executed by a suitable instruction execution system, such as a microprocessor or specially designed hardware. Those skilled in the art will appreciate that the methods and systems described above may be implemented using computer-executable instructions and/or embodied in processor control code, for example on a carrier medium such as a disk, CD or DVD-ROM, such as a read-only memory (firmware) ) or a data carrier such as an optical or electronic signal carrier. The system and its modules of this specification can be implemented not only by hardware circuits such as very large scale integrated circuits or gate arrays, semiconductors such as logic chips, transistors, etc., or programmable hardware devices such as field programmable gate arrays, programmable logic devices, etc. , can also be implemented by, for example, software executed by various types of processors, and can also be implemented by a combination of the above-mentioned hardware circuits and software (eg, firmware).
需要注意的是,以上对于图像真实性判断系统及其模块的描述,仅为描述方便,并不能把本说明书限制在所举实施例范围之内。可以理解,对于本领域的技术人员来说,在了解该系统的原理后,可能在不背离这一原理的情况下,对各个模块进行任意组合, 或者构成子系统与其他模块连接。诸如此类的变形,均在本说明书的保护范围之内。It should be noted that, the above description of the image authenticity judgment system and its modules is only for convenience of description, and does not limit the description to the scope of the illustrated embodiments. It can be understood that for those skilled in the art, after understanding the principle of the system, various modules may be combined arbitrarily, or a subsystem may be formed to connect with other modules without departing from the principle. Such deformations are all within the protection scope of this specification.
图2是根据本说明书的一些实施例所示的判断图像真实性的方法的示例性流程图,该方法应用于服务器端。在一些实施例中,流程200可以由图1所示的第一计算系统140实现。如图2所示,该流程200可以包括以下步骤:FIG. 2 is an exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification, and the method is applied to the server side. In some embodiments, the process 200 may be implemented by the first computing system 140 shown in FIG. 1 . As shown in FIG. 2, the process 200 may include the following steps:
步骤210,从客户端获取原始图像。在一些实施例中,该步骤210可以由第一获取模块执行。 Step 210, obtain the original image from the client. In some embodiments, this step 210 may be performed by the first acquisition module.
在一些实施例中,客户端可以是任何包含有拍摄设备的终端。例如手机、平板电脑或笔记本电脑等。在一些实施例中,原始图像可以是待检测对象的图像。例如,应用平台获取的待检测对象的图像。在一些实施例中,待检测对象可以是任何需要判断其是否为拍摄设备现场采集的对象,即判断其是拍摄设备采集的真实对象或拍摄设备采集之前提前伪造的虚假对象。示例地,待检测对象可以是待检测的人脸、掌纹或指纹等,还可以是待检测的证件,例如身份证、驾驶证等等。In some embodiments, the client can be any terminal that includes a camera device. Such as mobile phones, tablets or laptops, etc. In some embodiments, the original image may be an image of the object to be detected. For example, the image of the object to be detected obtained by the application platform. In some embodiments, the object to be detected may be any object that needs to be judged whether it is collected by the photographing device on-site, that is, it is judged whether it is a real object collected by the photographing device or a false object forged in advance before the photographing device collects. For example, the object to be detected may be a face, a palm print, a fingerprint, etc. to be detected, or a certificate to be detected, such as an ID card, a driver's license, and the like.
在一些实施例中,原始图像可以是拍摄设备录制的待检测对象的视频中包括的图像。其中,拍摄设备录制的待检测对象的视频可以是一段预设时长的视频。例如,3秒或5秒的视频。对应的,原始图像可以为3秒或5秒的视频中包括的图像。In some embodiments, the original image may be an image included in a video of the object to be detected recorded by the photographing device. The video of the object to be detected recorded by the photographing device may be a video of a preset duration. For example, a 3-second or 5-second video. Correspondingly, the original image may be an image included in a 3-second or 5-second video.
在一些实施例中,原始图像可以是拍摄设备拍摄的待检测对象的一张或多张图像。该多张图像可以是拍摄设备连续拍摄待检测对象生成的多张图像,也可以是拍摄设备按预设时间间隔拍摄待检测对象生成的多张图像。本说明书实施例并不对拍摄设备如何拍摄得到待检测对象的多张图像做具体限制。In some embodiments, the original image may be one or more images of the object to be detected captured by the capturing device. The multiple images may be multiple images generated by the photographing device continuously photographing the object to be detected, or may be multiple images generated by the photographing device photographing the object to be detected at preset time intervals. The embodiments of this specification do not specifically limit how the photographing device obtains multiple images of the object to be detected.
在本说明书中,原始图像可以包括静态的一幅或多幅照片,也可以包括视频,或者其混合。In this specification, the original image may include still one or more photographs, may also include video, or a mixture thereof.
在一些实施例中,第一获取模块可以通过多种方式获取原始图像。在一些实施例中,获取模块可以从存储设备中获取原始图像。其中,原始图像被预先生成并存储在存储设备中。例如,客户端110的拍摄设备采集待检测对象的视频和/或图像后发送到存储设备进行存储,此时获取模块可以直接从存储设备中获得原始图像。在一些实施例中,获取模块可以实时获取客户端110的拍摄设备采集待检测对象的视频和/或图像。In some embodiments, the first acquisition module may acquire the original image in various ways. In some embodiments, the acquisition module may acquire the original image from the storage device. Among them, the original image is pre-generated and stored in the storage device. For example, the photographing device of the client 110 collects the video and/or image of the object to be detected and sends it to the storage device for storage. At this time, the acquisition module can directly obtain the original image from the storage device. In some embodiments, the acquisition module may acquire in real time the video and/or image of the object to be detected acquired by the photographing device of the client 110 .
在一些实施例中,原始图像可以是由客户端基于拍摄参数序列生成的。具体的,从客户端获取原始图像可以包括:生成拍摄参数序列,将所述拍摄参数序列下发至客户端;从客户端获取原始图像,原始图像由客户端基于拍摄参数序列生成。In some embodiments, the original image may be generated by the client based on a sequence of shooting parameters. Specifically, acquiring the original image from the client may include: generating a shooting parameter sequence, and sending the shooting parameter sequence to the client; acquiring an original image from the client, where the original image is generated by the client based on the shooting parameter sequence.
在一些实施例中,拍摄参数序列可以是由多个拍摄参数组成的序列。在一些实 施例中,拍摄参数可以是指拍摄设备录制或拍摄时所使用的参数。在一些实施例中,拍摄参数可以包括色温参数、锐化程度参数、颜色饱和度参数、亮度参数、对比度参数、快门参数以及光圈参数等。对应的,当拍摄参数是色温参数时,拍摄参数序列可以是色温参数序列,即多个色温参数组成的序列。在一些实施例中,上述拍摄参数可以混合或组合使用,以生成拍摄参数序列。为简化说明为目的,本说明书实施例以拍摄参数序列为色温参数序列为例进行说明,应当知晓的,拍摄参数序列并不仅限制于该色温参数序列,例如还可以是颜色饱和度参数序列或者光圈参数序列,或者其他参数,或者各种参数的组合。本说明书实施例并不对此进行限制。In some embodiments, the sequence of shooting parameters may be a sequence of multiple shooting parameters. In some embodiments, the shooting parameters may refer to parameters used by the shooting device when recording or shooting. In some embodiments, the shooting parameters may include a color temperature parameter, a sharpening degree parameter, a color saturation parameter, a brightness parameter, a contrast parameter, a shutter parameter, an aperture parameter, and the like. Correspondingly, when the shooting parameter is a color temperature parameter, the shooting parameter sequence may be a color temperature parameter sequence, that is, a sequence composed of multiple color temperature parameters. In some embodiments, the above capture parameters may be mixed or combined to generate a sequence of capture parameters. For the purpose of simplifying the description, the embodiments of this specification take the shooting parameter sequence as the color temperature parameter sequence as an example for description. It should be known that the shooting parameter sequence is not limited to the color temperature parameter sequence, for example, it can also be a color saturation parameter sequence or an aperture. A sequence of parameters, or other parameters, or a combination of various parameters. The embodiments of the present specification do not limit this.
在一些实施例中,拍摄参数序列还可以包括时间信息和/或图像张数信息。在一些实施例中,时间信息用于反映拍摄设备采用对应的拍摄参数录制待检测对象的视频的相应时间段的信息。示例地,以上述原始图像为3s的视频中包括的图像,拍摄参数序列为色温参数序列为例,若将3s均分为3个时间段,即0-1s、1-2s以及2-3s;则时间信息可以反映采用对应的色温参数分别录制0-1s、1-2s以及2-3s时间段的视频的信息。在一些实施例中,图像张数信息用于反映拍摄设备采用对应的拍摄参数拍摄待检测对象的相应图像张数的信息。示例地,以原始图像为拍摄设备连续拍摄的15张图像,拍摄参数序列为色温参数序列为例,若将15张图像均分为3个图像段,即第1-5张图像、第5-10张图像以及第10-15张图像;则图像张数信息可以反映采用对应的色温参数分别拍摄第1-5张图像、第5-10张图像以及第10-15张图像的信息。In some embodiments, the shooting parameter sequence may further include time information and/or image number information. In some embodiments, the time information is used to reflect the information of the corresponding time period when the photographing device uses the corresponding photographing parameters to record the video of the object to be detected. For example, taking the above-mentioned original image as an image included in a 3s video and the shooting parameter sequence as a color temperature parameter sequence as an example, if the 3s is evenly divided into 3 time periods, namely 0-1s, 1-2s and 2-3s; Then the time information can reflect the information of the videos recorded in the time periods of 0-1s, 1-2s and 2-3s respectively using the corresponding color temperature parameters. In some embodiments, the information on the number of images is used to reflect the information on the number of corresponding images of the object to be detected that the photographing device uses with the corresponding photographing parameters. As an example, take the original image as 15 images continuously shot by the shooting device, and the shooting parameter sequence as the color temperature parameter sequence. 10 images and the 10th to 15th images; the number of images can reflect the information of the 1st to 5th images, the 5th to 10th images, and the 10th to 15th images, respectively, using the corresponding color temperature parameters.
在一些实施例中,服务器端可以基于客户端包含的拍摄参数集生成拍摄参数序列。具体的,生成拍摄参数序列可以包括:确定所述拍摄设备的识别信息;基于所述识别信息,确定所述拍摄设备的拍摄参数集;基于所述拍摄参数集,生成所述拍摄参数序列。In some embodiments, the server side may generate the shooting parameter sequence based on the shooting parameter set contained in the client terminal. Specifically, generating a shooting parameter sequence may include: determining identification information of the shooting device; determining a shooting parameter set of the shooting device based on the identification information; generating the shooting parameter sequence based on the shooting parameter set.
在一些实施例中,识别信息可以包括拍摄设备的型号或性能参数。在一些实施例中,第一获取模块可以基于客户端包含的操作系统确定拍摄设备的识别信息。例如客户端为手机,则可以基于手机包括的IOS或Android操作系统确定其拍摄设备对应的型号或性能参数。In some embodiments, the identification information may include the model or performance parameters of the photographing device. In some embodiments, the first obtaining module may determine the identification information of the photographing device based on an operating system included in the client. For example, if the client is a mobile phone, the model or performance parameter corresponding to the shooting device may be determined based on the IOS or Android operating system included in the mobile phone.
在一些实施例中,拍摄参数集可以是拍摄设备包含的所有或部分可设置的拍摄参数。在一些实施例中,不同型号的拍摄设备可以具备不同的拍摄参数集。仍以上述示例为例,则IOS手机与Android手机具备不同的拍摄参数集。In some embodiments, the set of shooting parameters may be all or part of the settable shooting parameters contained in the shooting device. In some embodiments, different models of photographing devices may have different sets of photographing parameters. Still taking the above example as an example, the IOS mobile phone and the Android mobile phone have different shooting parameter sets.
在一些实施例中,第一计算系统140(即服务器端)可以从拍摄参数集中随机选 取预设数量的拍摄参数,并基于该拍摄参数生成拍摄参数序列。示例地,仍以色温参数序列为例,若拍摄设备的可选色温参数范围包括400-420nm、460-470nm、568-572nm、6000-6500k、10000-12000k以及601-606nm,则第一计算系统140可以从上述6个色温参数点中选取5个色温参数点,如460-470nm(以下简称色温参数1)、568-572nm(以下简称色温参数2)、6000-6500k(以下简称色温参数3)、10000-12000k(以下简称色温参数4)以及601-606nm(以下简称色温参数5),以生成色温参数序列,例如色温参数序列可以是:色温参数3-色温参数1-色温参数5。In some embodiments, the first computing system 140 (i.e., the server side) may randomly select a preset number of shooting parameters from the shooting parameter set, and generate a shooting parameter sequence based on the shooting parameters. For example, still taking the color temperature parameter sequence as an example, if the selectable color temperature parameter ranges of the photographing device include 400-420nm, 460-470nm, 568-572nm, 6000-6500k, 10000-12000k, and 601-606nm, then the first computing system 140 You can select 5 color temperature parameter points from the above 6 color temperature parameter points, such as 460-470nm (hereinafter referred to as color temperature parameter 1), 568-572nm (hereinafter referred to as color temperature parameter 2), 6000-6500k (hereinafter referred to as color temperature parameter 3) , 10000-12000k (hereinafter referred to as color temperature parameter 4) and 601-606 nm (hereinafter referred to as color temperature parameter 5) to generate a color temperature parameter sequence, for example, the color temperature parameter sequence can be: color temperature parameter 3-color temperature parameter 1-color temperature parameter 5.
如前所述,拍摄参数序列可以包括时间信息和/或图像张数信息,因此,色温参数序列可以包括时间信息和/或图像张数信息。示例地,仍以上述原始图像为3s的视频中包括的图像,色温参数序列包括时间信息为例,则色温参数序列可以为s={1,2,3},其可以反映3s视频中的0-1s采用色温参数1进行录制,1-2s采用色温参数2进行录制,2-3s采用色温参数3进行录制。As mentioned above, the shooting parameter sequence may include time information and/or image number information, therefore, the color temperature parameter sequence may include time information and/or image number information. For example, still taking the above-mentioned original image as an image included in a 3s video and the color temperature parameter sequence including time information as an example, the color temperature parameter sequence can be s={1, 2, 3}, which can reflect the 0 in the 3s video. -1s is recorded with color temperature parameter 1, 1-2s is recorded with color temperature parameter 2, and 2-3s is recorded with color temperature parameter 3.
在一些实施例中,第一计算系统140(即服务器)可以通过网络将拍摄参数序列下发至客户端。In some embodiments, the first computing system 140 (ie, the server) may deliver the sequence of shooting parameters to the client through the network.
如前所述,原始图像可以由客户端基于拍摄参数序列生成。在一些实施例中,原始图像可以由客户端的拍摄设备基于拍摄参数序列,拍摄或录制相应的图像或视频生成。仍以上述示例为例,则客户端可以基于色温参数序列s={1,2,3},采用色温参数1录制视频的0-1s,采用色温参数2录制视频的1-2s,采用色温参数3录制视频的2-3s,录制完整的3s视频,进而将该3s视频包括的图像作为原始图像。As mentioned earlier, the raw image can be generated by the client based on the sequence of shooting parameters. In some embodiments, the original image may be generated by the client's photographing device by photographing or recording a corresponding image or video based on the photographing parameter sequence. Still taking the above example as an example, the client can use the color temperature parameter sequence s={1, 2, 3} to record 0-1s of the video with color temperature parameter 1, and record 1-2s of the video with color temperature parameter 2, and use the color temperature parameter 3 Record 2-3s of the video, record the complete 3s video, and then use the image included in the 3s video as the original image.
步骤220,根据预设提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像。在一些实施例中,该步骤220可以由提取模块执行。Step 220: Extract multiple images or image parts in the original image as multiple extracted images according to a preset extraction rule. In some embodiments, this step 220 may be performed by an extraction module.
在一些实施例中,多个提取图像可以是原始图像中的多个图像或者图像局部。在一些实施例中,多个提取图像可以是拍摄设备录制的待检测对象的视频或拍摄的待检测对象的多张图像中包括的多个图像。在一些实施例中,多个提取图像可以是多个图像中每个的图像局部。例如从每个图像中提取某个区域的图像作为图像局部。In some embodiments, the plurality of extracted images may be a plurality of images or portions of images in the original image. In some embodiments, the plurality of extracted images may be a video of the object to be detected recorded by the photographing device or a plurality of images included in a plurality of captured images of the object to be detected. In some embodiments, the plurality of extracted images may be image portions of each of the plurality of images. For example, an image of a certain area is extracted from each image as an image part.
在一些实施例中,预设提取规则可以根据实际需求进行具体设置。在一些实施例中,预设提取规则可以与原始图像的获取方式匹配。在一些实施例中,预设提取规则可以与拍摄参数序列匹配。在一些实施例中,预设提取规则可以与拍摄参数序列反映的时间信息和/或图像张数信息匹配。In some embodiments, the preset extraction rules may be specifically set according to actual requirements. In some embodiments, the preset extraction rules may match how the original images were acquired. In some embodiments, preset extraction rules may be matched to sequences of shooting parameters. In some embodiments, the preset extraction rule may be matched with time information and/or image number information reflected by the shooting parameter sequence.
仍以上述示例为例,原始图像为采用色温参数点1录制视频的0-1s、采用色温 参数点2录制视频的1-2s以及采用色温参数点3录制视频的2-3s的3s视频,则预设提取规则可以是从0-1s、1-2s以及2-3s的视频段中分别提取任意一张图像,生成多个提取图像。参考图3,图3是根据该示例示意的提取多个提取图像的示例性示意图。如图3所示,原始图像310为时长为3s的视频,则可以从0-1s、1-2s以及2-3s的视频段中分别提取最后一帧图像(灰色部分所示),作为多个提取图像320。Still taking the above example as an example, the original image is 0-1s of video recorded with color temperature parameter point 1, 1-2s of video recorded with color temperature parameter point 2, and 3s video of 2-3s recorded with color temperature parameter point 3, then The preset extraction rule may be to extract any image from the video segments of 0-1s, 1-2s, and 2-3s, respectively, to generate multiple extracted images. Referring to FIG. 3 , FIG. 3 is an exemplary schematic diagram illustrating extracting a plurality of extracted images according to this example. As shown in Figure 3, the original image 310 is a video with a duration of 3s, then the last frame of image (shown in gray) can be extracted from the video segments of 0-1s, 1-2s, and 2-3s respectively, as multiple Image 320 is extracted.
又例如,可以根据预设规则,提取一个图像的不同区域,每个区域作为一个提取图像,或者从多个图像的每个中提取多个区域作为多个提取图像。For another example, different regions of an image may be extracted according to preset rules, and each region may be used as an extracted image, or multiple regions may be extracted from each of multiple images as multiple extracted images.
步骤230,基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度。在一些实施例中,该步骤230可以由确定模块执行。Step 230: Based on the multiple extracted images, determine the degree of matching between the multiple extracted images and the preset sequence by using the trained machine learning model. In some embodiments, this step 230 may be performed by a determination module.
在一些实施例中,机器学习模型可以是预先训练好的模型。关于机器学习模型的训练过程可以参见图5及其相关描述,在此不再赘述。训练好的机器学习模型可以确定多个提取图像与预设序列的匹配度。In some embodiments, the machine learning model may be a pre-trained model. For the training process of the machine learning model, reference may be made to FIG. 5 and its related description, which will not be repeated here. A trained machine learning model can determine how well multiple extracted images match a preset sequence.
在一些实施例中,可以使用卷积神经网络模型确定每个提取图像的特征,将所输出的特征与预设序列进行比较。所使用的卷积神经网络模型可以经训练获得,训练时可以使用预设序列对应的拍摄参数获取图像,进一步提取后作为训练数据,以对应预设序列为标签,通过最优化损失函数的迭代方法进行训练。In some embodiments, a convolutional neural network model can be used to determine the features of each extracted image, and the output features can be compared to a preset sequence. The convolutional neural network model used can be obtained by training. During training, images can be obtained by using the shooting parameters corresponding to the preset sequence, and further extracted as training data, with the corresponding preset sequence as the label, through the iterative method of optimizing the loss function. to train.
在一些实施例中,可以使用卷积神经网络单元与序列到序列(Seq2Seq)单元结合的方式作为机器学习模型,具体细节可以参见图4及其相关描述,在此不再赘述。In some embodiments, a combination of a convolutional neural network unit and a sequence-to-sequence (Seq2Seq) unit may be used as a machine learning model. For details, see FIG. 4 and related descriptions, which will not be repeated here.
预设序列是用于对图像进行特征识别的条件值,它可以是多个值、多个向量或其他数据组成的序列,也可以将一个值或其他表示形式用序列统一代称。A preset sequence is a conditional value used for feature recognition of an image. It can be a sequence composed of multiple values, multiple vectors, or other data, or a value or other representation can be uniformly referred to as a sequence.
在一些实施例中,预设序列对应于客户端的拍摄设备。在一些实施例中,预设序列可以是对应于拍摄参数序列中参数的变化的序列。示例地,仍以上述示例为例,若拍摄参数序列为色温参数序列s={1,2,3},则预设序列可以是由色温参数1至2的变化、以及色温参数2至3的变化构成的变化序列。In some embodiments, the preset sequence corresponds to the client's camera device. In some embodiments, the preset sequence may be a sequence corresponding to changes in parameters in the sequence of shooting parameters. Illustratively, still taking the above example as an example, if the shooting parameter sequence is the color temperature parameter sequence s={1, 2, 3}, the preset sequence may be the change of color temperature parameters 1 to 2 and the change of color temperature parameters 2 to 3. A sequence of changes made up of changes.
在一些实施例中,预设序列中参数的变化的值可以由预先设置的编码信息确定。例如,预先设置的编码信息包括色温参数1变化至2由编码字符a表示,色温参数2变化至3由编码字符b表示。则预设序列可以为s′={a,b}。In some embodiments, the changed values of the parameters in the preset sequence may be determined by preset encoding information. For example, the preset coding information includes that the change of color temperature parameter 1 to 2 is represented by the code character a, and the change of the color temperature parameter 2 to 3 is represented by the code character b. Then the preset sequence can be s'={a,b}.
又例如,预设序列可能是一个代码值,用于区分拍摄设备。预设序列可以有多种表示形式,其作用没有本质区别,本说明书对此不作限制。As another example, the preset sequence may be a code value used to distinguish photographing equipment. The preset sequence may have various representation forms, and its functions are not substantially different, which is not limited in this specification.
在一些实施例中,匹配度可以反映多个提取图像与预设序列之间的相似度。在 一些实施例中,匹配度可以反映多个提取图像的预测拍摄参数的变化序列与预设序列之间的相似度,即多个提取图像的预测拍摄参数的变化序列与拍摄参数序列中参数的变化序列之间的相似度。可以理解的,匹配度的值越大,则两者的相似度越大,原始图像真实的可能性越大。In some embodiments, the matching degree may reflect the similarity between the plurality of extracted images and the preset sequence. In some embodiments, the matching degree may reflect the similarity between the sequence of changes of the predicted shooting parameters of the multiple extracted images and the preset sequence, that is, the difference between the sequence of changes of the predicted shooting parameters of the multiple extracted images and the parameters in the sequence of shooting parameters Similarity between sequences of changes. It can be understood that the greater the value of the matching degree, the greater the similarity between the two, and the greater the possibility that the original image is real.
对于匹配度的获取,可以通过多种变换的方式,例如将预设序列同时输入机器学习模型以直接输出匹配度等,这些方法没有本质的区别,本说明书对此不作限制。For the acquisition of the matching degree, various transformation methods can be used, such as inputting the preset sequence into the machine learning model at the same time to directly output the matching degree, etc. There is no essential difference between these methods, which is not limited in this specification.
根据以上描述可知,在一些实施例中,通过验证图像与拍摄设备的关联关系,可以有效地防止攻击者绕过拍摄设备直接上传预制的图像。特别是通过对于图像抽取不同部分基于机器学习模型进行分析,而非简单地基于图像附带的参数进行验证,因此即使攻击者劫持了拍摄设备并获取验证参数,也无法简单地生成可供验证的数据,大大提高了防御强度。As can be seen from the above description, in some embodiments, by verifying the association relationship between the image and the photographing device, an attacker can effectively prevent an attacker from bypassing the photographing device and directly uploading a prefabricated image. In particular, by analyzing different parts of the image extraction based on machine learning models, rather than simply verifying based on the parameters attached to the image, even if the attacker hijacks the shooting device and obtains the verification parameters, it is impossible to simply generate data for verification. , greatly improving the defense strength.
进一步地,在一些实施列中,通过生成和下发拍摄参数序列至客户端,进而客户端的拍摄设备生成相应的原始图像,因此,由拍摄设备真实拍摄的原始图像的拍摄参数必然与拍摄参数序列中的拍摄参数一致。由于预先准备的虚假图像不可能包含完全一致的拍摄参数,进而防止攻击者劫持拍摄设备,通过虚假图像完成身份认证。由于本说明书一些实施例的拍摄参数序列为随机下发,且即发即用,攻击者也无法提前通过该拍摄设备制作虚假图像,极大提高了现场验证的可靠性。Further, in some implementations, by generating and delivering the shooting parameter sequence to the client, the shooting device of the client generates the corresponding original image. Therefore, the shooting parameters of the original image actually shot by the shooting device must be the same as the shooting parameter sequence. The shooting parameters are the same. Since the pre-prepared fake images cannot contain exactly the same shooting parameters, the attacker can prevent the attacker from hijacking the shooting equipment and complete the identity authentication through the fake images. Since the shooting parameter sequence in some embodiments of this specification is randomly issued and ready to use, an attacker cannot make false images through the shooting device in advance, which greatly improves the reliability of on-site verification.
另一方面,本说明书实施例采用预测拍摄参数的变化序列与拍摄参数序列中参数的变化序列进行比对确定匹配度,即采用参数的变化进行比对确定匹配度,可以消除环境对具体参数值的确定影响,例如,以色温参数为例,环境(例如光线)会影响色温参数的确定,提高了后续通过匹配度确定真实性的准确度,进而提高身份认证的准确性。On the other hand, in the embodiment of the present specification, the change sequence of the predicted shooting parameters is compared with the change sequence of the parameters in the shooting parameter sequence to determine the matching degree, that is, the matching degree is determined by comparing the changes of the parameters, which can eliminate the influence of the environment on the specific parameter value. For example, taking the color temperature parameter as an example, the environment (such as light) will affect the determination of the color temperature parameter, which improves the accuracy of subsequent determination of authenticity through matching, thereby improving the accuracy of identity authentication.
步骤240,基于所述匹配度,判断所述原始图像的真实性。该步骤240可以由判断模块执行。 Step 240, based on the matching degree, determine the authenticity of the original image. This step 240 may be performed by a judgment module.
在一些实施例中,判断原始图像的真实性具体为:判断图像来自拍摄设备的真实性。根据上述步骤210的描述可知,原始图像来自拍摄设备即为真实图像,相反,则为虚假图像。In some embodiments, judging the authenticity of the original image is specifically: judging the authenticity of the image from a photographing device. According to the description of the above step 210, it can be known that the original image from the photographing device is a real image, and on the contrary, it is a fake image.
在一些实施例中,判断模块可以基于匹配度,判断原始图像的真实性。例如,当匹配度大于预设阈值时,原始图像为真实图像。In some embodiments, the judgment module may judge the authenticity of the original image based on the matching degree. For example, when the matching degree is greater than a preset threshold, the original image is a real image.
在一些实施例中,当判断模块判断原始图像为虚假图像时,第一计算系统140可以发送相关指令到客户端110,以终止客户端110的进一步操作(例如注册应用平 台)。In some embodiments, when the determination module determines that the original image is a fake image, the first computing system 140 may send a relevant instruction to the client 110 to terminate further operations of the client 110 (such as registering the application platform).
本步骤可以由其他执行主体执行,或者以其他变换的方式进行,对本说明书的技术方案没有实质影响。This step may be performed by other executive bodies, or performed in other transforming manners, and has no substantial impact on the technical solutions of this specification.
图4是根据本说明书的一些实施例所示的机器学习模型的示例性结构示意图。FIG. 4 is a schematic diagram of an exemplary structure of a machine learning model according to some embodiments of the present specification.
如图4所示,机器学习模型400可以至少包括多个卷积神经网络单元410和序列到序列单元420。所述多个卷积神经网络单元410中的每个可以用于对多个提取图像中的每个进行处理,获得每个提取图像对应的图像表示向量。以图4进行说明,多个卷积神经网络单元410中的每个可以对上述步骤240中提取的多个提取图像(例如提取图像1至提取图像n)中的每个进行处理,得到该提取图像的图像表示向量。在一些实施例中,卷积神经网络单元410可以采用包括基础卷积层4101和全连接层4102的常规卷积神经网络,例如,LeNet、AlexNet、GoogLeNet等等。As shown in FIG. 4 , the machine learning model 400 may include at least a plurality of convolutional neural network units 410 and a sequence-to-sequence unit 420 . Each of the plurality of convolutional neural network units 410 may be configured to process each of the plurality of extracted images to obtain an image representation vector corresponding to each of the extracted images. 4, each of the plurality of convolutional neural network units 410 may process each of the plurality of extracted images (eg, extracted image 1 to extracted image n) extracted in the above step 240 to obtain the extracted image. The image representation vector of the image. In some embodiments, the convolutional neural network unit 410 may employ a conventional convolutional neural network including a base convolutional layer 4101 and a fully connected layer 4102, eg, LeNet, AlexNet, GoogLeNet, and the like.
序列到序列单元420可以对所述图像表示向量进行处理,获得多个提取图像的预测拍摄参数的变化序列。具体地,序列到序列单元420可以对多个卷积神经网络单元410中的每个全连接层4102输出的图像表示向量进行处理,得到多个提取图像的预测拍摄参数的变化序列。The sequence-to-sequence unit 420 may process the image representation vector to obtain a sequence of changes of predicted shooting parameters of a plurality of extracted images. Specifically, the sequence-to-sequence unit 420 may process the image representation vector output by each fully connected layer 4102 in the multiple convolutional neural network units 410 to obtain a sequence of changes of predicted shooting parameters of multiple extracted images.
在一些实施例中,多个提取图像的预测拍摄参数的变化序列可以是指多个提取图像中每个提取图像之间的预测拍摄参数的变化构成的序列。在一些实施例中,预测拍摄参数与拍摄参数序列中包括的拍摄参数匹配。例如,拍摄参数序列为色温参数序列,则预测拍摄参数为色温参数。In some embodiments, the sequence of changes in the predicted shooting parameters of the plurality of extracted images may refer to a sequence formed by changes of the predicted shooting parameters between each of the multiple extracted images. In some embodiments, the predicted shot parameters match shot parameters included in the sequence of shot parameters. For example, if the shooting parameter sequence is a color temperature parameter sequence, the predicted shooting parameter is a color temperature parameter.
示例地,仍以上述多个提取图像为从0-1s、1-2s以及2-3s的视频段中分别提取的最后一帧图像为例,若该最后一帧图像分别为提取图像1、提取图像2以及提取图像3,拍摄参数序列为色温参数序列,则预测拍摄参数的变化序列可以是提取图像1与提取图像2之间预测色温参数的变化、和提取图像2与提取图像3之间预测色温参数的变化构成的序列。例如,序列到序列单元420输出的预测拍摄参数的变化序列可以为H={a,b},由上述示例可知,a表示预测的色温参数1变化至2,b表示预测的色温参数2变化至3。Illustratively, still take the above-mentioned multiple extracted images as the last frame image extracted from the video segments of 0-1s, 1-2s and 2-3s respectively, if the last frame of image is extracted image 1, extracted Image 2 and extracted image 3, the shooting parameter sequence is a color temperature parameter sequence, then the predicted shooting parameter change sequence can be the predicted color temperature parameter change between the extracted image 1 and the extracted image 2, and the predicted change between the extracted image 2 and the extracted image 3. A sequence of changes in the color temperature parameter. For example, the change sequence of the predicted shooting parameters output by the sequence-to-sequence unit 420 may be H={a, b}. It can be seen from the above example that a represents the change of the predicted color temperature parameter 1 to 2, and b represents the change of the predicted color temperature parameter 2 to 3.
由于卷积神经网络单元410输入的提取图像可以是从拍摄设备采集的原始图像中提取的,拍摄设备采集时所处位置的环境因素(例如光线明暗)可能会影响原始图像的颜色分布,由于一些拍摄参数(例如色温参数)反映图像的颜色分布,因此可能降低机器学习模型的预测准确性。Since the extracted image input by the convolutional neural network unit 410 may be extracted from the original image collected by the photographing device, the environmental factors (such as light and shade) of the location where the photographing device is collected may affect the color distribution of the original image, due to some Shooting parameters (such as color temperature parameters) reflect the color distribution of the image and therefore may reduce the prediction accuracy of the machine learning model.
在一些实施例中,为了解决上述问题,可以利用对比图像帧调节机器学习模型的参数,使得机器学习模型能准确得到多个提取图像的预测拍摄参数的变化序列,进而降低环境因素对机器学习模型400的影响。具体的,该对比图像帧可以是指定拍摄参数下拍摄的一帧图像,通过将该帧图像输入至卷积神经网络单元410得到其图像表示向量,进而基于图像表示向量与指定拍摄参数的向量之间的差异调整卷积神经网络单元410的参数,直至获得的图像表示向量与指定拍摄参数值相同。In some embodiments, in order to solve the above problems, the parameters of the machine learning model can be adjusted by using the comparison image frames, so that the machine learning model can accurately obtain the change sequence of the predicted shooting parameters of multiple extracted images, thereby reducing the environmental factors affecting the machine learning model. 400 impact. Specifically, the comparison image frame can be a frame of image captured under the specified shooting parameters, and the image representation vector is obtained by inputting the frame image into the convolutional neural network unit 410, and then the image representation vector is obtained based on the relationship between the image representation vector and the vector of the specified shooting parameters. The difference between the parameters of the convolutional neural network unit 410 is adjusted until the obtained image representation vector is the same as the specified shooting parameter value.
图5是根据本说明书的一些实施例所示的训练机器学习模型的示例性流程图。如前所述,机器学习模型400可以是由基础卷积层4101、全连接层4102以及序列到序列单元420构建的模型。在一些实施例中,该流程500可以包括以下步骤:FIG. 5 is an exemplary flowchart of training a machine learning model according to some embodiments of the present specification. As previously mentioned, the machine learning model 400 may be a model constructed by the base convolutional layer 4101 , the fully connected layer 4102 , and the sequence-to-sequence unit 420 . In some embodiments, the process 500 may include the following steps:
步骤510,获取多个携带标签的训练样本,所述训练样本包括基于样本拍摄参数获取的多个样本图像帧,所述标签包括所述多个样本图像帧之间样本拍摄参数的变化关系。Step 510: Acquire a plurality of training samples carrying labels, where the training samples include a plurality of sample image frames obtained based on the sample shooting parameters, and the labels include a change relationship of the sample shooting parameters among the plurality of sample image frames.
在一些实施例中,训练样本可以是输入至初始机器学习模型中用于训练机器学习模型的数据。在一些实施例中,训练样本可以包括基于样本拍摄参数获取的多个样本图像帧。示例地,仍以样本拍摄参数为色温参数,色温参数包括色温参数1至6为例,则其中一个训练样本可以是基于色温参数1获取的样本图像帧1、基于色温参数2获取的样本图像帧2、以及基于色温参数5获取的样本图像帧3。In some embodiments, the training samples may be data input into the initial machine learning model for training the machine learning model. In some embodiments, the training samples may include a plurality of sample image frames acquired based on sample capture parameters. For example, still taking the sample shooting parameters as the color temperature parameters, and the color temperature parameters including color temperature parameters 1 to 6 as an example, one of the training samples may be the sample image frame 1 obtained based on the color temperature parameter 1, and the sample image frame obtained based on the color temperature parameter 2. 2. The sample image frame 3 obtained based on the color temperature parameter 5.
在一些实施例中,标签可以包括多个样本图像帧之间样本拍摄参数的变化关系。在一些实施例中,标签可以是多个样本图像帧之间样本拍摄参数的变化序列。仍以上述示例为例,则标签可以为c={c,d},其中c为表征样本图像帧1和样本图像帧2之间由色温参数1至2的变化,d表示样本图像帧2和样本图像帧3之间由色温参数2至5的变化。In some embodiments, the label may include the variation relationship of the sample capture parameters among the plurality of sample image frames. In some embodiments, the label may be a sequence of changes in sample capture parameters between multiple sample image frames. Still taking the above example as an example, the label can be c={c, d}, where c represents the change between the sample image frame 1 and the sample image frame 2 by the color temperature parameters 1 to 2, and d represents the sample image frame 2 and the sample image frame 2. Variation between sample image frames 3 by color temperature parameters 2 to 5.
步骤520,基于所述多个携带有标签的训练样本训练初始机器学习模型,获得所述机器学习模型。Step 520: Train an initial machine learning model based on the plurality of labeled training samples to obtain the machine learning model.
根据图4的相关描述可知,一些拍摄参数(例如色温参数)反映图像的颜色分布。其对图像的影响是全局性的,例如同一张提取图像在不同感受野下的颜色分布应该近似或相同。而图4所示的机器学习模型400中的卷积神经网络单元410会偏重于图像中的轮廓信息,因此,在一些实施例中,可以对上述图4示意的机器学习模型400进行改进,使得其对全局特征进行识别。According to the related description in FIG. 4 , some shooting parameters (such as color temperature parameters) reflect the color distribution of the image. Its impact on the image is global, for example, the color distribution of the same extracted image under different receptive fields should be similar or the same. However, the convolutional neural network unit 410 in the machine learning model 400 shown in FIG. 4 will focus on the contour information in the image. Therefore, in some embodiments, the machine learning model 400 shown in FIG. 4 may be improved, so that It identifies global features.
如图6所示,在图4示意的机器学习模型400的基础上,构建的机器学习模型 600还可以包括第一采样卷积层4103和第二采样卷积层4104。在一些实施例中,可以在训练时在机器学习模型400中增加第一采样卷积层4103和第二采样卷积层4104。具体的,卷积神经网络单元410的基础卷积层4101分别连接第一采样卷积层4103和第二采样卷积层4104。As shown in FIG. 6 , on the basis of the machine learning model 400 illustrated in FIG. 4 , the constructed machine learning model 600 may further include a first sampling convolution layer 4103 and a second sampling convolution layer 4104. In some embodiments, a first sampled convolutional layer 4103 and a second sampled convolutional layer 4104 may be added to the machine learning model 400 during training. Specifically, the basic convolution layer 4101 of the convolutional neural network unit 410 is connected to the first sampling convolution layer 4103 and the second sampling convolution layer 4104 respectively.
通过该第一采样卷积层4103和第二采样卷积层4104可以在机器学习模型600的训练过程中调节机器学习模型600的参数,进而确保卷积神经网络单元410对于同一张提取图像在不同感受野下的颜色分布近似或相同。避免卷积神经网络单元410偏重于提取图像中的轮廓信息(例如偏重于提取图像中的物体轮廓),强化卷积神经网络单元410对全局特征的识别能力,进而提高卷积神经网络单元410对提取图像的识别能力。Through the first sampling convolution layer 4103 and the second sampling convolution layer 4104, the parameters of the machine learning model 600 can be adjusted during the training process of the machine learning model 600, thereby ensuring that the convolutional neural network unit 410 for the same extracted image in different The color distribution under the receptive field is similar or the same. Avoid the convolutional neural network unit 410 focusing on extracting contour information in the image (for example, focusing on extracting the outline of objects in the image), strengthen the convolutional neural network unit 410's ability to recognize global features, and then improve the convolutional neural network unit 410. Recognition ability to extract images.
在一些实施例中,第一采样卷积层4103和第二采样卷积层4104可以为空洞卷积层。在一些实施例中,第一采样卷积层4103和第二采样卷积层4104的卷积核大小相同。例如,第一采样卷积层4103和第二采样卷积层4104的卷积核大小均为3*3。在一些实施例中,第一采样卷积层4103和第二采样卷积层4104的卷积核的卷积采样点的间距不同。例如第一采样卷积层4103的采样间距为0,第二采样卷积层4104的采样间距为2。又例如第一采样卷积层4103的采样间距为2,第二采样卷积层4104的采样间距为0。In some embodiments, the first sampled convolutional layer 4103 and the second sampled convolutional layer 4104 may be atrous convolutional layers. In some embodiments, the convolution kernels of the first sampled convolutional layer 4103 and the second sampled convolutional layer 4104 have the same size. For example, the size of the convolution kernel of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 is both 3*3. In some embodiments, the convolution sampling points of the convolution kernels of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 have different spacings. For example, the sampling interval of the first sampling convolution layer 4103 is 0, and the sampling interval of the second sampling convolution layer 4104 is 2. For another example, the sampling interval of the first sampling convolution layer 4103 is 2, and the sampling interval of the second sampling convolution layer 4104 is 0.
当第一采样卷积层4103和第二采样卷积层4104的采样间距不相同时,对应的,第一采样卷积层4103和第二采样卷积层4104对相同提取图像的感受野不相同。示例地,仍以上述第一采样卷积层4103的采样间距为0,第二采样卷积层4104的采样间距为2为例,由于第二采样卷积层4104的采样间距更大,因此第二采样卷积层4103的感受野大于第二采样卷积层4104的感受野,通过第二采样卷积层4104获取更广的感受野,更好的利用了图像的全局特征。When the sampling intervals of the first sampling convolution layer 4103 and the second sampling convolution layer 4104 are different, correspondingly, the first sampling convolution layer 4103 and the second sampling convolution layer 4104 have different receptive fields for the same extracted image . By way of example, still taking the sampling interval of the first sampling convolutional layer 4103 as 0 and the sampling interval of the second sampling convolutional layer 4104 as 2 as an example, since the sampling interval of the second sampling convolutional layer 4104 is larger, the sampling interval of the second sampling convolutional layer 4104 is larger. The receptive field of the two-sampling convolutional layer 4103 is larger than that of the second sampling convolutional layer 4104, and a wider receptive field is obtained through the second sampling convolutional layer 4104, which makes better use of the global features of the image.
在一些实施例中,可以基于多个携带有标签的训练样本对初始机器学习模型进行端到端的训练,获得训练好的机器学习模型(如机器学习模型400和600)。具体的,可以不断地调整初始机器学习模型的参数,以减小各个训练样本对应的损失函数值,使得损失函数值满足预设条件。例如,损失函数值收敛,或损失函数值小于预设值。当损失函数满足预设条件时,模型训练完成,得到训练好的机器学习模型。In some embodiments, an initial machine learning model may be trained end-to-end based on a plurality of labeled training samples to obtain a trained machine learning model (eg, machine learning models 400 and 600 ). Specifically, the parameters of the initial machine learning model can be continuously adjusted to reduce the loss function value corresponding to each training sample, so that the loss function value satisfies the preset condition. For example, the loss function value converges, or the loss function value is smaller than a preset value. When the loss function satisfies the preset conditions, the model training is completed, and the trained machine learning model is obtained.
在一些实施例中,各个训练样本对应的损失函数值可以通过以下过程确定:通过初始机器学习模型处理多个样本图像帧,获得多个样本图像帧之间预测样本拍摄参数的变化关系,基于预测样本拍摄参数的变化关系和标签中的样本拍摄参数的变化关系的 差异,确定该训练样本对应的损失函数值。In some embodiments, the loss function value corresponding to each training sample may be determined through the following process: processing a plurality of sample image frames through an initial machine learning model, obtaining the change relationship of the predicted sample shooting parameters among the plurality of sample image frames, The difference between the change relationship of the sample shooting parameters and the change relationship of the sample shooting parameters in the label determines the loss function value corresponding to the training sample.
如前所述,机器学习模型600可以是由基础卷积层4101、全连接层4102、第一采样卷积层4103、第二采样卷积层4104以及序列到序列单元420构建的模型。该模型600除了通过上述方式确定训练样本对应的损失函数值以外,还可以基于第一采样卷积层4103和第二采样卷积层4104确定每个提取图像在不同感受野下的颜色分布差异,基于该差异确定训练样本对应的损失函数。As previously mentioned, the machine learning model 600 may be a model constructed by a base convolutional layer 4101 , a fully connected layer 4102 , a first sampled convolutional layer 4103 , a second sampled convolutional layer 4104 , and a sequence-to-sequence unit 420 . In addition to determining the loss function value corresponding to the training sample in the above-mentioned manner, the model 600 can also determine the color distribution difference of each extracted image under different receptive fields based on the first sampling convolution layer 4103 and the second sampling convolution layer 4104, The loss function corresponding to the training sample is determined based on the difference.
在一些实施例中,可以基于第一采样卷积层4103和第二采样卷积层4104输出的特征向量计算两者的KL散度,确定每个提取图像在不同感受野下的颜色分布差异,将其作为卷积神经网络单元410的约束参数,进而通过调节该约束参数确保卷积神经网络单元410对于同一张提取图像在不同感受野下的颜色分布近似。例如,调节约束参数为最小数值0。In some embodiments, the KL divergence of the first sampling convolutional layer 4103 and the second sampling convolutional layer 4104 can be calculated based on the feature vectors output by the two to determine the color distribution difference of each extracted image under different receptive fields, It is taken as the constraint parameter of the convolutional neural network unit 410, and then the convolutional neural network unit 410 can ensure that the color distribution of the same extracted image under different receptive fields is approximate by adjusting the constraint parameter. For example, adjust the constraint parameter to a minimum value of 0.
具体的,可以通过以下公式(1)确定卷积神经网络单元410的约束参数D KLSpecifically, the constraint parameter D KL of the convolutional neural network unit 410 can be determined by the following formula (1):
D KL=argminKL(P conv1(x)||P conv2(x))        (1) D KL =argminKL(P conv1 (x)||P conv2 (x)) (1)
其中,KL(P conv1(x)||P conv2(x))表示对P conv1(x)和P conv2(x)做KL散度计算,P conv1(x)表示第一采样卷积层输出的特征向量,P conv2(x)表示第二采样卷积层输出的特征向量,argmin表示对KL散度计算值取最小。最优地,约束参数为0,此时第一采样卷积层4103和第二采样卷积层4104输出的特征向量之间的差异最小,特征分布最相似,因此同一张提取图像在不同感受野下的颜色分布相同。 Among them, KL(P conv1 (x)||P conv2 (x)) represents the KL divergence calculation for P conv1 (x) and P conv2 (x), and P conv1 (x) represents the output of the first sampling convolutional layer. Feature vector, P conv2 (x) represents the feature vector output by the second sampling convolution layer, and argmin represents the minimum value of the KL divergence calculation. Optimally, the constraint parameter is 0. At this time, the difference between the feature vectors output by the first sampling convolution layer 4103 and the second sampling convolution layer 4104 is the smallest, and the feature distribution is the most similar. Therefore, the same extracted image has different receptive fields. The color distribution below is the same.
可以理解的,当上述基于标签构造的损失函数满足预设条件时,模型训练完成,可以得到训练好的机器学习模型400,或者基于标签和约束参数构造的损失函数均满足预设条件时,模型训练完成,得到训练好的机器学习模型600。It can be understood that when the above-mentioned loss function constructed based on labels satisfies the preset conditions, the model training is completed, and the trained machine learning model 400 can be obtained, or when the loss functions constructed based on labels and constraint parameters satisfy the preset conditions, the model After the training is completed, the trained machine learning model 600 is obtained.
在一些实施例中,卷积神经网络单元410和机器学习模型600可以进行联合训练,优化基础卷积层4101的参数,使基础卷积层4101所生成的特征向量更好地反映图像中与拍摄参数相关的整体特征,从而提高机器学习模型600的识别效果。In some embodiments, the convolutional neural network unit 410 and the machine learning model 600 may perform joint training to optimize the parameters of the basic convolutional layer 4101, so that the feature vector generated by the basic convolutional layer 4101 can better reflect the in-image and the shooting parameter-related overall features, thereby improving the recognition effect of the machine learning model 600 .
图7是根据本说明书的一些实施例所示的判断图像真实性的方法的另一示例性流程图,该方法应用于客户端。在一些实施例中,流程700可以由图1所示的客户端110实现。如图7所示,该流程700可以包括以下步骤:FIG. 7 is another exemplary flowchart of a method for judging the authenticity of an image according to some embodiments of the present specification, and the method is applied to a client. In some embodiments, the process 700 may be implemented by the client 110 shown in FIG. 1 . As shown in FIG. 7, the process 700 may include the following steps:
步骤710,获取服务器端生成并下发的拍摄参数序列。Step 710: Obtain the shooting parameter sequence generated and delivered by the server.
在一些实施例中,该步骤710可以由第二获取模块执行。In some embodiments, this step 710 may be performed by a second acquisition module.
在一些实施例中,拍摄参数序列可以是服务器端随机生成的。在一些实施例中, 拍摄参数序列可以是服务器端基于拍摄设备的拍摄参数集随机生成的;所述拍摄参数集对应于拍摄设备的识别信息。在一些实施例中,拍摄参数序列可以包括色温参数序列。关于步骤710的具体细节可以参见上述步骤210及其相关描述。In some embodiments, the shooting parameter sequence may be randomly generated on the server side. In some embodiments, the shooting parameter sequence may be randomly generated by the server based on a shooting parameter set of a shooting device; the shooting parameter set corresponds to the identification information of the shooting device. In some embodiments, the sequence of shooting parameters may include a sequence of color temperature parameters. For the specific details of step 710, reference may be made to the above-mentioned step 210 and its related description.
步骤720,基于所述拍摄参数序列生成所述原始图像。Step 720: Generate the original image based on the shooting parameter sequence.
在一些实施例中,该步骤720可以由生成模块执行。In some embodiments, this step 720 may be performed by a generation module.
关于步骤720的具体细节可以参见上述步骤210及其相关描述。For the specific details of step 720, reference may be made to the above-mentioned step 210 and its related description.
步骤730,将所述原始图像发送给服务器端。Step 730: Send the original image to the server.
在一些实施例中,该步骤730可以由发送模块执行。In some embodiments, this step 730 may be performed by a sending module.
在一些实施例中,客户端110可以通过网络将原始图像发送给服务器端(如第一计算系统140)。关于原始图像的具体细节可以参见上述步骤210,在此不再赘述。In some embodiments, the client 110 may send the original image to the server (eg, the first computing system 140 ) through the network. For the specific details of the original image, reference may be made to the foregoing step 210, and details are not repeated here.
步骤740,获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。Step 740: Obtain the information sent by the server that includes the result of the judgment on the authenticity of the original image.
在一些实施例中,客户端110可以通过网络获取所述服务器端(如第一计算系统140)发送的、包含对所述原始图像的真实性的判断结果的信息。在一些实施例中,原始图像的真实性的判断结果的信息可以包括原始图像是否真实的判断结果。In some embodiments, the client 110 may obtain the information sent by the server (eg, the first computing system 140 ) through the network, and including the judgment result of the authenticity of the original image. In some embodiments, the information of the judging result of the authenticity of the original image may include the judging result of whether the original image is authentic.
在一些实施例中,客户端还可以获取基于判断结果发送的原始图像的验证信息。例如人脸或证件是否符合要求。在一些实施例中,客户端还可以获取服务器基于判断结果发送的相关指令。例如,当判断结果为原始图像是虚假图像时,服务器可以发送终止指令,终止客户端的进一步操作(例如注册应用平台)。In some embodiments, the client may also obtain verification information of the original image sent based on the judgment result. For example, whether the face or documents meet the requirements. In some embodiments, the client may also obtain relevant instructions sent by the server based on the judgment result. For example, when it is determined that the original image is a fake image, the server may send a termination instruction to terminate further operations of the client (for example, registering the application platform).
以上,本说明书实施例从服务器端和客户端各自的角度对判断图像真实性的方法进行说明。以下,本说明书实施例从服务器端和客户端的整体角度对判断图像真实性的方法进行说明。In the above, the embodiments of this specification describe the method for judging the authenticity of an image from the perspectives of the server side and the client side. Hereinafter, the embodiments of the present specification describe the method for judging the authenticity of an image from the overall perspective of the server side and the client side.
图8是根据本说明书的一些实施例所示的服务器端和客户端的交互示意图。FIG. 8 is a schematic diagram of interaction between a server and a client according to some embodiments of the present specification.
如图8所示,该交互示意图800中服务器端和客户端的交互包括并不限于:服务器端从客户端获取设备型号,进而服务器端可以基于客户端的设备型号确定其拍摄设备包含的拍摄参数集,基于该拍摄参数集生成拍摄参数序列。服务器端将拍摄参数序列下发至客户端,客户端的拍摄设备基于该拍摄参数序列生成原始图像。服务器端从客户端获取原始图像,通过上述步骤220至240的方法判断原始图像的真实性,并发送包含有判断结果的信息至客户端。As shown in FIG. 8 , the interaction between the server and the client in the schematic interaction diagram 800 includes but is not limited to: the server obtains the device model from the client, and then the server can determine the shooting parameter set included in the shooting device based on the device model of the client, A shooting parameter sequence is generated based on the shooting parameter set. The server sends the shooting parameter sequence to the client, and the shooting device of the client generates the original image based on the shooting parameter sequence. The server side obtains the original image from the client side, judges the authenticity of the original image through the methods of steps 220 to 240 above, and sends information including the judgment result to the client side.
本说明书实施例还提供一种判断图像真实性的装置,所述装置包括处理器以及 存储器;所述存储器用于存储指令,所述处理器用于执行所述指令,以实现如前任一项所述判断图像真实性的方法对应的操作。An embodiment of the present specification further provides an apparatus for judging the authenticity of an image, the apparatus includes a processor and a memory; the memory is used for storing instructions, and the processor is used for executing the instructions, so as to achieve the above-mentioned items The operation corresponding to the method of judging the authenticity of the image.
本说明书实施例还提供一种计算机可读存储介质,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行如前任一项所述判断图像真实性的方法对应的操作。Embodiments of this specification also provide a computer-readable storage medium, where the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the method corresponding to the method for judging the authenticity of an image described in any preceding item. operate.
本说明书实施例可能带来的有益效果包括但不限于:(1)通过生成和下发拍摄参数序列至客户端,进而客户端的拍摄设备生成相应的原始图像,因此,由拍摄设备真实拍摄的原始图像的拍摄参数必然与拍摄参数序列中的拍摄参数一致。由于预先准备的虚假图像不可能包含完全一致的拍摄参数,进而防止攻击者劫持拍摄设备,通过虚假图像完成身份认证;(2)拍摄参数序列为随机下发,且即发即用,攻击者也无法提前通过该拍摄设备制作虚假图像,极大提高了现场验证的可靠性;(3)强化了卷积神经网络单元对图像的全局特征的识别能力,提高卷积神经网络单元对提取图像的识别能力,进而提高机器学习模型的预测准确性。需要说明的是,不同实施例可能产生的有益效果不同,在不同的实施例里,可能产生的有益效果可以是以上任意一种或几种的组合,也可以是其他任何可能获得的有益效果。The possible beneficial effects of the embodiments of this specification include, but are not limited to: (1) By generating and delivering a shooting parameter sequence to the client, the shooting device of the client generates a corresponding original image. Therefore, the original image actually shot by the shooting device The shooting parameters of the image must be consistent with the shooting parameters in the shooting parameter sequence. Since the pre-prepared fake images cannot contain exactly the same shooting parameters, the attacker can prevent the attacker from hijacking the shooting equipment and complete the identity authentication through the fake image; (2) The shooting parameter sequence is randomly issued, and it can be used immediately, and the attacker can also It is impossible to make false images through the shooting equipment in advance, which greatly improves the reliability of on-site verification; (3) The recognition ability of the convolutional neural network unit to the global features of the image is strengthened, and the recognition of the extracted image by the convolutional neural network unit is improved. ability to improve the prediction accuracy of machine learning models. It should be noted that different embodiments may have different beneficial effects, and in different embodiments, the possible beneficial effects may be any one or a combination of the above, or any other possible beneficial effects.
上文已对基本概念做了描述,显然,对于本领域技术人员来说,上述详细披露仅仅作为示例,而并不构成对本说明书的限定。虽然此处并没有明确说明,本领域技术人员可能会对本说明书进行各种修改、改进和修正。该类修改、改进和修正在本说明书中被建议,所以该类修改、改进、修正仍属于本说明书示范实施例的精神和范围。The basic concepts have been described above. Obviously, for those skilled in the art, the above detailed disclosure is merely an example, and does not constitute a limitation of the present specification. Although not explicitly described herein, various modifications, improvements, and corrections to this specification may occur to those skilled in the art. Such modifications, improvements, and corrections are suggested in this specification, so such modifications, improvements, and corrections still belong to the spirit and scope of the exemplary embodiments of this specification.
同时,本说明书使用了特定词语来描述本说明书的实施例。如“一个实施例”、“一实施例”、和/或“一些实施例”意指与本说明书至少一个实施例相关的某一特征、结构或特点。因此,应强调并注意的是,本说明书中在不同位置两次或多次提及的“一实施例”或“一个实施例”或“一个替代性实施例”并不一定是指同一实施例。此外,本说明书的一个或多个实施例中的某些特征、结构或特点可以进行适当的组合。Meanwhile, the present specification uses specific words to describe the embodiments of the present specification. Such as "one embodiment," "an embodiment," and/or "some embodiments" means a certain feature, structure, or characteristic associated with at least one embodiment of this specification. Therefore, it should be emphasized and noted that two or more references to "an embodiment" or "one embodiment" or "an alternative embodiment" in various places in this specification are not necessarily referring to the same embodiment . Furthermore, certain features, structures or characteristics of the one or more embodiments of this specification may be combined as appropriate.
此外,本领域技术人员可以理解,本说明书的各方面可以通过若干具有可专利性的种类或情况进行说明和描述,包括任何新的和有用的工序、机器、产品或物质的组合,或对他们的任何新的和有用的改进。相应地,本说明书的各个方面可以完全由硬件执行、可以完全由软件(包括固件、常驻软件、微码等)执行、也可以由硬件和软件组合执行。以上硬件或软件均可被称为“数据块”、“模块”、“引擎”、“单元”、“组件”或“系统”。此外,本说明书的各方面可能表现为位于一个或多个计算机可读介质中的计算机 产品,该产品包括计算机可读程序编码。Furthermore, those skilled in the art will appreciate that aspects of this specification may be illustrated and described in several patentable categories or situations, including any new and useful process, machine, product, or combination of matter, or combinations of them. of any new and useful improvements. Accordingly, various aspects of this specification may be performed entirely in hardware, entirely in software (including firmware, resident software, microcode, etc.), or in a combination of hardware and software. The above hardware or software may be referred to as a "data block", "module", "engine", "unit", "component" or "system". Furthermore, aspects of this specification may be embodied as a computer product comprising computer readable program code embodied in one or more computer readable media.
计算机存储介质可能包含一个内含有计算机程序编码的传播数据信号,例如在基带上或作为载波的一部分。该传播信号可能有多种表现形式,包括电磁形式、光形式等,或合适的组合形式。计算机存储介质可以是除计算机可读存储介质之外的任何计算机可读介质,该介质可以通过连接至一个指令执行系统、装置或设备以实现通讯、传播或传输供使用的程序。位于计算机存储介质上的程序编码可以通过任何合适的介质进行传播,包括无线电、电缆、光纤电缆、RF、或类似介质,或任何上述介质的组合。A computer storage medium may contain a propagated data signal with the computer program code embodied therein, for example, on baseband or as part of a carrier wave. The propagating signal may take a variety of manifestations, including electromagnetic, optical, etc., or a suitable combination. Computer storage media can be any computer-readable media other than computer-readable storage media that can communicate, propagate, or transmit a program for use by coupling to an instruction execution system, apparatus, or device. Program code on a computer storage medium may be transmitted over any suitable medium, including radio, cable, fiber optic cable, RF, or the like, or a combination of any of the foregoing.
本说明书各部分操作所需的计算机程序编码可以用任意一种或多种程序语言编写,包括面向对象编程语言如Java、Scala、Smalltalk、Eiffel、JADE、Emerald、C++、C#、VB.NET、Python等,常规程序化编程语言如C语言、Visual Basic、Fortran2003、Perl、COBOL2002、PHP、ABAP,动态编程语言如Python、Ruby和Groovy,或其他编程语言等。该程序编码可以完全在用户计算机上运行、或作为独立的软件包在用户计算机上运行、或部分在用户计算机上运行部分在远程计算机运行、或完全在远程计算机或处理设备上运行。在后种情况下,远程计算机可以通过任何网络形式与用户计算机连接,比如局域网(LAN)或广域网(WAN),或连接至外部计算机(例如通过因特网),或在云计算环境中,或作为服务使用如软件即服务(SaaS)。The computer program coding required for the operation of the various parts of this manual may be written in any one or more programming languages, including object-oriented programming languages such as Java, Scala, Smalltalk, Eiffel, JADE, Emerald, C++, C#, VB.NET, Python etc., conventional procedural programming languages such as C language, Visual Basic, Fortran2003, Perl, COBOL2002, PHP, ABAP, dynamic programming languages such as Python, Ruby and Groovy, or other programming languages. The program code may run entirely on the user's computer, or as a stand-alone software package on the user's computer, or partly on the user's computer and partly on a remote computer, or entirely on the remote computer or processing device. In the latter case, the remote computer can be connected to the user's computer through any network, such as a local area network (LAN) or wide area network (WAN), or to an external computer (eg, through the Internet), or in a cloud computing environment, or as a service Use eg software as a service (SaaS).
此外,除非权利要求中明确说明,本说明书所述处理元素和序列的顺序、数字字母的使用、或其他名称的使用,并非用于限定本说明书流程和方法的顺序。尽管上述披露中通过各种示例讨论了一些目前认为有用的发明实施例,但应当理解的是,该类细节仅起到说明的目的,附加的权利要求并不仅限于披露的实施例,相反,权利要求旨在覆盖所有符合本说明书实施例实质和范围的修正和等价组合。例如,虽然以上所描述的系统组件可以通过硬件设备实现,但是也可以只通过软件的解决方案得以实现,如在现有的处理设备或移动设备上安装所描述的系统。Furthermore, unless explicitly stated in the claims, the order of processing elements and sequences described in this specification, the use of alphanumerics, or the use of other names is not intended to limit the order of the processes and methods of this specification. While the foregoing disclosure discusses by way of various examples some embodiments of the invention that are presently believed to be useful, it is to be understood that such details are for purposes of illustration only and that the appended claims are not limited to the disclosed embodiments, but rather The requirements are intended to cover all modifications and equivalent combinations falling within the spirit and scope of the embodiments of this specification. For example, although the system components described above may be implemented by hardware devices, they may also be implemented by software-only solutions, such as installing the described systems on existing processing devices or mobile devices.
同理,应当注意的是,为了简化本说明书披露的表述,从而帮助对一个或多个发明实施例的理解,前文对本说明书实施例的描述中,有时会将多种特征归并至一个实施例、附图或对其的描述中。但是,这种披露方法并不意味着本说明书对象所需要的特征比权利要求中提及的特征多。实际上,实施例的特征要少于上述披露的单个实施例的全部特征。Similarly, it should be noted that, in order to simplify the expressions disclosed in this specification and thus help the understanding of one or more embodiments of the invention, in the foregoing description of the embodiments of this specification, various features may sometimes be combined into one embodiment, in the drawings or descriptions thereof. However, this method of disclosure does not imply that the subject matter of the description requires more features than are recited in the claims. Indeed, there are fewer features of an embodiment than all of the features of a single embodiment disclosed above.
一些实施例中使用了描述成分、属性数量的数字,应当理解的是,此类用于实施例描述的数字,在一些示例中使用了修饰词“大约”、“近似”或“大体上”来修饰。除非 另外说明,“大约”、“近似”或“大体上”表明所述数字允许有±20%的变化。相应地,在一些实施例中,说明书和权利要求中使用的数值参数均为近似值,该近似值根据个别实施例所需特点可以发生改变。在一些实施例中,数值参数应考虑规定的有效数位并采用一般位数保留的方法。尽管本说明书一些实施例中用于确认其范围广度的数值域和参数为近似值,在具体实施例中,此类数值的设定在可行范围内尽可能精确。Some examples use numbers to describe quantities of ingredients and attributes, it should be understood that such numbers used to describe the examples, in some examples, use the modifiers "about", "approximately" or "substantially" to retouch. Unless stated otherwise, "about", "approximately" or "substantially" means that a variation of ±20% is allowed for the stated number. Accordingly, in some embodiments, the numerical parameters set forth in the specification and claims are approximations that can vary depending upon the desired characteristics of individual embodiments. In some embodiments, the numerical parameters should take into account the specified significant digits and use a general digit reservation method. Notwithstanding that the numerical fields and parameters used in some embodiments of this specification to confirm the breadth of their ranges are approximations, in specific embodiments such numerical values are set as precisely as practicable.
针对本说明书引用的每个专利、专利申请、专利申请公开物和其他材料,如文章、书籍、说明书、出版物、文档等,特此将其全部内容并入本说明书作为参考。与本说明书内容不一致或产生冲突的申请历史文件除外,对本说明书权利要求最广范围有限制的文件(当前或之后附加于本说明书中的)也除外。需要说明的是,如果本说明书附属材料中的描述、定义、和/或术语的使用与本说明书所述内容有不一致或冲突的地方,以本说明书的描述、定义和/或术语的使用为准。For each patent, patent application, patent application publication, and other material, such as article, book, specification, publication, document, etc., cited in this specification, the entire contents of which are hereby incorporated by reference into this specification are hereby incorporated by reference. Application history documents that are inconsistent with or conflict with the contents of this specification are excluded, as are documents (currently or hereafter appended to this specification) limiting the broadest scope of the claims of this specification. It should be noted that, if there is any inconsistency or conflict between the descriptions, definitions and/or use of terms in the accompanying materials of this specification and the contents of this specification, the descriptions, definitions and/or use of terms in this specification shall prevail .
最后,应当理解的是,本说明书中所述实施例仅用以说明本说明书实施例的原则。其他的变形也可能属于本说明书的范围。因此,作为示例而非限制,本说明书实施例的替代配置可视为与本说明书的教导一致。相应地,本说明书的实施例不仅限于本说明书明确介绍和描述的实施例。Finally, it should be understood that the embodiments described in this specification are only used to illustrate the principles of the embodiments of this specification. Other variations are also possible within the scope of this specification. Accordingly, by way of example and not limitation, alternative configurations of the embodiments of this specification may be considered consistent with the teachings of this specification. Accordingly, the embodiments of this specification are not limited to those expressly introduced and described in this specification.

Claims (26)

  1. 一种判断图像真实性的方法,应用于服务器端,其中,所述方法包括:A method for judging the authenticity of an image, applied to the server side, wherein the method includes:
    从客户端获取原始图像;Get the original image from the client;
    根据预设的提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像;Extract multiple images or image parts in the original image as multiple extracted images according to preset extraction rules;
    基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度;Based on the multiple extracted images, determining the degree of matching between the multiple extracted images and the preset sequence by using the trained machine learning model;
    基于所述匹配度,判断所述原始图像的真实性;所述预设序列对应于所述客户端的拍摄设备,判断所述原始图像的真实性具体为:判断所述图像来自所述拍摄设备的真实性。Based on the matching degree, the authenticity of the original image is judged; the preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging that the image comes from the shooting device. authenticity.
  2. 如权利要求1所述的方法,其中,所述从客户端获取原始图像,包括:The method of claim 1, wherein the obtaining the original image from the client comprises:
    生成拍摄参数序列;Generate a sequence of shooting parameters;
    将所述拍摄参数序列下发至所述客户端;sending the shooting parameter sequence to the client;
    从所述客户端获取所述原始图像,所述原始图像由所述客户端基于所述拍摄参数序列生成。The original image is acquired from the client, and the original image is generated by the client based on the sequence of shooting parameters.
  3. 如权利要求2所述的方法,其中,所述生成拍摄参数序列,包括:The method of claim 2, wherein the generating a sequence of shooting parameters comprises:
    确定所述拍摄设备的识别信息;determining the identification information of the photographing device;
    基于所述识别信息,确定所述拍摄设备的拍摄参数集;determining a shooting parameter set of the shooting device based on the identification information;
    基于所述拍摄参数集,生成所述拍摄参数序列。Based on the shooting parameter set, the shooting parameter sequence is generated.
  4. 如权利要求3所述的方法,其中,所述基于所述拍摄参数集,生成所述拍摄参数序列,包括:The method of claim 3, wherein the generating the shooting parameter sequence based on the shooting parameter set comprises:
    从所述拍摄参数集中随机选取预设数量的拍摄参数,并基于该拍摄参数生成所述拍摄参数序列。A preset number of shooting parameters are randomly selected from the shooting parameter set, and the shooting parameter sequence is generated based on the shooting parameters.
  5. 如权利要求4所述的方法,其中,所述拍摄参数序列包括色温参数序列。The method of claim 4, wherein the sequence of shooting parameters includes a sequence of color temperature parameters.
  6. 如权利要求1所述的方法,其中,所述机器学习模型至少包括多个卷积神经网 络单元和一个序列到序列单元;The method of claim 1, wherein the machine learning model includes at least a plurality of convolutional neural network units and a sequence-to-sequence unit;
    所述基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度,包括:The determining the degree of matching between the multiple extracted images and the preset sequence based on the multiple extracted images through the trained machine learning model, including:
    通过所述多个卷积神经网络单元中的每个分别对所述多个提取图像中的每个进行处理,获得每个提取图像对应的图像表示向量;Processing each of the plurality of extracted images by each of the plurality of convolutional neural network units, respectively, to obtain an image representation vector corresponding to each of the extracted images;
    通过所述序列到序列单元对所述图像表示向量进行处理,获得所述多个提取图像的预测拍摄参数的变化序列;The image representation vector is processed by the sequence-to-sequence unit to obtain a change sequence of the predicted shooting parameters of the multiple extracted images;
    基于所述预测拍摄参数的变化序列和所述预设序列确定所述匹配度。The matching degree is determined based on the change sequence of the predicted shooting parameters and the preset sequence.
  7. 如权利要求1所述的方法,其中,所述机器学习模型通过如下方法训练得到:The method of claim 1, wherein the machine learning model is obtained by training as follows:
    获取多个携带标签的训练样本,所述训练样本包括基于样本拍摄参数获取的多个样本图像帧,所述标签包括所述多个样本图像帧之间样本拍摄参数的变化关系;Acquiring a plurality of training samples carrying labels, the training samples comprising a plurality of sample image frames obtained based on the sample shooting parameters, and the labels comprising the variation relationship of the sample shooting parameters between the plurality of sample image frames;
    基于所述多个携带有标签的训练样本训练初始机器学习模型,获得所述机器学习模型。An initial machine learning model is trained based on the plurality of labeled training samples to obtain the machine learning model.
  8. 一种判断图像真实性的系统,应用于服务器端,其中,所述系统包括:A system for judging the authenticity of an image, applied to a server side, wherein the system includes:
    第一获取模块,用于从客户端获取原始图像;The first acquisition module is used to acquire the original image from the client;
    提取模块,用于根据预设的提取规则提取所述原始图像中的多个图像或者图像局部作为多个提取图像;an extraction module, configured to extract multiple images or image parts in the original image as multiple extracted images according to a preset extraction rule;
    确定模块,用于基于所述多个提取图像,通过训练好的机器学习模型确定所述多个提取图像与预设序列的匹配度;A determination module, configured to determine the degree of matching between the multiple extracted images and the preset sequence by using the trained machine learning model based on the multiple extracted images;
    判断模块,用于基于所述匹配度,判断所述原始图像的真实性;所述预设序列对应于所述客户端的拍摄设备,判断所述原始图像的真实性具体为:判断所述图像来自所述拍摄设备的真实性。a judging module for judging the authenticity of the original image based on the matching degree; the preset sequence corresponds to the shooting device of the client, and judging the authenticity of the original image is specifically: judging that the image comes from The authenticity of the photographing equipment.
  9. 如权利要求8所述的系统,其中,所述第一获取模块进一步用于:The system of claim 8, wherein the first obtaining module is further configured to:
    生成拍摄参数序列;Generate a sequence of shooting parameters;
    将所述拍摄参数序列下发至所述客户端;sending the shooting parameter sequence to the client;
    从所述客户端获取所述原始图像,所述原始图像由所述客户端基于所述拍摄参数序列生成。The original image is acquired from the client, and the original image is generated by the client based on the sequence of shooting parameters.
  10. 如权利要求8所述的系统,其中,所述第一获取模块进一步用于:The system of claim 8, wherein the first obtaining module is further configured to:
    确定所述拍摄设备的识别信息;determining the identification information of the photographing device;
    基于所述识别信息,确定所述拍摄设备的拍摄参数集;determining a shooting parameter set of the shooting device based on the identification information;
    基于所述拍摄参数集,生成所述拍摄参数序列。Based on the shooting parameter set, the shooting parameter sequence is generated.
  11. 如权利要求8所述的系统,其中,所述第一获取模块进一步用于:The system of claim 8, wherein the first obtaining module is further configured to:
    从所述拍摄参数集中随机选取预设数量的拍摄参数,并基于该拍摄参数生成所述拍摄参数序列。A preset number of shooting parameters are randomly selected from the shooting parameter set, and the shooting parameter sequence is generated based on the shooting parameters.
  12. 如权利要求11所述的系统,其中,所述拍摄参数序列包括色温参数序列。The system of claim 11, wherein the sequence of shooting parameters includes a sequence of color temperature parameters.
  13. 如权利要求8所述的系统,其中,所述机器学习模型至少包括多个卷积神经网络单元和一个序列到序列单元;所述确定模块进一步用于:The system of claim 8, wherein the machine learning model includes at least a plurality of convolutional neural network units and a sequence-to-sequence unit; the determining module is further configured to:
    通过所述多个卷积神经网络单元中的每个分别对所述多个提取图像中的每个进行处理,获得每个提取图像对应的图像表示向量;Processing each of the plurality of extracted images by each of the plurality of convolutional neural network units, respectively, to obtain an image representation vector corresponding to each of the extracted images;
    通过所述序列到序列单元对所述图像表示向量进行处理,获得所述多个提取图像的预测拍摄参数的变化序列;The image representation vector is processed by the sequence-to-sequence unit to obtain a change sequence of the predicted shooting parameters of the multiple extracted images;
    基于所述预测拍摄参数的变化序列和所述预设序列确定所述匹配度。The matching degree is determined based on the change sequence of the predicted shooting parameters and the preset sequence.
  14. 如权利要求8所述的系统,其中,所述机器学习模型通过如下方法训练得到:The system of claim 8, wherein the machine learning model is trained by the following methods:
    获取多个携带标签的训练样本,所述训练样本包括基于样本拍摄参数获取的多个样本图像帧,所述标签包括所述多个样本图像帧之间样本拍摄参数的变化关系;Acquiring a plurality of training samples carrying labels, the training samples comprising a plurality of sample image frames obtained based on the sample shooting parameters, and the labels comprising the variation relationship of the sample shooting parameters between the plurality of sample image frames;
    基于所述多个携带有标签的训练样本训练初始机器学习模型,获得所述机器学习模型。An initial machine learning model is trained based on the plurality of labeled training samples to obtain the machine learning model.
  15. 一种判断图像真实性的方法,其中,应用于客户端,所述方法包括:A method for judging the authenticity of an image, wherein, applied to a client, the method includes:
    获取服务器端生成并下发的拍摄参数序列;Obtain the shooting parameter sequence generated and delivered by the server;
    基于所述拍摄参数序列生成所述原始图像;generating the original image based on the sequence of shooting parameters;
    将所述原始图像发送给服务器端;sending the original image to the server;
    获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。Obtain the information sent by the server and including the result of the judgment on the authenticity of the original image.
  16. 如权利要求15所述的方法,其中,在所述获取服务器端生成并下发的拍摄参数序列之前,所述方法还包括:The method according to claim 15, wherein, before the acquiring the shooting parameter sequence generated and delivered by the server, the method further comprises:
    向服务器端上传所述客户端的拍摄设备的识别信息。The identification information of the client's photographing device is uploaded to the server.
  17. 如权利要求15所述的方法,其中,所述拍摄参数序列是所述服务器端随机生成的。The method of claim 15, wherein the shooting parameter sequence is randomly generated by the server.
  18. 如权利要求17所述的方法,其中,所述拍摄参数序列是所述服务器端基于所述拍摄设备的拍摄参数集随机生成的;所述拍摄参数集对应于所述拍摄设备的识别信息。The method of claim 17, wherein the shooting parameter sequence is randomly generated by the server based on a shooting parameter set of the shooting device; the shooting parameter set corresponds to the identification information of the shooting device.
  19. 如权利要求18所述的方法,其中,所述拍摄参数序列包括色温参数序列。19. The method of claim 18, wherein the sequence of shooting parameters includes a sequence of color temperature parameters.
  20. 一种判断图像真实性的系统,应用于客户端,其中,所述系统包括:A system for judging the authenticity of an image, applied to a client, wherein the system includes:
    第二获取模块,用于获取服务器端生成并下发的拍摄参数序列;The second acquisition module is used to acquire the shooting parameter sequence generated and issued by the server;
    生成模块,用于基于所述拍摄参数序列生成所述原始图像;a generating module, configured to generate the original image based on the shooting parameter sequence;
    发送模块,用于将所述原始图像发送给服务器端;a sending module, configured to send the original image to the server;
    第三获取模块,用于获取所述服务器端发送的、包含对所述原始图像的真实性的判断结果的信息。The third obtaining module is configured to obtain the information sent by the server and including the judgment result of the authenticity of the original image.
  21. 如权利要求20所述的系统,其中,所述系统还包括上传模块,所述上传模块用于向服务器端上传所述客户端的拍摄设备的识别信息。The system according to claim 20, wherein the system further comprises an uploading module, the uploading module is configured to upload the identification information of the photographing device of the client to the server.
  22. 如权利要求20所述的系统,其中,所述拍摄参数序列是所述服务器端随机生成的。The system of claim 20, wherein the shooting parameter sequence is randomly generated by the server.
  23. 如权利要求22所述的系统,其中,所述拍摄参数序列是所述服务器端基于所述拍摄设备的拍摄参数集随机生成的;所述拍摄参数集对应于所述拍摄设备的识别信息。The system of claim 22, wherein the shooting parameter sequence is randomly generated by the server based on a shooting parameter set of the shooting device; the shooting parameter set corresponds to the identification information of the shooting device.
  24. 如权利要求23所述的系统,其中,所述拍摄参数序列包括色温参数序列。24. The system of claim 23, wherein the sequence of shooting parameters includes a sequence of color temperature parameters.
  25. 一种判断图像真实性的装置,所述装置包括处理器以及存储器;所述存储器用于存储指令,其中,所述处理器用于执行所述指令,以实现如权利要求1至7或者权利要求15至19中任一项所述判断图像真实性的方法对应的操作。An apparatus for judging the authenticity of an image, the apparatus includes a processor and a memory; the memory is used for storing instructions, wherein the processor is used for executing the instructions, so as to realize the claims 1 to 7 or claim 15 Operations corresponding to the method for judging the authenticity of an image described in any one of to 19.
  26. 一种计算机可读存储介质,所述存储介质存储计算机指令,当计算机读取存储介质中的计算机指令后,计算机执行如权利要求1至7或者权利要求15至19中任一项所述判断图像真实性的方法对应的操作。A computer-readable storage medium, the storage medium stores computer instructions, and after the computer reads the computer instructions in the storage medium, the computer executes the judgment image according to any one of claims 1 to 7 or claims 15 to 19 The method of authenticity corresponds to the operation.
PCT/CN2021/108607 2020-07-30 2021-07-27 Image authenticity determination method and system WO2022022493A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN202010750632.5 2020-07-30
CN202010750632.5A CN111881844B (en) 2020-07-30 2020-07-30 Method and system for judging image authenticity

Publications (1)

Publication Number Publication Date
WO2022022493A1 true WO2022022493A1 (en) 2022-02-03

Family

ID=73205684

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2021/108607 WO2022022493A1 (en) 2020-07-30 2021-07-27 Image authenticity determination method and system

Country Status (2)

Country Link
CN (1) CN111881844B (en)
WO (1) WO2022022493A1 (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111881844B (en) * 2020-07-30 2021-05-07 北京嘀嘀无限科技发展有限公司 Method and system for judging image authenticity
CN113807160B (en) * 2021-04-06 2024-02-06 京东科技控股股份有限公司 Method and device for detecting hijacking of image acquisition equipment and computer equipment
CN113158893A (en) * 2021-04-20 2021-07-23 北京嘀嘀无限科技发展有限公司 Target identification method and system
CN113111806A (en) * 2021-04-20 2021-07-13 北京嘀嘀无限科技发展有限公司 Method and system for object recognition
CN113111807B (en) * 2021-04-20 2024-06-07 北京嘀嘀无限科技发展有限公司 Target identification method and system
CN113111811A (en) * 2021-04-20 2021-07-13 北京嘀嘀无限科技发展有限公司 Target discrimination method and system
CN113364911B (en) * 2021-06-11 2023-03-07 上海兴容信息技术有限公司 Detection method and system for preset terminal
CN113971840A (en) * 2021-10-22 2022-01-25 支付宝(杭州)信息技术有限公司 Method and device for preventing image injection attack

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103646375A (en) * 2013-11-26 2014-03-19 上海交通大学 Method for authenticating primitiveness of picture photographed by intelligent mobile terminal
CN105590043A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Authentication method, device and system
CN108965824A (en) * 2018-08-13 2018-12-07 晋商博创(北京)科技有限公司 Video monitoring method, system, camera, server and client based on CPK
CN110598754A (en) * 2019-08-16 2019-12-20 泰致德科技发展(深圳)有限公司 Online monitoring method, device and equipment based on image detection and storage medium
CN111027573A (en) * 2019-12-09 2020-04-17 广西荔玉高速公路有限公司 Image authenticity identification method based on blind evidence obtaining technology
CN111161259A (en) * 2019-12-31 2020-05-15 支付宝(杭州)信息技术有限公司 Method and device for detecting whether image is tampered or not and electronic equipment
CN111881844A (en) * 2020-07-30 2020-11-03 北京嘀嘀无限科技发展有限公司 Method and system for judging image authenticity

Family Cites Families (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101751714A (en) * 2008-12-05 2010-06-23 深圳富泰宏精密工业有限公司 Multifunctional portable electronic device
DE102011054658A1 (en) * 2011-10-20 2013-04-25 Bioid Ag Method for distinguishing between a real face and a two-dimensional image of the face in a biometric capture process
US10820840B2 (en) * 2016-04-28 2020-11-03 Joshua Noel Hogan Optical coherence tomography for identity verification
US10657401B2 (en) * 2017-06-06 2020-05-19 Microsoft Technology Licensing, Llc Biometric object spoof detection based on image intensity variations
CN110677591B (en) * 2019-10-28 2021-03-23 Oppo广东移动通信有限公司 Sample set construction method, image imaging method, device, medium and electronic equipment

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103646375A (en) * 2013-11-26 2014-03-19 上海交通大学 Method for authenticating primitiveness of picture photographed by intelligent mobile terminal
CN105590043A (en) * 2014-10-22 2016-05-18 腾讯科技(深圳)有限公司 Authentication method, device and system
CN108965824A (en) * 2018-08-13 2018-12-07 晋商博创(北京)科技有限公司 Video monitoring method, system, camera, server and client based on CPK
CN110598754A (en) * 2019-08-16 2019-12-20 泰致德科技发展(深圳)有限公司 Online monitoring method, device and equipment based on image detection and storage medium
CN111027573A (en) * 2019-12-09 2020-04-17 广西荔玉高速公路有限公司 Image authenticity identification method based on blind evidence obtaining technology
CN111161259A (en) * 2019-12-31 2020-05-15 支付宝(杭州)信息技术有限公司 Method and device for detecting whether image is tampered or not and electronic equipment
CN111881844A (en) * 2020-07-30 2020-11-03 北京嘀嘀无限科技发展有限公司 Method and system for judging image authenticity

Also Published As

Publication number Publication date
CN111881844B (en) 2021-05-07
CN111881844A (en) 2020-11-03

Similar Documents

Publication Publication Date Title
WO2022022493A1 (en) Image authenticity determination method and system
US10332118B2 (en) Efficient prevention of fraud
US11539525B2 (en) Systems and methods for secure tokenized credentials
CN106599772B (en) Living body verification method and device and identity authentication method and device
US10839238B2 (en) Remote user identity validation with threshold-based matching
US20140270404A1 (en) Efficient prevention of fraud
US20170262472A1 (en) Systems and methods for recognition of faces e.g. from mobile-device-generated images of faces
US20140270409A1 (en) Efficient prevention of fraud
JP5710748B2 (en) Biometric authentication system
CN111476268A (en) Method, device, equipment and medium for training reproduction recognition model and image recognition
US20200374286A1 (en) Real time selfie systems and methods for automating user identify verification
Timmerman et al. Video camera identification from sensor pattern noise with a constrained convnet
US20200019691A1 (en) Biometric recognition method
WO2022078168A1 (en) Identity verification method and apparatus based on artificial intelligence, and computer device and storage medium
CN109005104A (en) A kind of instant communicating method, device, server and storage medium
CN111027400A (en) Living body detection method and device
US20220375259A1 (en) Artificial intelligence for passive liveness detection
CN114387548A (en) Video and liveness detection method, system, device, storage medium and program product
CN112491844A (en) Voiceprint and face recognition verification system and method based on trusted execution environment
WO2021049234A1 (en) Image analysis device, control method, and program
WO2021000407A1 (en) Character verification method and apparatus, and computer device and storage medium
US20200019689A1 (en) Biometric recognition method
WO2022222957A1 (en) Method and system for identifying target
CN114387674A (en) Living body detection method, living body detection system, living body detection apparatus, storage medium, and program product
WO2022104340A1 (en) Artificial intelligence for passive liveness detection

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 21851351

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

32PN Ep: public notification in the ep bulletin as address of the adressee cannot be established

Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16.05.2023)

122 Ep: pct application non-entry in european phase

Ref document number: 21851351

Country of ref document: EP

Kind code of ref document: A1