CN112597808A - Tamper detection method and system - Google Patents
Tamper detection method and system Download PDFInfo
- Publication number
- CN112597808A CN112597808A CN202011388928.3A CN202011388928A CN112597808A CN 112597808 A CN112597808 A CN 112597808A CN 202011388928 A CN202011388928 A CN 202011388928A CN 112597808 A CN112597808 A CN 112597808A
- Authority
- CN
- China
- Prior art keywords
- face
- image
- neural network
- training samples
- identification card
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000001514 detection method Methods 0.000 title claims abstract description 14
- 238000012549 training Methods 0.000 claims abstract description 76
- 238000013528 artificial neural network Methods 0.000 claims abstract description 42
- 238000000034 method Methods 0.000 claims abstract description 34
- 238000012545 processing Methods 0.000 claims description 17
- 239000002131 composite material Substances 0.000 claims description 10
- 230000006870 function Effects 0.000 claims description 7
- 238000010200 validation analysis Methods 0.000 claims description 7
- 238000004891 communication Methods 0.000 description 21
- 238000004590 computer program Methods 0.000 description 21
- 238000003860 storage Methods 0.000 description 21
- 238000010586 diagram Methods 0.000 description 9
- 238000013527 convolutional neural network Methods 0.000 description 7
- 230000003287 optical effect Effects 0.000 description 6
- 230000005540 biological transmission Effects 0.000 description 4
- 238000012360 testing method Methods 0.000 description 3
- 238000013459 approach Methods 0.000 description 2
- 238000013145 classification model Methods 0.000 description 2
- 238000005520 cutting process Methods 0.000 description 2
- 238000004519 manufacturing process Methods 0.000 description 2
- 238000012795 verification Methods 0.000 description 2
- 238000010276 construction Methods 0.000 description 1
- 230000001419 dependent effect Effects 0.000 description 1
- 230000000873 masking effect Effects 0.000 description 1
- 238000005065 mining Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000007639 printing Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/161—Detection; Localisation; Normalisation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q30/00—Commerce
- G06Q30/018—Certifying business or products
- G06Q30/0185—Product, service or business identity fraud
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/26—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
- G06V10/267—Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V40/00—Recognition of biometric, human-related or animal-related patterns in image or video data
- G06V40/10—Human or animal bodies, e.g. vehicle occupants or pedestrians; Body parts, e.g. hands
- G06V40/16—Human faces, e.g. facial parts, sketches or expressions
- G06V40/172—Classification, e.g. identification
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- Health & Medical Sciences (AREA)
- Business, Economics & Management (AREA)
- Multimedia (AREA)
- Oral & Maxillofacial Surgery (AREA)
- Mathematical Physics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Biomedical Technology (AREA)
- Computational Linguistics (AREA)
- Software Systems (AREA)
- Human Computer Interaction (AREA)
- Biophysics (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Image Analysis (AREA)
- Accounting & Taxation (AREA)
- Development Economics (AREA)
- Economics (AREA)
- Finance (AREA)
- Marketing (AREA)
- Strategic Management (AREA)
- General Business, Economics & Management (AREA)
- Entrepreneurship & Innovation (AREA)
Abstract
A tamper detection method and system, the method comprising: extracting image data of a face area of an identity card from an image of the identity card; applying a trained neural network to the extracted image data to calculate a confidence score; and comparing the confidence score to a threshold to determine whether the identification card has been tampered with. Training the neural network comprises: inputting each training sample in a plurality of training samples into a binary classifier to classify based on the face photos of the training samples, and iteratively optimizing the parameters of the neural network based on the classification results of the plurality of training samples.
Description
Technical Field
This document relates generally, but not exclusively, to identity fraud detection, including methods, devices and systems for detecting tampering with an identity card.
Background
Identity fraud is generally considered to be the unauthorized use of personal information of another person by one person to crime or deceive or defraud the other person or a third person. One example is identification card fraud, where forged or tampered identities are used for identity fraud.
For example, "electronically know your customer (eKYC)" usually needs to submit an image (e.g., a photograph or a scanned part) of an identification card, and it can be observed from some of these images that the identification card may have a picture of a substitute face, and the detailed information such as name, identification card number, birth date, and biometric data is the original information. In the eKYC process, a picture of a replacement face may match a scan of a live face, but the personal information clearly belongs to another person. Such fraud may pose a significant financial risk to the eKYC-dependent enterprise, as well as potential legal and security issues.
It may be desirable to provide methods, devices and systems that can detect such tampering to prevent identity card fraud.
Disclosure of Invention
An embodiment provides a tamper detection method, including: extracting image data of a face area of an identity card from an image of the identity card; applying a trained neural network to the extracted image data to calculate a confidence score; and comparing the confidence score to a threshold to determine whether the identification card has been tampered with. Training the neural network comprises: inputting each training sample in a plurality of training samples into a binary classifier to classify based on the face photos of the training samples, and iteratively optimizing the parameters of the neural network based on the classification results of the plurality of training samples.
Another embodiment provides a tamper detection system comprising: at least one processor; and a computer readable memory coupled to the processor and having instructions stored thereon. The instructions are executable by the processor to: extracting image data of a face area of an identity card from an image of the identity card; applying a trained neural network to the extracted image data to calculate a confidence score; determining whether the identity card is tampered based on a comparison of the confidence score to a threshold. The trained neural network includes: a binary classifier trained to classify each of a plurality of training samples based on a face photograph of the training sample; and neural network parameters that are iteratively optimized based on the classification results of the plurality of training samples.
Another embodiment provides an apparatus comprising: the receiving equipment is used for receiving the image of the identity card; and the processing equipment is used for extracting the image data of the face area of the identity card from the received image. The processing device provides the extracted image data to a trained neural network to calculate a confidence score, and determines whether the identification card has been tampered with based on a comparison of the confidence score to a threshold value. The trained neural network includes: a binary classifier trained to classify each of a plurality of training samples based on a face photograph of the training sample; and neural network parameters that are iteratively optimized based on the classification results of the plurality of training samples.
Drawings
Embodiments of the present disclosure will be better understood and appreciated by those of ordinary skill in the art from the following written description, by way of example only, and with reference to the accompanying drawings.
Fig. 1 shows a flow diagram illustrating a method of detecting tampering with an identity card according to an embodiment.
Fig. 2 shows a flow chart illustrating a detailed embodiment of the method of fig. 1.
FIG. 3 shows a flow diagram illustrating a detailed implementation of an automatic image generation process for preparing negative training samples.
Fig. 4 shows a schematic diagram of an apparatus suitable for implementing the method of fig. 1.
FIG. 5 shows a schematic diagram illustrating a computer system suitable for implementing the method of FIG. 1.
Skilled artisans will appreciate that elements in the figures are illustrated for simplicity and clarity and have not necessarily been drawn to scale. For example, the dimensions of some of the elements in the figures, block diagrams, or flowcharts may be exaggerated relative to other elements to help improve understanding of embodiments herein.
Detailed Description
The present disclosure provides a method for detecting tampering, in particular photo replacement, in identification card fraud. In the present disclosure, tamper detection is considered a classification task, where two classes of classification models, namely, genuine (i.e., untampered) cards and counterfeit (i.e., tampered) cards with photo replacement. Furthermore, the classification model is trained using regions of interest (e.g., face regions) of the identification card rather than the entire captured image. In other words, the tamper detection method can achieve high accuracy and high speed by focusing more attention on an area where tampering can be performed in the identification card.
Embodiments will now be described, by way of example only, with reference to the accompanying drawings. Like reference numbers and characters in the drawings indicate like elements or equivalents.
Some portions of the following description are presented in terms of algorithms and functional or symbolic representations of operations on data within a computer memory. These algorithmic descriptions and functional or symbolic representations are the means used by those skilled in the data processing arts to most effectively convey the substance of their work to others skilled in the art. An algorithm is here, and generally, considered to be a self-consistent sequence of steps leading to a desired result. These steps are those requiring physical manipulations of physical quantities such as electrical, magnetic or optical signals capable of being stored, transferred, combined, compared, and otherwise manipulated.
Unless specifically stated otherwise, and as will be apparent from the following, it is appreciated that throughout the present document, discussions utilizing terms such as "scanning," "computing," "determining," "applying," "extracting," "generating," "inputting," "outputting," "optimizing," or the like, refer to the action and processes of a computer system, or similar electronic device, that manipulates and transforms data represented as physical quantities within the computer system into other data similarly represented as physical quantities within the computer system or other information storage, transmission or display devices.
Also disclosed herein are apparatuses for performing the operations of the methods. Such apparatus may be specially constructed for the required purposes, or may comprise a computer or other device selectively activated or reconfigured by a computer program stored in the computer. The algorithms and displays presented herein are not inherently related to any particular computer or other apparatus. Various machines may be used with programs in accordance with the teachings herein. Alternatively, the construction of a more specialized apparatus for carrying out the required method steps may be appropriate. The structure of a conventional computer will appear from the description below.
Further, a computer program is implicitly disclosed herein, since it is clear to a person skilled in the art that the individual steps of the methods described herein can be implemented by computer code. The computer program is not intended to be limited to any particular programming language and implementation thereof. It will be appreciated that a variety of programming languages and code therefor may be used to implement the teachings of the disclosure as contained herein. Moreover, the computer program is not intended to be limited to any particular control flow. There are many other variations of computer programs that may use different control flows without departing from the scope of this disclosure.
Furthermore, one or more steps of a computer program may be executed in parallel rather than sequentially. Such a computer program may be stored on any computer readable medium. The computer-readable medium may include a storage device such as a magnetic or optical disk, a memory chip, or other storage device suitable for interfacing with a computer. The computer readable media may also include hardwired media such as those exemplified in the internet systems, or wireless media such as those exemplified in GSM, GPRS, 3G, 4G or 5G mobile phone systems, as well as other wireless systems such as bluetooth, ZigBee, Wi-Fi. When loaded and executed on such a computer, effectively creates means for implementing the steps of the preferred method.
The present disclosure may also be implemented as hardware elements. More specifically, in a hardware sense, an element is a functional hardware unit designed for use with other components or elements. For example, one element may be implemented using discrete electronic components, or an element may form part of an overall electronic circuit, such as an Application Specific Integrated Circuit (ASIC) or a Field Programmable Gate Array (FPGA). Many other possibilities exist. Those skilled in the art will appreciate that the system may also be implemented as a combination of hardware and software elements.
According to various embodiments, "circuitry" may be understood as any kind of logic implementing entity, which may be a dedicated circuit or a processor executing software stored in a memory, firmware, or any combination thereof. Thus, in embodiments, a "circuit" may be a hardwired logic circuit or a programmable logic circuit, e.g., a programmable processor such as a microprocessor (e.g., a Complex Instruction Set Computer (CISC) processor or a Reduced Instruction Set Computer (RISC) processor). The "circuitry" may also be a processor executing software, e.g. any kind of computer program, such as a computer program using virtual machine code such as Java. According to alternative embodiments, any other kind of implementation of the respective functions, which may be described in more detail herein, may also be understood as a "circuit".
Fig. 1 shows a flow diagram illustrating a method of detecting tampering with an identity card according to an embodiment. Examples of identification cards include, but are not limited to, driver's licenses, social security cards, passport pages, and other official release documents that can be accepted as valid identification. At step 102, image data of a face region of the identification card is extracted from an image of the identification card. As described in further detail below, this may involve a first crop of the image to contain substantially only the identification card, and a second crop to retain the portion of the image where the face region is located (e.g., the lower left corner of the identification card).
At step 104, the trained neural network is applied to the extracted image data to calculate a confidence score. The confidence score is then compared to a threshold value at step 106 to determine whether the identification card has been tampered with. For example, if the confidence score is greater than a threshold, the photograph on the identification card is determined to be a replacement photograph and the identification card is determined to be a counterfeit (i.e., tampered) identification card. On the other hand, if the confidence score is less than the threshold, the photograph on the identification card is determined to be the original photograph and the identification card is determined to be a genuine (i.e., untampered) identification card.
The neural network may be a Convolutional Neural Network (CNN) suitable for image classification tasks, such as ResNet-18 CNN, and the confidence score may be calculated from the trained CNN using a normalized exponential function. The threshold may be determined based on the marked validation dataset. More specifically, the verification dataset includes a normal identification card image and a counterfeit identification card image (i.e., replaced with a photograph), and then a Receiver Operating Characteristic (ROC) curve may be generated based on the label (i.e., 0 or 1) of the verification dataset and the classification/prediction result (i.e., 0 or 1) from the binary classifier. In one embodiment, the threshold may be set when the False Acceptance Rate (FAR) of the binary classifier is equal to 0.01, based on the ROC curve. In other embodiments, the threshold may be set when the FAR of the binary classifier is equal to 0.001. It will be appreciated that in alternative embodiments different FAR values may be used and the threshold may be varied accordingly.
In the present approach, the neural network may be trained with a plurality of training samples (e.g., tens of thousands of samples). Training the neural network may include inputting each of a plurality of training samples into a binary classifier to classify based on a face photograph of the training sample, and iteratively optimizing parameters of the neural network based on classification results of the plurality of training samples. During the training phase, the number and quality of the training samples may affect the accuracy of the classifier.
FIG. 2 shows a flow diagram 200 of a detailed implementation of the method of FIG. 1, including a data processing phase 202, a training phase 204, and a testing phase 206.
In the data processing stage 202, a portion 208 of the identification card 210 containing a picture of a human face, hereinafter also referred to as a face region, is extracted. The input image I1 of identification card 210 in data processing stage 202 may be a photograph or scanned of identification card 210. The image alignment step 212 is performed by detecting the four corners of the identification card 210 from the input image l1 to perform a first crop and aligning the cropped image I2 based on the four corners. In an exemplary embodiment, the resolution of the cropped image I2 is 400 × 300 pixels, and some background around the identity card 210 may be included in the image I2. Then, a face cropping step (i.e., second cropping) is performed to extract a face region 208 having a size of 200 × 300 pixels from the cropped image I2, and then resize the face region 208 into a rectangle having a size of 192 × 256 pixels.
The data processing stage 202 may be used to prepare training samples for the training stage 204. In addition, the data processing stage may be used to extract the face region of the identification card during the testing stage 206 or actual use of the tamper detection method.
Referring to the training phase 204, the training samples may include a plurality of positive training samples 214 and a plurality of negative training samples 216. Using the data processing stage 202 as described above, a positive training sample 214 is obtained from a real image of the identification cards, each having a respective picture of a face. In one embodiment, about 15 thousand such positive training samples are used, but it is understood that the number may vary depending on the training results (e.g., accuracy of implementation).
The negative training samples 216 may similarly be obtained from real images of identification cards, with a picture of the face of each card replaced. However, it has been recognized that the process of preparing such real images in a physical/manual manner requires repeated printing and cropping of physical face photographs, superimposing the face photographs on the face area of the identification card and taking the photographs, and therefore this approach is time consuming and expensive when a large number of negative training samples (e.g., 15 tens of thousands) are required. In a preferred embodiment, to efficiently prepare a large number of negative training samples, the negative training samples are obtained using the data processing stage 202 above from synthetic images, each generated by superimposing a photograph of a human face on an identification card image based on an automatic image generation process, as described in further detail below with reference to FIG. 3.
In one embodiment, the positive training samples 214 are labeled 0 and the negative training samples 216 are labeled 1. Each of the positive training samples 214 and the negative training samples 216 has a size of 192 × 256 pixels (corresponding to a rectangle) instead of the size of 224 × 224 pixels typically used in CNNs such as ResNet-18 CNN. The size of 192 x 256 pixels may better match the face area of the identification card because the face area is generally rectangular in shape rather than square. Each training sample is provided to a neural network based classifier, such as CNN classifier 218, for binary classification based on the photographs of the faces of the training samples. In other words, the face region images are used to train a neural network based classifier, rather than a full identification card image. Thus, in use, the trained classifier may focus more on the face region of the identification card image, ignoring unnecessary regions of the identification card image, and may significantly improve the accuracy of detecting the replacement photograph. The classification result may be a value 0 or a value 1. The training phase 202 also includes iteratively optimizing parameters of the neural network based on the classification results of the plurality of training samples.
In the testing stage 206 or actual use, the submitted identification card image first passes through the data processing stage 202 as described above to obtain a cropped image 220 of the face area of the identification card. The cropped image 220 is then provided to the trained neural network based classifier 218 to obtain a confidence score. If the confidence score is greater than the defined threshold, the submitted image is considered to be the image of the identification card that replaced the face photograph, i.e., the identification card has been tampered with. On the other hand, if the confidence score is less than the defined threshold, the submitted image is considered to be an image of the original identification card, i.e., the identification card has not been tampered with.
The confidence score may be calculated using a normalized exponential function. In one example, the threshold is determined from the tagged validation dataset such that the false acceptance rate of the classifier 218 based on the threshold is equal to 0.01. In another embodiment, the threshold is determined from the tagged validation data set such that the false acceptance rate of the classifier 218 based on the threshold is equal to 0.001. In alternative embodiments, different false acceptance rates may be used.
FIG. 3 shows a flow diagram 300 illustrating a detailed embodiment of an automatic image generation process for preparing negative training samples. First, a face image library including only face images from the identification card is created, and one face image (face _ image) is randomly selected in step 302. An identification card image library including identification card images is also created and one identification card image (card image) is randomly selected in step 304. Next, face detection is performed on the selected face image to obtain a first rectangular frame (face _ roi) in step 306. Similarly, face detection is performed on the selected card image to obtain a second rectangular frame (card _ roi) corresponding to the face photograph in step 308. Then, in step 310, the first rectangular box (face _ roi) is clipped to extract a face photo (face _ crop). In an embodiment, the cropping may be performed by one of: (a) randomly moving four end points of the face _ roi within a small range to obtain a random quadrangle and cutting the quadrangle; (b) randomly generating n end points around the face _ roi, connecting the end points pairwise to obtain a random n-polygon, and cutting the n-polygon; (c) a face segmentation algorithm is performed on the face _ roi to obtain an avatar edge (avatar edge) and to trim the avatar edge.
The cropped face picture (face _ crop) is then further processed. For example, in step 312, the average luminance of the second rectangular frame (card _ roi) is calculated, and the luminance of the face _ crop is adjusted to the average luminance of the card _ roi. In addition, shading may be added to the face _ crop at any angle. Then, at step 314, the size of the face _ crop is scaled to match the size of the card _ roi, and the face _ crop is superimposed on the card _ roi to generate a composite image. Further, at step 316, gaussian noise and gaussian blur are randomly added to the composite image to obtain a final fitted photograph (tamper image).
The automatic image generation process also includes hard sample mining. The goal is to find those negative examples that current neural network-based classifiers cannot correctly identify. This process collects "hard" negative examples, which are then sent to further optimize the next generation/iteration of neural network-based classifiers. Specifically, at step 318, the pointer _ image is provided to the neural network-based classifier to determine whether the image is an image with a replacement photograph. If the classifier correctly determines that the pointer image is indeed an image with a replacement photograph, which is a normal result, then the sample is considered a simple/regular sample and is ignored (step 320). On the other hand, if the classifier incorrectly determines that the pointer image is a photograph with the original photograph (i.e., no replacement), which is a false negative result, then the sample is considered a "hard" sample and saved (step 322). The collected hard samples may then be used to retrain the neural network-based classifier by updating and optimizing neural network parameters to substantially reduce or minimize false negative conditions.
Fig. 4 shows a schematic diagram of an apparatus 400 suitable for implementing the method of fig. 1. The device 400 includes a receiving device 402 connected to a processing device 404. Receiving device 402 may receive image 406 of an identification card. The processing device 404 may extract image data of the face region of the identification card from the received image 406, provide the extracted image data to a trained neural network 408 to calculate a confidence score 410, and determine whether the identification card has been tampered with based on a comparison of the confidence score 410 to a threshold 412. The trained neural network 408 includes: a binary classifier 414 trained to classify each of a plurality of training samples based on a face photograph of the training sample; and iteratively optimized neural network parameters based on the classification results of the plurality of training samples 416.
As described, the present methods, systems, and apparatus can detect identification card fraud by detecting tampering of a face photograph based on an image of an identification card. Tamper detection is performed by a classifier trained to focus on regions of interest that generally correspond in shape, size, and location to the face photograph regions of the identification card, thereby improving the speed and accuracy of classification. In some embodiments, the large number of samples used to train the classifier are generated by strictly automated processing with high quality. In addition, the classifier can be trained on hard samples that typically give false negatives of the selection, thereby improving classification accuracy.
Fig. 5 depicts an exemplary computing device 500, hereinafter interchangeably referred to as computer system 500, in which one or more such computing devices 500 may be used in the apparatus 400. The following description of computing device 500 is provided by way of example only and is not intended to be limiting.
As shown in fig. 5, exemplary computing device 500 includes a processor 504 for executing software routines. Although single processors are shown for clarity, computing device 500 may also comprise a multi-processor system. The processor 504 is connected to a communication infrastructure 506 to communicate with other components of the computing device 500. The communication infrastructure 506 may include, for example, a communication bus, a crossbar, or a network.
In alternative embodiments, secondary memory 510 may additionally or alternatively include other similar means for allowing computer programs or other instructions to be loaded into computing device 500. Such means may include, for example, a removable storage unit 522 and an interface 520. Examples of removable storage unit 522 and interface 520 include a program cartridge and cartridge interface (e.g., such as that found in video game devices), a removable memory chip (e.g., an EPROM, or PROM) and associated socket, and other removable storage units 522 and interfaces 520 which allow software and data to be transferred from removable storage unit 522 to computer system 500.
As shown in fig. 5, computing device 500 further includes: a display interface 502 that performs operations for presenting images to an associated display 530; and an audio interface 532 that performs operations for playing audio content via an associated speaker 534.
As used herein, the term "computer program product" may refer, in part, to removable storage unit 518, removable storage unit 522, a hard disk installed in hard disk drive 512, or a carrier wave of software that passes through communications path 526 (wireless link or cable) to communications interface 524. Computer-readable storage media refers to any non-transitory tangible storage medium that provides recorded instructions and/or data to computing device 500 for execution and/or processing. Examples of such storage media include floppy disks, magnetic tapes, CD-ROMs, DVDs, Blu-ray (Blu-ray)TM) An optical disk, a hard drive, a ROM or integrated circuit, a USB memory, a magneto-optical disk, or a computer readable card such as a PCMCIA card, whether or not such devices are internal or external to computing device 500. Can also participate in the direction meterExamples of transitory or non-tangible computer-readable transmission media in which computing device 500 provides software, applications, instructions, and/or data include radio or infrared transmission channels and network connections to another computer or networked device, as well as the internet or ethernet, etc., including e-mail transmissions and information recorded on websites and the like.
Computer programs (also called computer program code) are stored in main memory 508 and/or secondary memory 510. Computer programs may also be received via communications interface 524. Such computer programs, when executed, enable computing device 500 to perform one or more features of embodiments discussed herein. In various embodiments, the computer programs, when executed, enable processor 504 to perform the features of the embodiments described above. Accordingly, such computer programs represent controllers of the computer system 500.
The software may be stored in a computer program product and loaded into computing device 500 using removable storage drive 514, hard drive 512, or interface 520. Alternatively, the computer program product may be downloaded to computer system 500 via communications path 526. The software, when executed by the processor 504, causes the computing device 500 to perform the functions of the embodiments described herein.
It should be understood that the embodiment of fig. 5 is given by way of example only. Thus, in some embodiments, one or more features of computing device 500 may be omitted. Also, in some embodiments, one or more features of computing device 500 may be combined together. Additionally, in some embodiments, one or more features of computing device 500 may be separated into one or more components.
It should be understood that the elements shown in fig. 5 are used to provide means to perform the various functions and operations of the server as described in the above embodiments.
In an embodiment, a server may generally be described as a physical device including at least one processor and at least one memory including computer program code. The at least one memory and the computer program code are configured to, with the at least one processor, cause the physical device to perform the necessary operations.
It will be appreciated by persons skilled in the art that numerous variations and/or modifications may be made to what is shown in a particular embodiment of the disclosure without departing from the scope of the disclosure as broadly described. For example, the threshold may be adjusted based on actual demand, for example. The described embodiments are, therefore, to be considered in all respects as illustrative and not restrictive.
Claims (20)
1. A tamper detection method, comprising:
extracting image data of a face area of an identity card from an image of the identity card;
applying a trained neural network to the extracted image data to compute a confidence score, wherein training the neural network comprises:
inputting each training sample in a plurality of training samples into a binary classifier to classify the face photos based on the training samples; and
iteratively optimizing parameters of the neural network based on classification results of the plurality of training samples; and
comparing the confidence score to a threshold to determine whether the identification card has been tampered with.
2. The method of claim 1, wherein the confidence score is calculated using a normalized exponential function.
3. The method of claim 1 or 2, wherein the threshold is determined from a labeled validation dataset such that the false acceptance rate of the binary classifier based on the threshold is equal to 0.01.
4. The method of claim 1 or 2, wherein the threshold is determined from a labeled validation dataset such that a false acceptance rate of the binary classifier based on the threshold is equal to 0.001.
5. The method of any of the preceding claims, wherein the training samples comprise:
a plurality of positive training samples obtained from real images of a plurality of identification cards, each having a respective picture of a face;
a plurality of negative training samples obtained from a plurality of composite images, wherein each composite image is generated by superimposing a photograph of a human face on an identification card photograph based on an automatic image generation process.
6. The method of claim 5, wherein superimposing the picture of the face on the identification card based on the automatic image generation process comprises:
randomly selecting an image with the face photo and an image with the identification card photo from a corresponding image library;
extracting the face photo and the identification card photo from the selected image, wherein the identification card photo comprises an interested area where the face photo is located;
adjusting the size and brightness of the extracted face picture based on the size and brightness of the region of interest of the identification card picture; and
and superposing the human face picture after size and brightness adjustment on the interested area of the identity card picture.
7. The method of claim 6, adjusting the size and brightness of the extracted face photograph further comprising:
shadows are added to the extracted face photos at random angles.
8. The method of claim 6 or 7, further comprising:
gaussian noise and Gaussian blur are added to the composite image.
9. The method of any one of the preceding claims,
each training sample comprises a rectangle, an
The face area of the identification card comprises a rectangle of the same size as the training sample.
10. The method of any of the preceding claims, wherein training the neural network further comprises:
identifying a set of negative training samples that provide false negative classification results;
adjusting a parameter of the neural network; and
retraining the neural network based on the identified set of negative training samples and the adjusted parameters.
11. A tamper detection system comprising:
at least one processor; and
a computer-readable memory coupled with the processor and having instructions stored thereon that are executable by the processor to:
extracting image data of a face area of an identity card from an image of the identity card;
applying a trained neural network to the extracted image data to compute a confidence score, wherein the trained neural network comprises:
a binary classifier trained to classify each of a plurality of training samples based on a face photograph of the training sample; and
a neural network parameter iteratively optimized based on classification results of the plurality of training samples; and
determining whether the identity card is tampered based on a comparison of the confidence score to a threshold.
12. The system of claim 11, wherein the instructions are executable by the processor to calculate the confidence score using a normalized exponential function.
13. The system of claim 11 or 12, wherein the memory further comprises instructions executable by the processor to: the threshold is determined from the labeled validation dataset such that the false acceptance rate of the binary classifier based on the threshold is equal to 0.01.
14. The system of claim 11 or 12, wherein the memory further comprises instructions executable by the processor to: the threshold is determined from the labeled validation dataset such that the false acceptance rate of the binary classifier based on the threshold is equal to 0.001.
15. The system of any of claims 11 to 14, wherein the training samples comprise:
a plurality of positive training samples obtained from real images of a plurality of identification cards, each having a respective picture of a face;
a plurality of negative training samples obtained from a plurality of composite images, wherein each composite image comprises a photograph of a human face superimposed on a photograph of an identification card.
16. The system of claim 15, wherein the face photograph of each composite image has a size and brightness that is adjusted based on a size and brightness of a region of interest in which the face photograph of the identification card photograph is located.
17. The system of claim 16, wherein the face photograph of each composite image includes shadows added at random angles.
18. The system of claim 16 or 17, wherein each composite image comprises added gaussian noise and gaussian blur.
19. The system of any of claims 11 to 18, wherein each training sample comprises a rectangle and the face area of the identification card comprises a rectangle of the same size as the training sample.
20. An apparatus, comprising:
the receiving equipment is used for receiving the image of the identity card; and
a processing device to extract image data of a face region of the identification card from the received image, the processing device to provide the extracted image data to a trained neural network to calculate a confidence score and determine whether the identification card has been tampered with based on a comparison of the confidence score to a threshold, wherein the trained neural network comprises:
a binary classifier trained to classify each of a plurality of training samples based on a face photograph of the training sample; and
neural network parameters are iteratively optimized based on the classification results of the plurality of training samples.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
SG10202005989Q | 2020-06-23 | ||
SG10202005989Q | 2020-06-23 |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112597808A true CN112597808A (en) | 2021-04-02 |
Family
ID=75187595
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011388928.3A Pending CN112597808A (en) | 2020-06-23 | 2020-12-01 | Tamper detection method and system |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112597808A (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180107887A1 (en) * | 2016-10-14 | 2018-04-19 | ID Metrics Group Incorporated | Tamper detection for identification documents |
CN109754393A (en) * | 2018-12-19 | 2019-05-14 | 众安信息技术服务有限公司 | A kind of tampered image identification method and device based on deep learning |
CN110414670A (en) * | 2019-07-03 | 2019-11-05 | 南京信息工程大学 | A kind of image mosaic tampering location method based on full convolutional neural networks |
CN111191539A (en) * | 2019-12-20 | 2020-05-22 | 江苏常熟农村商业银行股份有限公司 | Certificate authenticity verification method and device, computer equipment and storage medium |
-
2020
- 2020-12-01 CN CN202011388928.3A patent/CN112597808A/en active Pending
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20180107887A1 (en) * | 2016-10-14 | 2018-04-19 | ID Metrics Group Incorporated | Tamper detection for identification documents |
CN109754393A (en) * | 2018-12-19 | 2019-05-14 | 众安信息技术服务有限公司 | A kind of tampered image identification method and device based on deep learning |
CN110414670A (en) * | 2019-07-03 | 2019-11-05 | 南京信息工程大学 | A kind of image mosaic tampering location method based on full convolutional neural networks |
CN111191539A (en) * | 2019-12-20 | 2020-05-22 | 江苏常熟农村商业银行股份有限公司 | Certificate authenticity verification method and device, computer equipment and storage medium |
Non-Patent Citations (1)
Title |
---|
PENGZHOU等: ""Learning Rich Features for Image Manipulation Detection"", 《IEEE XPLORE》, 31 December 2018 (2018-12-31), pages 1 - 9 * |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2022161286A1 (en) | Image detection method, model training method, device, medium, and program product | |
US9152860B2 (en) | Methods and apparatus for capturing, processing, training, and detecting patterns using pattern recognition classifiers | |
US11023708B2 (en) | Within document face verification | |
US10210415B2 (en) | Method and system for recognizing information on a card | |
KR102370694B1 (en) | Anti-counterfeiting detection method and device, electronic device, storage medium | |
KR102406432B1 (en) | Identity authentication methods and devices, electronic devices and storage media | |
US20210019519A1 (en) | Detection of fraudulently generated and photocopied credential documents | |
JP2019534526A (en) | How to detect falsification of ID | |
US10403076B2 (en) | Method for securing and verifying a document | |
WO2019061658A1 (en) | Method and device for positioning eyeglass, and storage medium | |
US9489566B2 (en) | Image recognition apparatus and image recognition method for identifying object | |
US10043071B1 (en) | Automated document classification | |
CN110059607B (en) | Living body multiplex detection method, living body multiplex detection device, computer equipment and storage medium | |
Prabhu et al. | Recognition of Indian license plate number from live stream videos | |
CN115035533B (en) | Data authentication processing method and device, computer equipment and storage medium | |
Gonzalez et al. | Improving presentation attack detection for ID cards on remote verification systems | |
CN112597808A (en) | Tamper detection method and system | |
CN112434727A (en) | Identity document authentication method and system | |
CN111626244B (en) | Image recognition method, device, electronic equipment and medium | |
CN114820476A (en) | Identification card identification method based on compliance detection | |
CN112597810A (en) | Identity document authentication method and system | |
CN112613345A (en) | User authentication method and system | |
EP3671558A1 (en) | Authentication system, apparatus, and method therefor | |
CN111723612A (en) | Face recognition and face recognition network training method and device, and storage medium | |
CN111209863A (en) | Living body model training and human face living body detection method, device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination |