WO2024028450A1 - Système et procédé d'extraction de données - Google Patents

Système et procédé d'extraction de données Download PDF

Info

Publication number
WO2024028450A1
WO2024028450A1 PCT/EP2023/071575 EP2023071575W WO2024028450A1 WO 2024028450 A1 WO2024028450 A1 WO 2024028450A1 EP 2023071575 W EP2023071575 W EP 2023071575W WO 2024028450 A1 WO2024028450 A1 WO 2024028450A1
Authority
WO
WIPO (PCT)
Prior art keywords
document
data extraction
user interface
application
electronic device
Prior art date
Application number
PCT/EP2023/071575
Other languages
English (en)
Inventor
Eldho ABRAHAM
Thomas Christopher Wolfgang LANDGREBE
Joshua MERRITT
Hristijan GJORGJIEVSKI
Original Assignee
Amadeus S.A.S.
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Amadeus S.A.S. filed Critical Amadeus S.A.S.
Publication of WO2024028450A1 publication Critical patent/WO2024028450A1/fr

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/12Detection or correction of errors, e.g. by rescanning the pattern
    • G06V30/127Detection or correction of errors, e.g. by rescanning the pattern with the intervention of an operator
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/12Detection or correction of errors, e.g. by rescanning the pattern
    • G06V30/133Evaluation of quality of the acquired characters
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/16Image preprocessing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/42Document-oriented image-based pattern recognition based on the type of document

Definitions

  • the present invention relates generally to a data extraction system and method. More specifically, the present invention relates to a system and method for controlling an electronic device to extract data from an official document (including passports and the like) for purposes such as identity verification and check-in automation.
  • an official document including passports and the like
  • the travel industry is moving towards more automated and contactless procedures for customers to check into their chosen mode of transportation-such as an international flight.
  • One approach to automated check-in involves the customer utilising a software application on their own device to scan the official document (such as a passport) that would otherwise be manually checked by a check-in agent.
  • the application extracts data from the official document and transmits the data to the airline’s check-in system for verification and actioning.
  • biometric enrolment the data extraction application, in addition to extracting bibliographic data, captures biometric information; typically in the form of a facial scan taken from the customer’s photograph in the official document.
  • identity verification the customer can submit an additional digital photograph through the data extraction application that the check-in system compares against the recorded facial scan.
  • Biometric enrolment systems allow the customer to check into an international flight even before arriving at the airport, and then to board the flight after only being photographed by the airport security system.
  • the airport security system is able to verify the customer’s identity and check-in status by comparing the preboarding photograph against the stored biometric information.
  • the present invention aims to provide a data extraction system and method that are more convenient for the customer to use. According to the present invention, there are provided a data extraction system and method as defined in the claims.
  • a data extraction system comprising a data extraction application, the data extraction application including a trained document classification model and computer program code which, when executed by a processor of an electronic device, causes the processor to: present a user interface on the electronic device instructing a user to capture an image of a document; submit the captured image of the document to the document classification model; receive from the document classification model a document category value for the document; process the document category to select a next data extraction operation to be performed by the data extraction application; and launch the selected data extraction operation.
  • the present disclosure provides a data extraction application that utilizes an on- device document classification machine learning model to classify an official document from a captured image thereof, and control a subsequent data extraction process based on the classification.
  • the present disclosure not only relieves the traveller from manually entering the type of document into the application (such as by selecting the document type from a list), but also improves the efficiency of the overall data extraction operation.
  • the processor selects the next data extraction operation from a set of data extraction operations, the set comprising: presenting a retry user interface instructing the user to capture another image of the document; performing optical character recognition on the document, presenting a manipulation user interface instructing the user to physically manipulate the document; and presenting a successful extraction user interface informing the user of a successful data extraction from the document.
  • the successful extraction user interface presents, on the electronic device, data extracted from the document.
  • the successful extraction user interface may include a confirmation user interface element manipulable by the user to confirm that data extracted from the document is correct.
  • the manipulation user interface instructs the user to present another face of the document to the electronic device for capture.
  • the computer program code may further cause the processor to perform an image capture routine on the another face of the document once captured by the electronic device.
  • the computer program code further causes the processor to perform optical character recognition on the another face of the document once captured by the electronic device.
  • the manipulation user interface instructs the user to present another part of the same face of the document to the electronic device for capture.
  • the computer program code further causes the processor to: input features of the document classification model as heuristics to a heuristic- filtering algorithm; and execute the heuristic-filtering algorithm on the captured image prior to performing optical character recognition on the captured image.
  • the features of the document classification model typically include the presence of a machine-readable zone and the structure of the machine readable zone.
  • the data extraction application and the document classification model are comprised in a web application that is executable by a web browser installed on the electronic device.
  • a method for extracting data from a document comprising: delivering a data extraction application and a trained document classification model to an electronic device; the data extraction application presenting a user interface on the electronic device instructing a user to capture an image of a document; the data extraction application submitting the captured image of the document to the document classification model; the data extraction application receiving from the document classification model a document category value for the document; the data extraction application processing the document category value to select a next data extraction operation to be performed by the data extraction application; and the data extraction application launching the selected data extraction operation.
  • the data extraction application selects the next data extraction operation from a set of data extraction operations, the set comprising: presenting a retry user interface instructing the user to capture another image of the document; performing optical character recognition on the document, presenting a manipulation user interface instructing the user to physically manipulate the document; and presenting a successful extraction user interface informing the user of a successful data extraction from the document.
  • the successful extraction user interface presents, on the electronic device, data extracted from the document.
  • the successful extraction user interface may include a confirmation user interface element manipulable by the user to confirm that data extracted from the document is correct.
  • the manipulation user interface instructs the user to present another face of the document to the electronic device for capture.
  • the data extraction application performs an image capture routine on the another face of the document once captured by the electronic device.
  • the data extraction application performs optical character recognition on the another face of the document once captured by the electronic device.
  • the manipulation user interface typically instructs the user to present another part of the same face of the document to the electronic device for capture.
  • the document extraction application inputs features of the document classification model as heuristics to a heuristic-filtering algorithm; and executes the heuristic-filtering algorithm on the captured image prior to performing optical character recognition on the captured image.
  • the features of the document classification model may include the presence of a machine-readable zone and the structure of the machine readable zone.
  • the data extraction application and the document classification model are comprised in a web application that is executable by a web browser installed on the electronic device.
  • Figure l is a schematic representation of a computing environment in which aspects of the present invention can be implemented
  • FIG. 2 is a block diagram illustrating the modules of a Data Extraction Application in accordance with an embodiment of the present invention
  • Figure 3 is a flow chart illustrating a data extraction and transmission process performed by a Data Extraction Application in accordance with an embodiment of the present invention
  • Figure 4 is an illustration of a User Interface generated by the Data Extraction Application in accordance with an embodiment of the present invention
  • FIG. 5 is an illustration of a Verify Details User Interface generated by the Data Extraction Application in accordance with an embodiment of the present invention
  • FIG. 6 is a flow chart illustrating a Card Flip Operation performed by the Data Extraction Application in accordance with an embodiment of the present invention
  • FIG. 7 is an illustration of a Card Flip User Interface generated by the Data Extraction Application in accordance with an embodiment of the present invention.
  • Figure 8 is an illustration of Lighting Compensation and Binarization Operations performed by the Data Extraction Application in accordance with an embodiment of the present invention
  • FIG. 9 is an illustration of CCL and Clustering Operations performed by the Data Extraction Application in accordance with an embodiment of the present invention.
  • Figure 10 is an illustration of Heuristics Filtering Operations performed by the Data Extraction Application in accordance with an embodiment of the present invention
  • Figure 11 illustrates an example homography calculation performed by the Data Extraction Application in accordance with an embodiment of the present invention
  • Figure 12 is a block diagram of a computer system suitable for implementing an embodiment of the present invention.
  • FIG. 1 illustrates a computing environment 100 in which aspects of the present invention are implemented.
  • the environment 100 is a networked environment comprising an Automated Check-In Server 102 in communication with a Client System 104 over one or more communication networks 106. Aspects of the computer processing described below are performed by a Server Application 108 executing on the Automated Check-In Server 102 and a Data Extraction Application 112 executing on the Client System 104.
  • Automated Check-In Server 102 further includes a Data Storage 110 on which data collected by the Data Extraction Application 112 and transmitted to the Automated Check-In Server 102 is stored.
  • Data storage 110 is typically a storage medium such as a hard drive (or collection of hard drives).
  • a database management system (not shown) executing on Automated Check-In- Server 102 implements a database on Data Storage 110 for storing and retrieving data.
  • Automated Check-In Server 102 has been illustrated as a single system.
  • Automated Check-In Server 102 can, however, be a scalable server system comprising multiple nodes which can be commissioned/decommissioned based on processing demands.
  • server systems are server computers that provide greater resources (e.g. processing, memory, network bandwidth) in comparison to client systems.
  • Data Storage 110 is illustrated as part of the Automated Check-In Server.
  • the Data Storage 110 could be a separate system in operative networked communication with the Automated Check-In Server 102.
  • the Data Storage could be a networked-attached storage device, an entirely separate storage system accessed via a database management system, or any other appropriate data storage mechanism.
  • the Server Application 108 performs various operations in response to commands received from (and initiated at) the Data Extraction Application 112. As such, when executed by the Automated Check-In Server 102, the Server Application 108 configures the Automated Check-In Server 102 to provide server-side functionality to the Data Extraction Operation. To provide this functionality, the Server Application 108 comprises one or more suitable application programs, libraries, or other software infrastructure.
  • the Server Application 108 will typically be, or interact with, a web server such as a server implemented with the node.js runtime environment.
  • the Data Extraction Application 112 is a native application of the Client System 104
  • the Server Application 108 will typically be, or interact with, an application server.
  • Automated Check-In Server system 102 may be provided with both web server and application server applications to enable it to serve both web browser and native client applications.
  • the Automated Check-In Server 102 and Client System 104 communicate data between each other either directly or indirectly through one or more Communications Networks 106.
  • Communications network 106 may comprise a local area network (LAN), a public network (such as the Internet), or a combination of networks.
  • LAN local area network
  • public network such as the Internet
  • Client System 104 can be any type of computer system, including a desktop computer or laptop computer, it will more commonly be a smartphone or a tablet device with an integrated or connected camera.
  • the Data Extraction Application 112 configures the Client System 104 to provide client-side data extraction functionality and interact with the Automated Check- In Server 102.
  • the Data Extraction Application 112 may be provided to the Client System 104 as a web application that is executed by a general web browser application (such as Chrome, Edge, Safari or the like) that is installed thereon.
  • a general web browser application such as Chrome, Edge, Safari or the like
  • the Data Extraction Application 112 accesses the Server Application 108 via an appropriate uniform resource locator (URL) and communicates with the Server Application 108 via general world-wide-web protocols (e.g. http, https, ftp) and application programming interfaces (APIs) (e.g. REST APIs).
  • URL uniform resource locator
  • APIs application programming interfaces
  • the Data Extraction Application 112 is a native application, it is typically programmed to communicate with the Server Application 108 using defined API calls.
  • a given Client System 104 may have more than one client application 112 installed thereon, for example both a general web browser application and a dedicated programmatic client application.
  • a web application implementation can have certain advantages over a native application implementation.
  • the web application implementation can be more easily integrated with the backends of the various automated check-in systems that different airlines and airports offer.
  • the Data Extraction Application 112 utilizes a Document Classification Model 116 to perform data extraction operations.
  • the Document Classification Model 116 is a model that results from training a machine learning network such as a convolutional neural network. The particular training performed to build the Document Classification Model 116 is discussed below.
  • the Server Application 108 preferably provides the Document Classification Model 116 to the Client System 104 contemporaneously with the code of the Data Extraction Application 112. For example, when the Data Extraction Application 112 is a web application, the Server Application 108 transmits the Document Classification Model 116 to the Client System 104 along with the Javascript code of the Data Extraction Application 112. Providing the Document Classification Model 116 to the Client System 104 allows the Data Extraction Application 112 to perform inferencing locally, which speeds processing time and improves reliability.
  • the Data Extraction Application 112 includes a number of software modules which are described below by reference to Figure 2.
  • the Data Extraction Application 112 utilizes the Document Classification Model 116 to perform inferencing on an image of an official document. As described below, the customer operates the camera 120 of the Client System 104 to capture an image of an official document typically following instructions that the Data Extraction Application 112 provides on the display of the Client System. The Document Classification Model 116 is trained to recognise different types of official document.
  • the Document Classification Model 116 is trained to recognise documents that do not necessarily comply with the Doc 9303 standard.
  • the Document Classification Model 116 is trained on (and can thus recognise) Chinese identity cards with a one-row MRZ.
  • the TD1 is mostly used in identity cards.
  • the MRZ is on the reverse side of a TD1 document, which results in the necessity to capture both the front and the back of the document when performing data extraction.
  • Each issuing country can add optional content to the document; usually added on the reverse side of the document adjacent to the MRZ.
  • the MRZ of a TD1 document spans 3 lines, each of 30 characters. Standardised data elements are included in the MRZ, along with one or more check digits (for data verification) and any optional information.
  • a TD2 document is also used for identity cards and is larger, in terms of area, than a TD1 document.
  • the MRZ is on the front face of a TD2 document and comprises 2 lines, each of 35 characters.
  • the MRZ of a TD2 document includes standardised data elements, check digit/s and optional information.
  • the TD3 document is used in most travel passports issued by the majority of issuing agencies. Although TD3 documents are typically in the form of a booklet, the document includes a card with the information presented thereon.
  • the TD3 document is a larger size compared to a TD1 document, and includes a 2-line MRZ, each of 44 characters.
  • the MRZ of a TD3 document includes standardised fields, check digit/s and optional information.
  • TD3 documents also include a photograph of the document’s owner located in the Visual Inspection Zone (VIZ) of the document.
  • VIPZ Visual Inspection Zone
  • the exemplified Document Classification Model 116 is capable of predicting whether an image of a document is of one of the following categories:
  • the Document Classification Model 116 outputs the result of a prediction operation as a data value of the type: document category.
  • the exemplified document categories are TDI Front, TDl Back, TD3 and Chinese. Those skilled in the art will appreciate that with appropriate training, the Document Classification Model 116 can function to predict documents of other categories.
  • the Document Classification Model 116 like the Data Extraction Application 112 is distributed to the Server Application 108 (for eventual serving to the Client System 104) in the form of portable pre-compiled binary-code, such as Web Assembly (or WASM).
  • portable pre-compiled binary-code such as Web Assembly (or WASM).
  • the Data Extraction Application 112 further includes a Preprocessing module 118.
  • Preprocessing Module 118 includes computer-executable code for preprocessing a document image prior to performing OCR on MRZ text.
  • the Preprocessing Module 118 includes a number of sub-modules that allow the Data Extraction Application 112 to locate an MRZ in a document image. In the exemplified embodiment, these sub-modules are:
  • the Data Extraction Application 112 further includes an Information Parser Module 130.
  • Information Parser Module 130 includes computer-executable code for performing OCR on the text of the MRZ that the Perprocessing Module 118 identifies.
  • the Information Parser Module 130 includes a Regular Expression (Regex) Module 134 for processing the OCR text that the OCR Module 130 generates. As described below, the Regex Module 134 utilizes the document category value when processing the OCR text.
  • Regular Expression (Regex) Module 134
  • Figure 3 conceptually depicts one embodiment of a computer-executable process 300 that the Data Extraction Application 112 performs to allow a traveller to scan an official document and transmit extracted data to an Automated Check-In Server 102.
  • the process commences at step 302, at which the Data Extraction Application 112 captures an image of the official document using the Client System 104’s camera.
  • the Data Extraction Application 112 presents a user interface 400 ( Figure 4) on the display of the Client System 104 that includes textual and graphical instructions to the traveller to manipulate the camera so as to locate the official document in a frame 402 of the user interface.
  • User Interface 400 also includes a Button 404 that the traveller operates to add a boarding pass to the data that is transmitted to the Automated Check-In Server 101.
  • step 304 Upon the Data Extraction Application 112 detecting the traveller taking a photograph of the official document, the proceeds to step 304, at which the Data Extraction Application 112 inputs the captured document image to the Document Classification Model 116.
  • the Document Classification Model 116 performs an initial inferencing operation on the captured image to determine if the Model recognises the image as one of an official document. In the event that the Document Classification Model 116 does not recognise the image as an official document, the process proceeds to step 307, at which the Data Extraction Application displays a Retry User Interface on the Client System.
  • the Retry User Interface can be similar (or identical) in appearance to the User Interface 400 by including textual and graphical instructions to assist the traveller to capture an image of the official document (step 302).
  • step 308 in the event that the Document Classification Model 116 recognises the captured image as an image of an official document.
  • the Document Classification Model 116 performs inferencing on the image to classify the image into a document category and assign an appropriate document category value to the document.
  • the Data Extraction Application 112 receives this document category value from the Document Classification Model 116.
  • step 310 the Data Extraction Application 112 processes the document category value.
  • the Data Extraction Application 112 performs a determination of whether the document category value indicates that the document is a TD1 document.
  • step 314 the Data Extraction Application 112 performs a Card Flip Operation.
  • the Card Flip Operation that the Data Extraction Application 112 performs is described below.
  • step 316 the Data Extraction Application 112 performs a determination of whether the document category value indicates that the document is a TD3 document.
  • step 318 the Data Extraction Application 112 performs an operation to capture the face of the TD3 document.
  • operation 318 involves the Data Extraction Application 112 pre-processing the image to identify an MRZ therein. Operation 318 also involves the Data Extraction Application 112 extracting the traveller’s photograph from the document.
  • step 320 at which the Data Extraction Application 112 performs optical character recognition (OCR) on the MRZ so as to capture the text in a suitable format (such as ASCII text).
  • OCR optical character recognition
  • the method then proceeds to step 322.
  • the method also proceeds to step 322 after the Data Extraction Application 112 performs the Card Flip Operation on a TD1 document.
  • the Data Extraction Application 112 displays a Verify Details User Interface on the display of the Client System 104.
  • An example Verify Details User Interface 500 is illustrated in Figure 5.
  • the Verify Details User Interface 500 lists the data that the Data Extraction Application 112 extracted from the official document.
  • the extracted data comprises the passenger’s: First Name, Last Name, Gender, Nationality, Date of birth, Document Number, Expiration Date, Issuing Authority and Document Type.
  • the data is extracted from the MRZ using the OCR operation.
  • the extracted data also includes the traveller’s photograph 502 from the official document.
  • the Verify Details User Interface 500 includes a Submit Details Button 504 that the traveller operates (in the event that the extracted data is correct) to have the Data Extraction Application 112 transmit the extracted data to the Automated Check-In Server 102.
  • the Verify Details User Interface includes an Edit Details Button 506 that the traveller operates to manually edit any incorrect extracted data.
  • step 324 the Data Extraction Application displays a Selfie Capture User Interface on the display of the Client System 104.
  • a Selfie Capture User Interface is illustrated in Figure 4, in the form of a “Take a Selfie Button 404”.
  • the Data Extraction Application 112 instructs the user to take a selfie (or select an existing photograph) and transmits the selfie to the Automated Check-In- Server 102.
  • the Server Application 108 utilises the submitted selfie to perform identify verification by comparing the selfie with the photograph 502 from the official document.
  • FIG. 6 An example of a Card Flip Operation 314 is illustrated by reference to Figure 6. As noted above, the Data Extraction Application 112 performs a Card Flip Operation when the document category value indicates that the document is a TD1 document.
  • the process 314 commences at step 602, at which the Document Classification Model 116 performs inferencing on the document image and determines whether the image is of the front face of the TD1 document. If the Document Classification Model determines that the image is of the front face of the document, the process proceeds to step 604, at which the Data Extraction Application 112 performs a Capture Face Operation.
  • the Capture Face Operation 604 involves capturing an image of the traveller’s photograph in the document.
  • step 606 the Data Extraction Application 112 displays a Card Flip User Interface on the display of the Client System 104.
  • An example of a Card Flip User Interface 700 is illustrated in Figure 7.
  • Interface 700 includes textual and graphical instructions to the traveller to flip the document over and take a photograph of the other side.
  • Interface 700 includes a Frame 702 in which the traveller is instructed to locate the image of the ID document.
  • step 608 the Data Extraction Application 112 performs OCR on the captured image.
  • the MRZ is on the rear face of the document.
  • step 610 the Document Classification Model 116 determines that the image is of the rear face of the TD1 document. the process proceeds to step 610, at which the Data Extraction Application 112 performs OCR on the captured image, as per. step 608.
  • step 612 the Data Extraction Application 112 displays the Card Flip User Interface on the display of the Client System 104, as per step 606.
  • step 614 the Data Extraction Application 112 performs a Capture Face operation, as per step 604.
  • Certain data augmentation techniques were utilised to train a neural network into a Document Classification Model that could be deployed as a web application and reliably classify official documents client-side.
  • an initial data set of images of TD1 -Front, TDl-Back and TD3 documents were augmented using background randomization and scan randomisation.
  • a training data set of Chinese Identification Documents was generated by performing row augmentation using interpolation from a set of TD3 sample documents. This resulted in a set of training examples with the visual characteristics of a TD3 document, but a single row MRZ.
  • the augmented training set of TD3 documents trained the neural network to recognise a document with a photograph and a 3 -row MRZ.
  • the augmented training set of TDl-Back images trained the neural network to recognise a 2-row MRZ.
  • FIG. 8 An example of the Lighting Compensation and Binarization operations performed on an image prior to OCR is illustrated in Figure 8. The operations involve performing ROI localization relative to a detected face.
  • Heuristics Filtering operations 950 performed on an image prior to OCR is illustrated in Figure 10. Heuristics Filtering is shown for both a TD1 document and a TD3 document, both with an MRZ in the OCRB font.
  • the detected characters and their corresponding clusters are filtered by tuning the following characteristics of the MRZ and its OCRB font:
  • the filtering algorithm involves:
  • Figure 12 provides a block diagram of a computer processing system 1200 configurable to implement embodiments and/or features described herein.
  • System 1200 is a general purpose computer processing system. It will be appreciated that Figure 12 does not illustrate all functional or physical components of a computer processing system. For example, no power supply or power supply interface has been depicted, however system 1200 will either carry a power supply or be configured for connection to a power supply (or both). It will also be appreciated that the particular type of computer processing system will determine the appropriate hardware and architecture, and alternative computer processing systems suitable for implementing features of the present disclosure may have alternative components to those depicted.
  • Computer processing system 1200 includes at least one processing unit 1202.
  • the processing unit 1202 may be a single computer processing device (e.g. a central processing unit, graphics processing unit, or other computational device), or may include a plurality of computer processing devices. In some instances all processing will be performed by processing unit 1202, however in other instances processing may also be performed by remote processing devices accessible and useable (either in a shared or dedicated manner) by the system 1200.
  • system 1200 includes a system memory 1206 (e.g. a BIOS), volatile memory 1208 (e.g. random access memory such as one or more DRAM modules), and non-volatile memory 1210 (e.g. one or more hard disk or solid state drives).
  • system memory 1206 e.g. a BIOS
  • volatile memory 1208 e.g. random access memory such as one or more DRAM modules
  • non-volatile memory 1210 e.g. one or more hard disk or solid state drives.
  • System 1200 also includes one or more interfaces, indicated generally by 1212, via which system 1200 interfaces with various devices and/or networks.
  • other devices may be integral with system 1200, or may be separate.
  • connection between the device and system 1200 may be via wired or wireless hardware and communication protocols, and may be a direct or an indirect (e.g. networked) connection.
  • Wired connection with other devices/networks may be by any appropriate standard or proprietary hardware and connectivity protocols.
  • system 1200 may be configured for wired connection with other devices/communications networks by one or more of USB; FireWire; eSATA; Thunderbolt; Ethernet; OS/2; Parallel; Serial; HDMI; DVI; VGA; SCSI. Other wired connections are possible.
  • Wireless connection with other devices/networks may similarly be by any appropriate standard or proprietary hardware and communications protocols.
  • system 1200 may be configured for wireless connection with other devices/communications networks using one or more of infrared; Bluetooth; Wi-Fi; near field communications (NFC); Global System for Mobile Communications (GSM), Enhanced Data GSM Environment (EDGE), long term evolution (LTE), wideband code division multiple access (W-CDMA), code division multiple access (CDMA).
  • GSM Global System for Mobile Communications
  • EDGE Enhanced Data GSM Environment
  • LTE long term evolution
  • W-CDMA wideband code division multiple access
  • CDMA code division multiple access
  • devices to which system 1200 connects include one or more input devices to allow data to be input into/received by system 1200 for processing by the processing unit 1202, and one or more output device to allow data to be output by system 1200.
  • Example devices are described below, however it will be appreciated that not all computer processing systems will include all mentioned devices, and that additional and alternative devices to those mentioned may well be used.
  • system 1200 may include or connect to one or more input devices by which information/data is input into (received by) system 1200.
  • input devices may include keyboards, mice, trackpads, microphones, accelerometers, proximity sensors, GPS devices and the like.
  • System 1200 may also include or connect to one or more output devices controlled by system 1200 to output information.
  • output devices may include devices such as a CRT displays, LCD displays, LED displays, plasma displays, touch screen displays, speakers, vibration modules, LEDs/other lights, and such like.
  • System 1200 may also include or connect to devices which may act as both input and output devices, for example memory devices (hard drives, solid state drives, disk drives, compact flash cards, SD cards and the like) which system 1200 can read data from and/or write data to, and touch screen displays which can both display (output) data and receive touch signals (input).
  • memory devices hard drives, solid state drives, disk drives, compact flash cards, SD cards and the like
  • touch screen displays which can both display (output) data and receive touch signals (input).
  • System 1200 may also connect to one or more communications networks (e.g. the Internet, a local area network, a wide area network, a personal hotspot etc.) to communicate data to and receive data from networked devices, which may themselves be other computer processing systems.
  • communications networks e.g. the Internet, a local area network, a wide area network, a personal hotspot etc.
  • System 1200 may be any suitable computer processing system such as, by way of non-limiting example, a server computer system, a desktop computer, a laptop computer, a netbook computer, a tablet computing device, a mobile/smart phone, a personal digital assistant, a personal media player, a set-top box, a games console, [note repetition in computer processing system description]
  • system 1200 will include at least user input and output devices 1214 and a communications interface 1216 for communication with a network such as network 106 of environment 100.
  • System 1200 stores or has access to computer applications (also referred to as software or programs), i.e. computer readable instructions and data which, when executed by the processing unit 1202, configure system 1200 to receive, process, and output data.
  • Instructions and data can be stored on non-transient machine readable medium accessible to system 1200.
  • instructions and data may be stored on non-transient memory 1210. Instructions and data may be transmitted to/received by system 1200 via a data signal in a transmission channel enabled (for example) by a wired or wireless network connection.
  • Applications accessible to system 1200 will typically include an operating system application such as Microsoft Windows®, Apple OSX, Apple IOS, Android, Unix, or Linux.
  • an operating system application such as Microsoft Windows®, Apple OSX, Apple IOS, Android, Unix, or Linux.
  • System 1200 also stores or has access to applications which, when executed by the processing unit 1202, configure system 1200 to perform various computer- implemented processing operations described herein.
  • client system 104 includes a Data extraction application 112 which configures the client system 104 to perform the described client system operations.
  • Automated Check-In Server 102 includes a Server application 108 which configures the server system 102 to perform the described server system operations.

Abstract

La présente invention concerne un système et un procédé d'extraction de données permettant d'extraire des données d'un document. Le système d'extraction de données comprend une application d'extraction de données, l'application d'extraction de données comprenant un modèle de classification de documents entraîné et un code de programme informatique qui, lorsqu'il est exécuté par un processeur d'un dispositif électronique, amène le processeur à : présenter une interface utilisateur sur le dispositif électronique ordonnant à un utilisateur de capturer une image d'un document ; soumettre l'image capturée du document au modèle de classification de documents ; recevoir du modèle de classification de documents une valeur de catégorie de document pour le document ; traiter la valeur de catégorie de document pour sélectionner une opération d'extraction de données suivante à effectuer par l'application d'extraction de données ; et lancer l'opération d'extraction de données sélectionnée.
PCT/EP2023/071575 2022-08-03 2023-08-03 Système et procédé d'extraction de données WO2024028450A1 (fr)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
IN202241044316 2022-08-03
IN202241044316 2022-08-03

Publications (1)

Publication Number Publication Date
WO2024028450A1 true WO2024028450A1 (fr) 2024-02-08

Family

ID=87571039

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/EP2023/071575 WO2024028450A1 (fr) 2022-08-03 2023-08-03 Système et procédé d'extraction de données

Country Status (1)

Country Link
WO (1) WO2024028450A1 (fr)

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090052751A1 (en) * 2007-08-20 2009-02-26 International Business Machines Corporation Using a surface based computing device for verification of an identification document
US20140270536A1 (en) * 2013-03-13 2014-09-18 Kofax, Inc. Systems and methods for classifying objects in digital images captured using mobile devices
WO2023154393A1 (fr) * 2022-02-11 2023-08-17 Gtriip Global, Inc. Procédés et systèmes de vérification d'identification numérique et d'enregistrement sans contact

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090052751A1 (en) * 2007-08-20 2009-02-26 International Business Machines Corporation Using a surface based computing device for verification of an identification document
US20140270536A1 (en) * 2013-03-13 2014-09-18 Kofax, Inc. Systems and methods for classifying objects in digital images captured using mobile devices
WO2023154393A1 (fr) * 2022-02-11 2023-08-17 Gtriip Global, Inc. Procédés et systèmes de vérification d'identification numérique et d'enregistrement sans contact

Similar Documents

Publication Publication Date Title
EP3437019B1 (fr) Reconnaissance optique de caractères dans des documents structurés
US8995774B1 (en) Automated document recognition, identification, and data extraction
US9984471B2 (en) Label and field identification without optical character recognition (OCR)
US10043231B2 (en) Methods and systems for detecting and recognizing text from images
US9679225B2 (en) Extracting card data with linear and nonlinear transformations
CN109034050B (zh) 基于深度学习的身份证图像文本识别方法及装置
US11093748B2 (en) Visual feedback of process state
US9298981B1 (en) Categorizer assisted capture of customer documents using a mobile device
US20220050890A1 (en) Remotely verifying an identity of a person
US20160188783A1 (en) Computer-aided card design validation
US20150347839A1 (en) Identification verification using a device with embedded radio-frequency identification functionality
US10339373B1 (en) Optical character recognition utilizing hashed templates
JP2016541049A (ja) カードのocr画像のクライアント側フィルタリング
JP2016048444A (ja) 帳票識別プログラム、帳票識別装置、帳票識別システム、および帳票識別方法
US20220067585A1 (en) Method and device for identifying machine learning models for detecting entities
US20170093850A1 (en) Pharmacy authentication methods and systems
CN109388935A (zh) 单证验证方法及装置、电子设备及可读存储介质
US20170277423A1 (en) Information processing method and electronic device
WO2024028450A1 (fr) Système et procédé d'extraction de données
US20240005640A1 (en) Synthetic document generation pipeline for training artificial intelligence models
EP4244830A1 (fr) Segmentation sémantique pour classification de traits dans une application d'encrage
US11145064B2 (en) Technologies for detecting crop marks in electronic documents
US20220327437A1 (en) Information processing apparatus, information processing method, and computer readable recording medium
US11853925B2 (en) Information processing apparatus, information processing method, and computer readable recording medium to display information for a passenger at an airport
JP2021125040A (ja) 帳票仕分システム、帳票仕分方法、及びプログラム

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 23754192

Country of ref document: EP

Kind code of ref document: A1