WO2021012570A1 - 数据录入方法、装置、设备及存储介质 - Google Patents
数据录入方法、装置、设备及存储介质 Download PDFInfo
- Publication number
- WO2021012570A1 WO2021012570A1 PCT/CN2019/122812 CN2019122812W WO2021012570A1 WO 2021012570 A1 WO2021012570 A1 WO 2021012570A1 CN 2019122812 W CN2019122812 W CN 2019122812W WO 2021012570 A1 WO2021012570 A1 WO 2021012570A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- image
- text
- information
- text information
- word segmentation
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/33—Querying
- G06F16/3331—Query processing
- G06F16/334—Query execution
- G06F16/3344—Query execution using natural language analysis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q40/00—Finance; Insurance; Tax strategies; Processing of corporate or income taxes
- G06Q40/08—Insurance
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V20/00—Scenes; Scene-specific elements
- G06V20/60—Type of objects
- G06V20/62—Text, e.g. of license plates, overlay texts or captions on TV images
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
- G06V30/41—Analysis of document content
- G06V30/414—Extracting the geometrical structure, e.g. layout tree; Block segmentation, e.g. bounding boxes for graphics or text
Definitions
- This application relates to the field of data processing, and in particular to a data entry method, device, equipment and storage medium.
- paper files are not only easy to lose, but also easy to be damaged due to accidents such as water and fire, and the search efficiency of paper files is relatively low when searching. .
- the existing data storage method is usually to manually input the data on paper files into the data management system.
- the existing manual entry method is not only inefficient in entry, but also prone to errors and low entry accuracy.
- This application provides a data entry method, device, equipment, and storage medium, which provide an important reference for identifying fraudulent insurance fraud.
- this application provides a data entry method, which includes:
- the text information and the keywords corresponding to the text information are stored in the target database to complete data entry.
- this application also provides a data entry device, which includes:
- the to-be-processed image module is used to obtain the to-be-processed image corresponding to the text file, the to-be-processed image includes a text area;
- a bounding box extraction module for recognizing the text area in the image to be processed to obtain size information and position information of the text area; and determining the bounding box of the text area according to the size information and position information , And use the to-be-processed image in the bounding box as the bounding image;
- a text recognition module configured to input the boundary image into a pre-trained image text recognition model for text recognition, so as to output text information corresponding to the boundary image;
- the keyword extraction module is used to input the text information into a pre-trained keyword extraction model for keyword extraction, so as to obtain keywords corresponding to the text information;
- the data entry module is used to store the text information and the keywords corresponding to the text information in the target database to complete data entry.
- the present application also provides a computer device, the computer device includes a memory and a processor; the memory is used to store a computer program; the processor is used to execute the computer program and execute the The computer program realizes the above-mentioned data entry method.
- the present application also provides a computer-readable storage medium that stores a computer program, and when the computer program is executed by a processor, the processor implements the above-mentioned data entry method .
- This application discloses a data entry method, device, equipment, and storage medium.
- the image to be processed is obtained by obtaining the image to be processed corresponding to a text file, and then the bounding box of the text area in the image to be processed is extracted to obtain the boundary image, and the boundary image is input into the image text
- the recognition model performs text recognition, outputs text information, and then enters the text information into the keyword extraction model for keyword extraction to obtain keywords.
- the text information and keywords corresponding to the text information are stored in the target database to complete data entry .
- this solution realizes the input of text information by processing the to-be-processed image corresponding to the text file, which improves the efficiency and accuracy of data input.
- FIG. 1 is a schematic flowchart of a method for training an image text recognition model provided by an embodiment of the present application
- Figure 2 is a schematic flowchart of a data entry method provided by an embodiment of the present application.
- FIG. 3 is a schematic flowchart of sub-steps provided by an embodiment of the present application.
- FIG. 5 is a schematic flowchart of the sub-steps of extracting keywords from text information according to an embodiment of the present application
- FIG. 6 is a schematic flowchart of the steps of another data entry method provided by an embodiment of the present application.
- FIG. 7 is a schematic flowchart of sub-steps provided by an embodiment of the present application.
- FIG. 8 is a schematic block diagram of a model training device provided by an embodiment of the present application.
- FIG. 9 is a schematic block diagram of a data entry device provided in an embodiment of the present application.
- FIG. 10 is a schematic block diagram of another data entry device provided by an embodiment of the present application.
- FIG. 11 is a schematic block diagram of the structure of a computer device according to an embodiment of the application.
- the embodiments of the present application provide a data entry method, device, computer equipment, and storage medium.
- the data entry method can be applied to a terminal or a server to improve the accuracy and efficiency of data entry.
- FIG. 1 is a schematic flowchart of a method for training an image text recognition model provided by an embodiment of the present application.
- the image text recognition model is obtained by model training based on a convolutional neural network.
- a convolutional neural network Of course, other networks can also be used for training.
- GoogLeNet is used for model training to obtain the target recognition model.
- CNN Convolutional Neural Network
- Deep Residual Network Deep Residual Network
- DResNet Deep Residual Network
- LSTM Long Short-Term Memory
- the training method of the image text recognition model is used to train the image text recognition model for application in the data entry method.
- the training method of the image text recognition model includes step S101 and step S102.
- the text image sample is an image including a text area.
- the content of the text area can be contract content, ID card content, or other text content. The following will take the content of the text area as the contract content as an example for detailed description.
- the text image samples may be different types of contract images such as sales contract images, transportation contract images, technical contract images, etc. These contract images constitute text image samples for training the image text recognition model. Use a variety of different contract images as samples to improve the recognition accuracy of the image text recognition model.
- S102 Based on the convolutional neural network, perform model training according to the text image sample to obtain an image text recognition model, and use the image text recognition model as a preset image text recognition model.
- the constructed sample data is used for model training through GoogLeNet.
- GoogLeNet Specifically, directional propagation training can be used.
- the convolutional layer and pooling layer of GoogLeNet are used to extract features from the input sample data, and the fully connected layer is used as a classifier.
- the output of this classifier is the probability value of different images and texts.
- the convolutional neural network takes the trained sample data as input and goes through the forward propagation step (convolution, ReLU activation and pooling operations to forward propagation in the fully connected layer) , And finally get the output probability of each category.
- the terminal can be an electronic device such as a mobile phone, a tablet computer, a notebook computer, a desktop computer, a personal digital assistant, and a wearable device
- the server can be an independent server or a server cluster.
- the compression processing specifically includes pruning processing, quantization processing, and Huffman coding processing on the image text recognition model, so as to reduce the size of the image text recognition model, and then it is convenient to save it in a terminal with a smaller capacity.
- the training method provided in the above embodiment obtains an image text model by acquiring a variety of text image samples, and then based on a convolutional neural network, model training is performed according to the text image samples to obtain an image text model, and the obtained image text model is used as a pre-trained image text model Used in data entry methods, which can improve the accuracy of text recognition in images.
- FIG. 2 is a schematic flowchart of a data entry method provided by an embodiment of the present application.
- the data entry method is used to identify and enter data to improve entry efficiency and accuracy.
- the following takes the data entry method applied to the server as an example to introduce.
- the data entry method specifically includes: step S201 to step S205.
- the image to be processed includes a text area and a non-text area.
- the image to be processed corresponding to the text file can be collected by an image collection device such as a camera.
- the server acquires the image to be processed collected by the image acquisition device to facilitate subsequent operations on the image to be processed.
- S202 Extract a bounding box of the text area in the image to be processed to obtain a bounding image.
- the boundary image is an image to be processed in the boundary box.
- the boundary image is obtained, and then text recognition is performed on the extracted boundary image, which reduces the amount of calculation during contract image recognition, thereby improving the efficiency of contract image recognition.
- extracting the bounding box of the text region in the image to be processed to obtain the bounding image specifically includes sub-steps S202a and S202b.
- S202a Recognizing a text area in the image to be processed to obtain size information and position information of the text area.
- recognizing the text area in the image to be processed refers to automatic analysis of the text, table information, and position relationship in the image to be processed.
- the position information of the text area can be obtained by recognizing the text area in the image to be processed, and the size information of the text area can be determined according to the position information of the text area.
- the text area in the image to be processed is first recognized to obtain the area coordinates of the text in the text area.
- the area coordinates refer to the pixel position coordinates of the text area on the image to be processed; according to the area of the text in the text area
- the coordinates are calculated for the outsourcing area to obtain size information of the text area.
- the outsourcing area refers to the smallest area including the text area.
- S202b Determine a bounding box of the text area according to the size information and position information, and use an image to be processed in the bounding box as a boundary image.
- the bounding box of the text area can be determined according to the size information and position information of the text area, and then the image to be processed in the bounding box is extracted based on the bounding box, and the extracted image to be processed is used as the boundary image.
- the boundary image is input to a pre-trained image text recognition model for text recognition, and the image text recognition model outputs the text information included in the boundary image.
- S204 Input the text information into a pre-trained keyword extraction model to perform keyword extraction, so as to obtain keywords corresponding to the text information.
- the text information recognized from the boundary image is input into a pre-trained keyword extraction model to perform keyword extraction, so as to obtain corresponding keywords.
- the text information recognized from the border image is: The contract is valid from June 28, 2017 to June 28, 2019.
- the keyword extraction model performs keyword extraction, the corresponding keywords are obtained as: effective date June 28, 2017, and termination date June 28, 2018.
- performing keyword extraction on text information specifically includes sub-steps S204a to S204c.
- S204a Perform word segmentation on the text information and obtain a word segmentation result.
- the word segmentation result includes at least one word segmentation.
- the maximum matching algorithm can be used to segment the text information to obtain each segmentation contained in the text information, and the set of each segmentation contained in the text information is used as the segmentation result.
- the maximum matching algorithm refers to the dictionary as the basis, taking the longest word in the dictionary as the first scan string, and scanning in the dictionary. For example, if the longest word in the dictionary is "People's Republic of China" and a total of 7 Chinese characters, the maximum number of starting characters for matching is 7 Chinese characters. Then decrease it word by word, and look it up in the corresponding dictionary. It is understandable that in other embodiments, other methods may be used to segment the text information, for example, the Viterbi algorithm.
- S204b Input at least one of the word segmentation into a pre-trained keyword extraction model to obtain the importance weight corresponding to each word segmentation.
- the word vector corresponding to each word segmentation is input into the keyword extraction model, and the keyword extraction model outputs the importance weight of each word segmentation. The greater the importance weight, the stronger the importance of the word segmentation.
- the keyword extraction model can be obtained by selecting a preset number of standard contract samples on the Internet to perform model training on the deep learning model. For example, select 1,000 standard contract samples for model training.
- the keyword extraction model includes a two-way long and short-term memory (BLSTM) algorithm model, a maximum pooling (Max Pooling) algorithm model, and a Softmax algorithm model that are sequentially connected in sequence.
- the keywords can be keywords such as contract signing date, effective date, expiration date, and timeliness.
- the keyword extraction model extracts the keywords corresponding to the text information, so that when querying the text information, there is no need to traverse each word in the text information, avoiding excessive query data and improving query efficiency.
- step S204c includes step S204c1 and step S204c2.
- each word segmentation in the word segmentation result is sorted in descending order, and a preset number of word segments are selected as the keywords of the text information. For example, the top three words ranked by importance weight can be selected as the keywords of the ranking result.
- the signing date of a certain insurance contract is April 10, 2018, the effective date is April 20, 2018, the termination date is April 10, 2019, and the time limit is one year.
- the importance weight of the signing date is 0.28
- the importance weight of the effective date is 0.22
- the importance weight of the termination date is 0.42
- the importance weight of timeliness is 0.08.
- the sorting results are as follows: termination date April 10, 2019, contract date April 10, 2018, effective date April 20, 2018, and one-year statute of limitations.
- the keywords are the expiry date April 10, 2019, the signing date April 10, 2018, and the effective date April 20, 2018.
- the text information and the keywords are stored in the target database, thereby completing the entry of the text information, that is, completing the entry of contract-related information, which improves entry efficiency and entry accuracy.
- the data entry method obtains a to-be-processed image including a text area; then extracts the bounding box of the text area in the to-be-processed image to obtain a boundary image; and inputs the boundary image into a pre-trained image text recognition model for text Recognition, thereby outputting text information; inputting the text information into the keyword extraction model for keyword extraction, thereby obtaining keywords; finally storing the text information and keywords corresponding to the text information in the target database to complete data entry.
- Fig. 6 is a schematic flowchart of the steps of another data entry method provided by an embodiment of the present application.
- the data entry method can improve the accuracy and efficiency of data entry, and improve the efficiency of data query.
- the data entry method specifically includes: step S301 to step S308.
- the image to be processed includes a text area and a non-text area.
- the image to be processed corresponding to the text file can be collected by an image collection device such as a camera.
- the server acquires the image to be processed collected by the image acquisition device to facilitate subsequent operations on the image to be processed.
- the boundary image is an image to be processed in the boundary box.
- the boundary image is obtained, and then text recognition is performed on the extracted boundary image, which reduces the amount of calculation in contract image recognition, thereby improving the efficiency of contract image recognition.
- the sub-steps S302a to S302c are included.
- S302a Perform image smoothing processing and wavelet filtering processing on the image to be processed in the bounding box to obtain a denoised image.
- image smoothing processing and wavelet filtering processing can eliminate the noise points of the boundary image, thereby making the boundary image less blurred.
- the image smoothing process can use the neighborhood average method.
- the neighborhood average method is to assign a pixel and the average value of all pixels in its neighborhood to the corresponding pixel in the output image, so as to achieve the purpose of smoothing.
- the process is to make a window slide on the image, and the value of the center position of the window is used
- the average value of each point value in the window is replaced, that is, the gray value of a pixel is replaced by the average gray value of several pixels.
- image balance processing methods such as median filtering, can also be used.
- S302b Perform direction correction processing on the denoising image to obtain a corrected image.
- the received contract may have multiple rotation angles, it is necessary to rotate the contract to the correct direction to facilitate the next operation. Perform direction correction processing on the denoised image to make the contract rotate in the correct direction to obtain a corrected image.
- an image compression orthographic network can be used to rotate the denoised image to complete the direction correction of the denoised image, so that the contract text in the denoised image is in the correct direction and the corrected image is obtained.
- the image compression orthographic network is trained through machine learning and has the function of image rotation.
- S302c Perform background removal processing on the corrected image to obtain the background removal image as a boundary image.
- the corrected image includes a contract and a background image
- the interference of the background image in the corrected image can be removed by background removal processing.
- the boundary image is input to a pre-trained image text recognition model for text recognition, thereby outputting the text information included in the boundary image.
- the text classification model may identify that the classification category of the insurance contract includes information such as insurance subject matter and insurance protection risk.
- the subject of insurance includes: personal insurance or property insurance.
- Insurance protection risks include: life insurance, personal accident insurance or health insurance, etc.
- the insurance subject and insurance protection risks of the recognized text information are life insurance and life insurance respectively.
- the text information recognized from the boundary image is input into a pre-trained keyword extraction model to perform keyword extraction, so as to obtain corresponding keywords.
- step S304 and step S305 are not limited in this solution.
- S306 According to the classification category, store the text information and the keywords corresponding to the text information in a sub-database corresponding to the target database to complete data entry.
- each classification category corresponds to a sub-database
- the collection of each sub-database constitutes the target database.
- the sub-database corresponding to the classification category in the target database can be determined, and the text information and the keyword information corresponding to the text information are stored in the sub-database, thereby completing the efficient contract information Entry, entry accuracy is high and labor costs are reduced.
- the data query request includes query keywords, which may include classification categories, effective dates, and so on.
- the query request may be a user input voice information through a questioning mode, and a text with a query request converted from the voice information, or a data query request directly sent by the user to the server.
- S308 Match the keywords in the target database according to the query keywords to obtain target text information corresponding to the query keywords.
- mapping is performed in the target database according to the query keywords in the data query request. If the query keyword matches the keyword in the target database, the text information corresponding to the matched keyword is output as the target text information, thereby efficiently completing the query of the target contract.
- the query keyword includes a classification category
- the data entry method obtains the image to be processed including the text area; then extracts the bounding box of the text area in the image to be processed to obtain the boundary image; performs text recognition on the boundary image to output text information; Recognize the information category to obtain the classification category of the text information; perform keyword extraction on the text information to obtain the keywords; finally store the text information and the keywords corresponding to the text information in the target database according to the classification category to complete data entry . Then, the data query request is received, and the keyword in the target database is matched according to the data query request, thereby obtaining text information.
- the image to be processed corresponding to the text file the input of text information is realized, and the efficiency and accuracy of data input are improved. Categorize text information to improve the orderliness of data entry, and also improve query efficiency during data query.
- Fig. 8 is a schematic block diagram of a model training device provided by an embodiment of the present application.
- the model training device can be configured in a server and used to execute the aforementioned image text recognition model training method.
- the model training device 400 includes: a sample acquisition module 401 and a model training module 402.
- the sample acquisition module 401 is configured to acquire a text image sample, the text image sample being an image including a text area.
- the model training module 402 is configured to perform model training according to the text image samples based on the convolutional neural network to obtain an image text recognition model, and use the image text recognition model as a preset image text recognition model.
- FIG. 9 is a schematic block diagram of a data entry device according to an embodiment of the present application.
- the data entry device is used to execute the aforementioned data entry method.
- the data entry device can be configured in a server or a terminal.
- the server can be an independent server or a server cluster.
- the terminal can be an electronic device such as a mobile phone, a tablet computer, a notebook computer, a desktop computer, a personal digital assistant, and a wearable device.
- the data entry device 500 includes: a to-be-processed image module 501, a bounding box extraction module 502, a text recognition module 503, a keyword extraction module 504, and a data entry module 505.
- the to-be-processed image module 501 obtains a to-be-processed image corresponding to a text file, and the to-be-processed image includes a text area.
- the bounding box extraction module 502 is configured to extract the bounding box of the text area in the image to be processed to obtain a bounding image, and the bounding image is the image to be processed in the bounding box.
- the bounding box extraction module 502 includes a region recognition sub-module 5021 and a bounding box determination sub-module 5022.
- the area recognition sub-module 5021 is configured to recognize the text area in the image to be processed to obtain size information and position information of the text area.
- the bounding box determination sub-module 5022 is configured to determine the bounding box of the text area according to the size information and position information, and use the image to be processed in the bounding box as the bounding image.
- the text recognition module 503 is configured to input the boundary image into a pre-trained image text recognition model for text recognition, so as to output text information corresponding to the boundary image.
- the keyword extraction module 504 is configured to input the text information into a pre-trained keyword extraction model to perform keyword extraction, so as to obtain keywords corresponding to the text information.
- the keyword extraction module 504 includes a word segmentation result submodule 5041, a weight acquisition submodule 5042, a keyword determination submodule 5043.
- the word segmentation result sub-module 5041 is used to segment the text information and obtain the word segmentation result.
- the weight obtaining sub-module 5042 is configured to input at least one of the word segmentation into a pre-trained keyword extraction model to obtain the importance weight corresponding to each word segmentation.
- the keyword determination submodule 5043 is configured to select the word segmentation corresponding to the importance weight as the keyword of the text information according to the importance weight.
- the keyword determination submodule 5043 is specifically configured to: sort each of the word segmentation according to the importance weight to obtain a ranking result; and select the word segmentation as a keyword of the text information based on the ranking result .
- the data entry module 505 is configured to store the text information and keywords corresponding to the text information in a target database to complete data entry.
- FIG. 10 is a schematic block diagram of another data entry device provided by an embodiment of the present application.
- the data entry device is used to execute the aforementioned data entry method.
- the data entry device can be configured in the server.
- the data entry device 600 includes: an image to be processed module 601, a bounding box extraction module 602, a text recognition module 603, a category determination module 604, a keyword extraction module 605, a data entry module 606, and a request receiving module 607 ⁇ Query matching module 608.
- the image to be processed module 601 obtains an image to be processed corresponding to a text file, and the image to be processed includes a text area.
- the bounding box extraction module 602 is configured to extract a bounding box of the text area in the image to be processed to obtain a bounding image, and the boundary image is the image to be processed in the bounding box.
- the bounding box extraction module 602 includes a denoising image sub-module 6021, a corrected image sub-module 6022, and a background image sub-module 6023.
- the denoising image sub-module 6021 is used to perform image smoothing processing and wavelet filtering processing on the image to be processed in the bounding box to obtain a denoising image.
- the image correction sub-module 6022 is used to perform direction correction processing on the denoising image to obtain a corrected image.
- the background image removal sub-module 6023 is used to perform background removal processing on the corrected image to obtain the background image removed as a boundary image.
- the text recognition module 603 is configured to input the boundary image into a pre-trained image text recognition model for text recognition, so as to output text information corresponding to the boundary image.
- the category determination module 604 is configured to input the text information into a pre-trained text classification model for category recognition, so as to output a classification category corresponding to the text information.
- the keyword extraction module 605 is configured to input the text information into a pre-trained keyword extraction model to perform keyword extraction, so as to obtain keywords corresponding to the text information.
- the data entry module 606 is configured to store the text information and keywords corresponding to the text information in a sub-database corresponding to the target database according to the classification category, so as to complete data entry.
- the request receiving module 607 is configured to receive a data query request, where the data query request includes query keywords.
- the query matching module 608 is configured to match keywords in the target database according to the query keywords to obtain target text information corresponding to the query keywords.
- the above-mentioned data entry device can be implemented in the form of a computer program, and the computer program can be run on a computer device as shown in FIG. 11.
- FIG. 11 is a schematic block diagram of a structure of a computer device provided by an embodiment of the present application.
- the computer equipment can be a server or a terminal.
- the computer device includes a processor, a memory, and a network interface connected through a system bus, where the memory may include a non-volatile storage medium and an internal memory.
- the non-volatile storage medium can store an operating system and a computer program.
- the computer program includes program instructions, and when the program instructions are executed, the processor can execute any data entry method.
- the processor is used to provide computing and control capabilities and support the operation of the entire computer equipment.
- the internal memory provides an environment for the operation of the computer program in the non-volatile storage medium, and when the computer program is executed by the processor, the processor can execute any data entry method.
- the network interface is used for network communication, such as sending assigned tasks.
- the network interface is used for network communication, such as sending assigned tasks.
- FIG. 11 is only a block diagram of a part of the structure related to the solution of the present application, and does not constitute a limitation on the computer device to which the solution of the present application is applied.
- the specific computer device may Including more or fewer parts than shown in the figure, or combining some parts, or having a different arrangement of parts.
- the processor may be a central processing unit (Central Processing Unit, CPU), the processor may also be other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), and application specific integrated circuits (Application Specific Integrated Circuits). Circuit, ASIC), Field-Programmable Gate Array (FPGA) or other programmable logic devices, discrete gates or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor may be a microprocessor or the processor may also be any conventional processor.
- the processor is used to run a computer program stored in a memory to implement the following steps:
- the to-be-processed image includes a text area; extract the bounding box of the text area in the to-be-processed image to obtain a bounding image, and the bounding image is within the bounding box Image to be processed; input the boundary image into a pre-trained image text recognition model for text recognition to output text information corresponding to the boundary image; input the text information into a pre-trained keyword extraction model for keyword extraction , To obtain the keywords corresponding to the text information; store the text information and the keywords corresponding to the text information in the target database to complete data entry.
- the processor when the processor implements the extraction of the bounding box of the text area in the image to be processed to obtain a bounding image, it is used to implement:
- the image to be processed is used as the boundary image.
- the processor when the processor implements the extraction of the bounding box of the text area in the to-be-processed image to obtain a bounding image, it is used to implement:
- the processor when the processor implements the input of the text information into a pre-trained keyword extraction model for keyword extraction to obtain keywords corresponding to the text information, the processor is used to implement:
- the word segmentation result includes at least one word segmentation
- input at least one word segmentation into a pre-trained keyword extraction model to obtain the importance weight corresponding to each word segmentation;
- the word segmentation corresponding to the importance weight is selected as a keyword of the text information.
- the processor when the processor implements the selection of the word segmentation corresponding to the importance weight as a keyword of the text information according to the importance weight, the processor is configured to implement:
- the processor is used to run a computer program stored in the memory to implement the following steps:
- the to-be-processed image includes a text area; extract the bounding box of the text area in the to-be-processed image to obtain a bounding image, and the bounding image is within the bounding box Image to be processed; input the boundary image into a pre-trained image text recognition model for text recognition to output text information corresponding to the boundary image; input the text information into a pre-trained keyword extraction model for keyword extraction , To obtain keywords corresponding to the text information; store the text information and keywords corresponding to the text information in the target database to complete data entry; receive a data query request, the data query request includes Query keywords; match keywords in the target database according to the query keywords to obtain target text information corresponding to the query keywords.
- the processor is used to run a computer program stored in the memory to implement the following steps:
- the text image sample is an image including a text area; based on a convolutional neural network, perform model training according to the text image sample to obtain an image text recognition model, and use the image text recognition model as a preset Image text recognition model.
- the embodiments of the present application also provide a computer-readable storage medium, the computer-readable storage medium stores a computer program, the computer program includes program instructions, and the processor executes the program instructions to implement the present application Any data entry method provided in the embodiment.
- the computer-readable storage medium may be the internal storage unit of the computer device described in the foregoing embodiment, such as the hard disk or memory of the computer device.
- the computer-readable storage medium may also be an external storage device of the computer device, such as a plug-in hard disk, a smart memory card (SMC), or a secure digital (Secure Digital, SD) equipped on the computer device. ) Card, Flash Card, etc.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Business, Economics & Management (AREA)
- General Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Accounting & Taxation (AREA)
- Finance (AREA)
- Databases & Information Systems (AREA)
- Geometry (AREA)
- Economics (AREA)
- Evolutionary Computation (AREA)
- Evolutionary Biology (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Development Economics (AREA)
- Computer Graphics (AREA)
- Marketing (AREA)
- Strategic Management (AREA)
- Technology Law (AREA)
- General Business, Economics & Management (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computational Linguistics (AREA)
- Character Input (AREA)
- Image Analysis (AREA)
Abstract
Description
Claims (20)
- 一种数据录入方法,所述方法包括:获取文本文件对应的待处理图像,所述待处理图像包括文本区域;对所述待处理图像中的文本区域进行识别,以获取所述文本区域的尺寸信息和位置信息;根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像;将所述边界图像输入预先训练的图像文本识别模型进行文本识别,以输出与所述边界图像对应的文本信息;将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词;将所述文本信息和与所述文本信息对应的关键词存储至目标数据库中,以完成数据录入。
- 根据权利要求1所述的数据录入方法,其中,所述根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像,包括:对所述边界框内的待处理图像进行图像平滑处理和小波滤波处理,以得到去噪图像;对所述去噪图像进行方向矫正处理,以得到矫正图像;对所述矫正图像进行去背景处理,以得到去背景图像作为边界图像。
- 根据权利要求1所述的数据录入方法,其中,所述将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词,包括:对所述文本信息进行分词并得到分词结果,所述分词结果包括至少一个分词;将至少一个所述分词分别输入预先训练好的关键词提取模型,以获取各所述分词对应的重要性权重;根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词。
- 根据权利要求3所述的数据录入方法,其中,所述根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词,包括:根据所述重要性权重对各所述分词进行排序,以获得排序结果;基于所述排序结果选取分词作为所述文本信息的关键词。
- 根据权利要求1所述的数据录入方法,其中,还包括:接收数据查询请求,所述数据查询请求包括查询关键词;根据所述查询关键词匹配所述目标数据库中的关键词,以获取与所述查询关键词对应的目标文本信息。
- 根据权利要求1所述的数据录入方法,其中,还包括:获取文本图像样本,所述文本图像样本为包括文本区域的图像;基于卷积神经网络,根据所述文本图像样本进行模型训练以得到图像文本识别模型,并将所述图像文本识别模型作为预设的图像文本识别模型。
- 根据权利要求1所述的数据录入方法,其中,所述关键词包括合同签约日期、生效日期、终止日期和时效。
- 一种数据录入装置,所述装置包括:待处理图像模块,用于获取文本文件对应的待处理图像,所述待处理图像包括文本区域;边界框提取模块,用于对所述待处理图像中的文本区域进行识别,以获取所述文本区域的尺寸信息和位置信息;及根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像;文本识别模块,用于将所述边界图像输入预先训练的图像文本识别模型进行文本识别,以输出与所述边界图像对应的文本信息;关键词提取模块,用于将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词;数据录入模块,用于将所述文本信息和与所述文本信息对应的关键词存储至目标数据库中,以完成数据录入。
- 一种计算机设备,所述计算机设备包括存储器和处理器;所述存储器用于存储计算机程序;所述处理器,用于执行所述计算机程序并在执行所述计算机程序时实现如下步骤:获取文本文件对应的待处理图像,所述待处理图像包括文本区域;对所述待处理图像中的文本区域进行识别,以获取所述文本区域的尺寸信息和位置信息;根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像;将所述边界图像输入预先训练的图像文本识别模型进行文本识别,以输出与所述边界图像对应的文本信息;将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词;将所述文本信息和与所述文本信息对应的关键词存储至目标数据库中,以完成数据录入。
- 如权利要求9所述的计算机设备,其中,所述处理器在实现所述根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像时,用于实现:对所述边界框内的待处理图像进行图像平滑处理和小波滤波处理,以得到去噪图像;对所述去噪图像进行方向矫正处理,以得到矫正图像;对所述矫正图像进行去背景处理,以得到去背景图像作为边界图像。
- 如权利要求9所述的计算机设备,其中,所述处理器在实现所述将所述 文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词时,用于实现:对所述文本信息进行分词并得到分词结果,所述分词结果包括至少一个分词;将至少一个所述分词分别输入预先训练好的关键词提取模型,以获取各所述分词对应的重要性权重;根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词。
- 如权利要求11所述的计算机设备,其中,所述处理器在实现所述根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词时,用于实现:根据所述重要性权重对各所述分词进行排序,以获得排序结果;基于所述排序结果选取分词作为所述文本信息的关键词。
- 如权利要求9所述的计算机设备,其中,所述处理器还用于实现:接收数据查询请求,所述数据查询请求包括查询关键词;根据所述查询关键词匹配所述目标数据库中的关键词,以获取与所述查询关键词对应的目标文本信息。
- 如权利要求9所述的计算机设备,其中,所述处理器还用于实现:获取文本图像样本,所述文本图像样本为包括文本区域的图像;基于卷积神经网络,根据所述文本图像样本进行模型训练以得到图像文本识别模型,并将所述图像文本识别模型作为预设的图像文本识别模型。
- 如权利要求9所述的计算机设备,其中,所述关键词包括合同签约日期、生效日期、终止日期和时效。
- 一种计算机可读存储介质,所述计算机可读存储介质存储有计算机程序,所述计算机程序被处理器执行时使所述处理器实现如下步骤:获取文本文件对应的待处理图像,所述待处理图像包括文本区域;对所述待处理图像中的文本区域进行识别,以获取所述文本区域的尺寸信息和位置信息;根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像;将所述边界图像输入预先训练的图像文本识别模型进行文本识别,以输出与所述边界图像对应的文本信息;将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词;将所述文本信息和与所述文本信息对应的关键词存储至目标数据库中,以完成数据录入。
- 如权利要求16所述的计算机可读存储介质,其中,所述处理器在实现所述根据所述尺寸信息和位置信息确定所述文本区域的边界框,并将所述边界框内的待处理图像作为边界图像时,用于实现:对所述边界框内的待处理图像进行图像平滑处理和小波滤波处理,以得到 去噪图像;对所述去噪图像进行方向矫正处理,以得到矫正图像;对所述矫正图像进行去背景处理,以得到去背景图像作为边界图像。
- 如权利要求16所述的计算机可读存储介质,其中,所述处理器在实现所述将所述文本信息输入预先训练的关键词提取模型进行关键词提取,以获取与所述文本信息对应的关键词时,用于实现:对所述文本信息进行分词并得到分词结果,所述分词结果包括至少一个分词;将至少一个所述分词分别输入预先训练好的关键词提取模型,以获取各所述分词对应的重要性权重;根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词。
- 如权利要求18所述的计算机可读存储介质,其中,所述处理器在实现所述根据所述重要性权重,选取与所述重要性权重对应的所述分词作为所述文本信息的关键词时,用于实现:根据所述重要性权重对各所述分词进行排序,以获得排序结果;基于所述排序结果选取分词作为所述文本信息的关键词。
- 如权利要求16所述的计算机可读存储介质,其中,所述处理器还用于实现:接收数据查询请求,所述数据查询请求包括查询关键词;根据所述查询关键词匹配所述目标数据库中的关键词,以获取与所述查询关键词对应的目标文本信息。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910663209.9A CN110555372A (zh) | 2019-07-22 | 2019-07-22 | 数据录入方法、装置、设备及存储介质 |
CN201910663209.9 | 2019-07-22 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021012570A1 true WO2021012570A1 (zh) | 2021-01-28 |
Family
ID=68735720
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/122812 WO2021012570A1 (zh) | 2019-07-22 | 2019-12-03 | 数据录入方法、装置、设备及存储介质 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110555372A (zh) |
WO (1) | WO2021012570A1 (zh) |
Cited By (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112966583A (zh) * | 2021-02-26 | 2021-06-15 | 深圳壹账通智能科技有限公司 | 图像处理方法、装置、计算机设备和存储介质 |
CN113094473A (zh) * | 2021-04-30 | 2021-07-09 | 平安国际智慧城市科技股份有限公司 | 关键词的权重计算方法、装置、计算机设备及存储介质 |
CN113505588A (zh) * | 2021-07-14 | 2021-10-15 | 京东科技控股股份有限公司 | 信息抽取方法、装置和电子设备 |
CN114187605A (zh) * | 2021-12-13 | 2022-03-15 | 苏州方兴信息技术有限公司 | 一种数据集成方法、装置和可读存储介质 |
WO2022267387A1 (zh) * | 2021-06-21 | 2022-12-29 | 梅卡曼德(北京)机器人科技有限公司 | 图像识别方法、装置、电子设备和存储介质 |
WO2023035332A1 (zh) * | 2021-09-08 | 2023-03-16 | 深圳前海环融联易信息科技服务有限公司 | 一种日期提取方法、装置、计算机设备及存储介质 |
CN115862036A (zh) * | 2022-12-14 | 2023-03-28 | 北京瑞莱智慧科技有限公司 | 信息干扰模型训练方法、信息干扰方法、相关装置及介质 |
CN116522011A (zh) * | 2023-05-16 | 2023-08-01 | 深圳九星互动科技有限公司 | 一种基于大数据的推送方法及推送系统 |
CN116757886A (zh) * | 2023-08-16 | 2023-09-15 | 南京尘与土信息技术有限公司 | 数据分析方法及分析装置 |
Families Citing this family (11)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN111507850A (zh) * | 2020-03-25 | 2020-08-07 | 上海商汤智能科技有限公司 | 核保方法及相关装置、设备 |
CN111401042B (zh) * | 2020-03-26 | 2023-04-14 | 支付宝(杭州)信息技术有限公司 | 一种训练文本关键内容提取模型的方法和系统 |
CN111507333B (zh) * | 2020-04-21 | 2023-09-15 | 腾讯科技(深圳)有限公司 | 一种图像矫正方法、装置、电子设备和存储介质 |
CN111652229B (zh) * | 2020-05-25 | 2023-09-12 | 泰康保险集团股份有限公司 | 一种信息输入方法、装置、电子设备及存储介质 |
CN111723816B (zh) * | 2020-06-28 | 2023-10-27 | 北京联想软件有限公司 | 一种教学笔记的获取方法及电子设备 |
CN111950353B (zh) * | 2020-06-30 | 2024-04-19 | 深圳市雄帝科技股份有限公司 | 印章文本识别方法、装置及电子设备 |
CN112381038B (zh) * | 2020-11-26 | 2024-04-19 | 中国船舶工业系统工程研究院 | 一种基于图像的文本识别方法、系统和介质 |
CN112560411A (zh) * | 2020-12-21 | 2021-03-26 | 深圳供电局有限公司 | 一种人员信息智能录入方法及其系统 |
CN113033832B (zh) * | 2021-01-21 | 2024-04-26 | 深圳市轱辘车联数据技术有限公司 | 一种汽修数据录入方法、装置、终端设备及可读存储介质 |
CN112906801A (zh) * | 2021-02-26 | 2021-06-04 | 北京房江湖科技有限公司 | 一种图像处理的方法、装置和存储介质 |
CN116894937B (zh) * | 2023-06-25 | 2024-02-06 | 德联易控科技(北京)有限公司 | 获取车轮定位仪参数的方法、系统及电子设备 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102317955A (zh) * | 2009-04-20 | 2012-01-11 | 万涛国际有限公司 | 基于图像的数据管理方法与系统 |
CN109492143A (zh) * | 2018-09-21 | 2019-03-19 | 平安科技(深圳)有限公司 | 图像数据处理方法、装置、计算机设备及存储介质 |
US20190155887A1 (en) * | 2017-11-22 | 2019-05-23 | Adp, Llc | Methods and devices for enabling computers to automatically enter information into a unified database from heterogeneous documents |
CN109977956A (zh) * | 2019-04-29 | 2019-07-05 | 腾讯科技(深圳)有限公司 | 一种图像处理方法、装置、电子设备以及存储介质 |
CN110032990A (zh) * | 2019-04-23 | 2019-07-19 | 杭州智趣智能信息技术有限公司 | 一种发票文本识别方法、系统及相关组件 |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107609057B (zh) * | 2017-08-25 | 2020-12-22 | 百度在线网络技术(北京)有限公司 | 一种获取商标图像的文字数据的方法与装置 |
CN108764226B (zh) * | 2018-04-13 | 2022-05-03 | 顺丰科技有限公司 | 图像文本识别方法、装置、设备及其存储介质 |
CN108664473A (zh) * | 2018-05-11 | 2018-10-16 | 平安科技(深圳)有限公司 | 文本关键信息的识别方法、电子装置及可读存储介质 |
-
2019
- 2019-07-22 CN CN201910663209.9A patent/CN110555372A/zh active Pending
- 2019-12-03 WO PCT/CN2019/122812 patent/WO2021012570A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102317955A (zh) * | 2009-04-20 | 2012-01-11 | 万涛国际有限公司 | 基于图像的数据管理方法与系统 |
US20190155887A1 (en) * | 2017-11-22 | 2019-05-23 | Adp, Llc | Methods and devices for enabling computers to automatically enter information into a unified database from heterogeneous documents |
CN109492143A (zh) * | 2018-09-21 | 2019-03-19 | 平安科技(深圳)有限公司 | 图像数据处理方法、装置、计算机设备及存储介质 |
CN110032990A (zh) * | 2019-04-23 | 2019-07-19 | 杭州智趣智能信息技术有限公司 | 一种发票文本识别方法、系统及相关组件 |
CN109977956A (zh) * | 2019-04-29 | 2019-07-05 | 腾讯科技(深圳)有限公司 | 一种图像处理方法、装置、电子设备以及存储介质 |
Cited By (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112966583A (zh) * | 2021-02-26 | 2021-06-15 | 深圳壹账通智能科技有限公司 | 图像处理方法、装置、计算机设备和存储介质 |
CN113094473A (zh) * | 2021-04-30 | 2021-07-09 | 平安国际智慧城市科技股份有限公司 | 关键词的权重计算方法、装置、计算机设备及存储介质 |
WO2022267387A1 (zh) * | 2021-06-21 | 2022-12-29 | 梅卡曼德(北京)机器人科技有限公司 | 图像识别方法、装置、电子设备和存储介质 |
CN113505588A (zh) * | 2021-07-14 | 2021-10-15 | 京东科技控股股份有限公司 | 信息抽取方法、装置和电子设备 |
WO2023035332A1 (zh) * | 2021-09-08 | 2023-03-16 | 深圳前海环融联易信息科技服务有限公司 | 一种日期提取方法、装置、计算机设备及存储介质 |
CN114187605A (zh) * | 2021-12-13 | 2022-03-15 | 苏州方兴信息技术有限公司 | 一种数据集成方法、装置和可读存储介质 |
CN115862036A (zh) * | 2022-12-14 | 2023-03-28 | 北京瑞莱智慧科技有限公司 | 信息干扰模型训练方法、信息干扰方法、相关装置及介质 |
CN115862036B (zh) * | 2022-12-14 | 2024-02-23 | 北京瑞莱智慧科技有限公司 | 信息干扰模型训练方法、信息干扰方法、相关装置及介质 |
CN116522011A (zh) * | 2023-05-16 | 2023-08-01 | 深圳九星互动科技有限公司 | 一种基于大数据的推送方法及推送系统 |
CN116522011B (zh) * | 2023-05-16 | 2024-02-13 | 深圳九星互动科技有限公司 | 一种基于大数据的推送方法及推送系统 |
CN116757886A (zh) * | 2023-08-16 | 2023-09-15 | 南京尘与土信息技术有限公司 | 数据分析方法及分析装置 |
CN116757886B (zh) * | 2023-08-16 | 2023-11-28 | 南京尘与土信息技术有限公司 | 数据分析方法及分析装置 |
Also Published As
Publication number | Publication date |
---|---|
CN110555372A (zh) | 2019-12-10 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021012570A1 (zh) | 数据录入方法、装置、设备及存储介质 | |
CN108804512B (zh) | 文本分类模型的生成装置、方法及计算机可读存储介质 | |
WO2021072885A1 (zh) | 识别文本的方法、装置、设备及存储介质 | |
CN109829453B (zh) | 一种卡证中文字的识别方法、装置以及计算设备 | |
WO2019085064A1 (zh) | 医疗理赔拒付方法、装置、终端设备及存储介质 | |
CN111460250B (zh) | 用于画像的数据的清洗方法、装置、介质及电子设备 | |
WO2018040068A1 (zh) | 基于知识图谱的语意分析系统及方法 | |
US20150379343A1 (en) | System and Method for Data Extraction and Searching | |
US11430202B2 (en) | Enhanced optical character recognition (OCR) image segmentation system and method | |
WO2020253506A1 (zh) | 合同内容的提取方法及装置、计算机设备、存储介质 | |
CN105808649A (zh) | 一种搜索结果排序方法及其设备 | |
US11727701B2 (en) | Techniques to determine document recognition errors | |
US9256805B2 (en) | Method and system of identifying an entity from a digital image of a physical text | |
CN113657274A (zh) | 表格生成方法、装置、电子设备、存储介质及产品 | |
CN114780746A (zh) | 基于知识图谱的文档检索方法及其相关设备 | |
CN114241501B (zh) | 影像文档处理方法、装置及电子设备 | |
US10963690B2 (en) | Method for identifying main picture in web page | |
WO2024051427A1 (zh) | 硬币识别方法、系统及存储介质 | |
CN111382254A (zh) | 电子名片推荐方法、装置、设备及计算机可读存储介质 | |
CN108073567A (zh) | 一种特征词提取处理方法、系统及服务器 | |
CN114443864A (zh) | 跨模态数据的匹配方法、装置及计算机程序产品 | |
CN114463746A (zh) | 目标识别模型训练以及细胞识别方法、装置及电子设备 | |
CN112163090A (zh) | 一种法律裁判文书的案由分类方法和终端 | |
CN110610213A (zh) | 一种邮件分类方法、装置、设备及计算机可读存储介质 | |
CN117493645B (zh) | 一种基于大数据的电子档案推荐系统 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19938368 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19938368 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 16.08.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19938368 Country of ref document: EP Kind code of ref document: A1 |