WO2021068682A1 - Method and apparatus for intelligently filtering table text, and computer-readable storage medium - Google Patents

Method and apparatus for intelligently filtering table text, and computer-readable storage medium Download PDF

Info

Publication number
WO2021068682A1
WO2021068682A1 PCT/CN2020/112334 CN2020112334W WO2021068682A1 WO 2021068682 A1 WO2021068682 A1 WO 2021068682A1 CN 2020112334 W CN2020112334 W CN 2020112334W WO 2021068682 A1 WO2021068682 A1 WO 2021068682A1
Authority
WO
WIPO (PCT)
Prior art keywords
image
text
feature
image set
key
Prior art date
Application number
PCT/CN2020/112334
Other languages
French (fr)
Chinese (zh)
Inventor
石明川
李路路
Original Assignee
平安科技(深圳)有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 平安科技(深圳)有限公司 filed Critical 平安科技(深圳)有限公司
Publication of WO2021068682A1 publication Critical patent/WO2021068682A1/en

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/414Extracting the geometrical structure, e.g. layout tree; Block segmentation, e.g. bounding boxes for graphics or text
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/11Region-based segmentation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • G06T7/136Segmentation; Edge detection involving thresholding
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/413Classification of content, e.g. text, photographs or tables
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]

Definitions

  • This application relates to the field of artificial intelligence technology, and in particular to a method, device and computer-readable storage medium for intelligently filtering table text.
  • An intelligent filtering method for table text provided in this application includes:
  • the present application also provides an electronic device that includes a memory and a processor, the memory stores a table text filtering program that can be run on the processor, and the table text filtering program is executed by the processor
  • the memory stores a table text filtering program that can be run on the processor
  • the table text filtering program is executed by the processor
  • the present application also provides a computer-readable storage medium having a table text filtering program stored on the computer-readable storage medium, and the table text filtering program can be executed by one or more processors to realize the following table
  • the steps of the text intelligent filtering method obtaining a document-based table image set, and preprocessing the table image set to obtain a standard table image set;
  • This application also provides an intelligent filtering device for form text, which includes:
  • the image preprocessing module is used to obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set;
  • An enhancement processing module configured to perform enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set
  • the feature extraction module is used to perform feature image extraction on the table key image area set to obtain a feature table image set
  • the filtering module is configured to use a pre-built table text filtering model to perform text position detection on the feature table image set, and filter the text if the position of the text in the feature table image of the feature table image set is detected After that, the characteristic table image is saved, and if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  • FIG. 1 is a schematic flowchart of a method for intelligently filtering table text provided by an embodiment of this application;
  • FIG. 2 is a schematic diagram of the internal structure of an electronic device provided by an embodiment of the application.
  • FIG. 3 is a schematic diagram of modules of a table text intelligent filtering device provided by an embodiment of the application.
  • This application provides an intelligent filtering method for table text.
  • FIG. 1 it is a schematic flowchart of a method for intelligently filtering table text provided by an embodiment of this application.
  • the method can be executed by a device, and the device can be implemented by software and/or hardware.
  • the intelligent filtering method for table text includes:
  • the document includes a word document.
  • the word document will contain a large amount of text content in the form of a table.
  • the text content in the form of a table is scanned to obtain a table image, according to the table The image combination forms a table image set.
  • this application obtains the word document in the following two ways: method one, downloading it from major search engines using keyword terms; method two, downloading it from major professional academic websites, for example, China HowNet.
  • the preprocessing operation includes: performing image gray-scale processing on the table image set according to various scale methods to obtain a gray-scale table image set, and use the contrast stretching method to perform the gray-scale table image set. Contrast enhancement is performed on the gray-scale table image set, and the standard table image set is obtained by performing an image thresholding operation on the gray-scale table image set after the contrast enhancement.
  • the preprocessing operation is as follows:
  • the image gray-scale processing is to convert a color image into a gray-scale image.
  • the brightness information of the grayscale image can fully express the overall and local characteristics of the image, and the grayscale processing of the image can greatly reduce the amount of calculation for subsequent work.
  • the table image set in each scale method is converted into a gray scale table image set, and the steps of implementing each scale method are: converting the R, G, and B components of the pixels in the table image set into YUV
  • the Y component of the color space is the brightness value, and the calculation method of the Y component is shown in the following formula:
  • R, G, and B are the R, G, and B values of the image pixel in the RGB color mode, respectively.
  • the contrast refers to the contrast between the maximum value and the minimum value of the brightness in the imaging system, where low contrast makes image processing more difficult.
  • a contrast stretching method is adopted, which uses a method of increasing the dynamic range of gray levels to achieve the purpose of image contrast enhancement.
  • the contrast stretching is also called gray-scale stretching, which is a commonly used gray-scale transformation method at present.
  • the present application performs gray scale stretching on a specific area according to the piecewise linear transformation function in the contrast stretching method, so as to further improve the contrast of the output image.
  • contrast stretching it essentially realizes gray value conversion.
  • This application implements the gray value transformation through linear stretching.
  • the linear stretching refers to a pixel-level operation with a linear relationship between the input and output gray values.
  • the gray conversion formula is as follows:
  • the image thresholding process is an efficient algorithm for binarizing the contrast-enhanced gray-scale table image set through the OTSU algorithm.
  • the preferred embodiment of the present application presets the gray level t to be the segmentation threshold of the foreground and background of the gray image, and assumes that the proportion of the number of front spots in the image is w 0 , the average gray level is u 0 , and the proportion of background points in the image is w 1 , The average gray level is u 1 , then the total average gray level of the gray image is:
  • the gray level t at this time is the optimal threshold, and the gray level value of the gray level image after the contrast enhancement is greater than the gray level t It is set to 255, and the gray value smaller than the gray t is set to 0, thereby obtaining the standard table image set.
  • the image enhancement algorithm includes a threshold segmentation method and a Retinex algorithm.
  • this application uses a threshold segmentation method to segment the foreground text and background pattern in the standard table image set.
  • the core idea of the threshold segmentation method is to traverse each pixel in the image by setting a threshold T. When the gray value of the pixel is greater than T, it is considered as foreground text, otherwise it is considered as background pattern.
  • this application adopts the region growth method to perform segmentation processing. Wherein, the special text includes characters, symbols, etc.
  • the core idea of the area growth method is to aggregate pixels or sub-regions into a larger area according to pre-defined criteria, starting from a set of growth points (the growth point can be a single pixel or a small area), and will be related to the nature of the growth point. Similar adjacent pixels or regions are merged with the growth point to form a new growth point, and this process is repeated until no growth is possible.
  • the Retinex algorithm is used to calculate the key information image regions in the standard table image set after segmentation, to obtain the table key image regions, so as to combine to form the table key image region set, wherein the Retinex algorithm include:
  • S(x,y) represents the table key image area
  • R(x,y) represents the reflected light image
  • L(x,y) represents the brightness image
  • x represents the abscissa of the table key image area
  • y represents the table key The ordinate of the image area.
  • the core idea of the Retinex algorithm is: an image is composed of a brightness image and a reflection image, expressed as the product of a pixel and a corresponding pixel between the brightness image and the reflection image of the image, and image enhancement can be achieved by reducing the influence of the brightness image on the reflection image purpose.
  • feature image extraction is performed on the key image region set of the table through a residual block neural network.
  • the residual block neural network includes an input layer, a hidden layer and an output layer.
  • the present application inputs the table key image area set into the residual block neural network input layer, and uses the hidden layer of the residual block neural network to perform a convolution operation on the table key image area set, Obtain the feature map set of the table key image area set, and output the feature map set through the output layer of the residual block neural network, thereby obtaining the feature table image set.
  • the embodiment of the present application also includes adding a shortcut connection to the residual block neural network, the shortcut connection is a direct connection or a shortcut connection, that is, the F(x)+x function of the residual block neural network is substituted The original H(x) function to achieve fast connection.
  • the table text filtering model includes a text detection network.
  • the text position detection includes: generating a geometric diagram in the feature table image set, scaling the geometric diagram according to a preset ratio, and inputting the zoomed geometric diagram into the table text filtering model After training, the zoomed geometric figure loss L g is obtained ; the class-balanced cross-entropy is used to calculate the text loss L s in the zoomed geometric figure; the zoomed geometric figure loss and text loss are input to the preview
  • the loss function value is obtained from the loss function, and the text position detection is performed on the feature table image set according to the loss function value.
  • the loss function value is less than the preset threshold, the position of the text in the feature table image is detected, and the text is filtered and the feature table image is saved. If the loss function value is greater than or equal to the preset When the threshold is used, the position of the text in the characteristic table image is not detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  • the preset threshold in this application is 0.01.
  • the loss function includes:
  • L represents the loss function value
  • L s and L g represent text loss and geometric graph loss, respectively
  • ⁇ g represents the importance level value between the two losses.
  • inputting the zoomed geometric figure into the table text filtering model for training to obtain the zoomed geometric figure loss L g includes: inputting the zoomed geometric figure into In the input layer of the tabular text filtering model, feature merging is performed on the zoomed geometric map through the hidden layer of the tabular text filtering model to obtain a feature map, and the output layer of the tabular text filtering model compares all features.
  • the feature map performs frame regression, thereby outputting the loss L g of the geometric map.
  • the hidden layer includes a convolutional layer and a pooling layer.
  • the invention also provides an intelligent filtering device for table text.
  • FIG. 2 it is a schematic diagram of the internal structure of an electronic device provided by an embodiment of this application.
  • the electronic device 1 may be a PC (Personal Computer, personal computer), or a terminal device such as a smart phone, a tablet computer, or a portable computer, or a server.
  • the electronic device 1 at least includes a memory 11, a processor 12, a communication bus 13, and a network interface 14.
  • the memory 11 includes at least one type of readable storage medium.
  • the readable storage medium includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), magnetic memory, magnetic disk, optical disk, and the like.
  • the memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a hard disk of the electronic device 1.
  • the memory 11 may also be an external storage device of the electronic device 1, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SmartMediaCard, SMC), a Secure Digital (SD) card, and a flash memory. Card (FlashCard) etc.
  • the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device.
  • the memory 11 can be used not only to store application software and various data installed in the electronic device 1, such as the code of the table text filtering program 01, etc., but also to temporarily store data that has been output or will be output.
  • the processor 12 may be a central processing unit (CPU), controller, microcontroller, microprocessor, or other data processing chip, for running program codes or processing data stored in the memory 11, For example, execute the form text filtering program 01 and so on.
  • CPU central processing unit
  • controller microcontroller
  • microprocessor or other data processing chip
  • the communication bus 13 is used to realize the connection and communication between these components.
  • the network interface 14 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface), and is usually used to establish a communication connection between the device 1 and other electronic devices.
  • the device 1 may also include a user interface.
  • the user interface may include a display (Display) and an input unit such as a keyboard (Keyboard).
  • the optional user interface may also include a standard wired interface and a wireless interface.
  • the display may be an LED display, a liquid crystal display, a touch liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light emitting diode) touch device, etc.
  • the display can also be appropriately called a display screen or a display unit, which is used to display the information processed in the electronic device 1 and to display a visualized user interface.
  • Figure 2 only shows the electronic device 1 with the components 11-14 and the form text filtering program 01.
  • Figure 1 does not constitute a limitation on the electronic device 1, and may include comparisons. Fewer or more parts are shown, or some parts are combined, or different parts are arranged.
  • the table text filtering program 01 is stored in the memory 11; when the processor 12 executes the table text filtering program 01 stored in the memory 11, the following steps are implemented:
  • Step 1 Obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set.
  • the document includes a word document.
  • the word document will contain a large amount of text content in the form of a table.
  • the text content in the form of a table is scanned to obtain a table image, according to the table The image combination forms a table image set.
  • this application obtains the word document in the following two ways: method one, downloading it from major search engines using keyword terms; method two, downloading it from major professional academic websites, for example, China HowNet.
  • the preprocessing operation includes: performing image gray-scale processing on the table image set according to various scale methods to obtain a gray-scale table image set, and use the contrast stretching method to perform the gray-scale table image set. Contrast enhancement is performed on the gray-scale table image set, and the standard table image set is obtained by performing an image thresholding operation on the gray-scale table image set after the contrast enhancement.
  • the preprocessing operation is as follows:
  • the image gray-scale processing is to convert a color image into a gray-scale image.
  • the brightness information of the grayscale image can fully express the overall and local characteristics of the image, and the grayscale processing of the image can greatly reduce the amount of calculation for subsequent work.
  • the table image set in each scale method is converted into a gray scale table image set, and the steps of implementing each scale method are: converting the R, G, and B components of the pixels in the table image set into YUV
  • the Y component of the color space is the brightness value, and the calculation method of the Y component is shown in the following formula:
  • R, G, and B are the R, G, and B values of the image pixel in the RGB color mode, respectively.
  • the contrast refers to the contrast between the maximum value and the minimum value of the brightness in the imaging system, where low contrast makes image processing more difficult.
  • a contrast stretching method is adopted, which uses a method of increasing the dynamic range of gray levels to achieve the purpose of image contrast enhancement.
  • the contrast stretching is also called gray-scale stretching, which is a commonly used gray-scale transformation method at present.
  • the present application performs gray scale stretching on a specific area according to the piecewise linear transformation function in the contrast stretching method, so as to further improve the contrast of the output image.
  • contrast stretching it essentially realizes gray value conversion.
  • This application implements the gray value transformation through linear stretching.
  • the linear stretching refers to a pixel-level operation with a linear relationship between the input and output gray values.
  • the gray conversion formula is as follows:
  • the image thresholding process is an efficient algorithm for binarizing the contrast-enhanced gray-scale table image set through the OTSU algorithm.
  • the preferred embodiment of the present application presets the gray level t to be the segmentation threshold of the foreground and background of the gray image, and assumes that the proportion of the number of front spots in the image is w 0 , the average gray level is u 0 , and the proportion of background points in the image is w 1 , The average gray level is u 1 , then the total average gray level of the gray image is:
  • the gray level t at this time is the optimal threshold, and the gray level value of the gray level image after the contrast enhancement is greater than the gray level t It is set to 255, and the gray value smaller than the gray t is set to 0, thereby obtaining the standard table image set.
  • Step 2 Using an image enhancement algorithm to perform enhancement processing on the standard table image set to obtain a table key image area set.
  • the image enhancement algorithm includes a threshold segmentation method and a Retinex algorithm.
  • this application uses a threshold segmentation method to segment the foreground text and background pattern in the standard table image set.
  • the core idea of the threshold segmentation method is to traverse each pixel in the image by setting a threshold T. When the gray value of the pixel is greater than T, it is considered as foreground text, otherwise it is considered as background pattern.
  • this application adopts the region growth method to perform segmentation processing. Wherein, the special text includes characters, symbols, etc.
  • the core idea of the area growth method is to aggregate pixels or sub-regions into a larger area according to pre-defined criteria, starting from a set of growth points (the growth point can be a single pixel or a small area), and will be related to the nature of the growth point. Similar adjacent pixels or regions are merged with the growth point to form a new growth point, and this process is repeated until no growth is possible.
  • the Retinex algorithm is used to calculate the key information image regions in the standard table image set after segmentation, to obtain the table key image regions, so as to combine to form the table key image region set, wherein the Retinex algorithm include:
  • S(x,y) represents the table key image area
  • R(x,y) represents the reflected light image
  • L(x,y) represents the brightness image
  • x represents the abscissa of the table key image area
  • y represents the table key The ordinate of the image area.
  • the core idea of the Retinex algorithm is: an image is composed of a brightness image and a reflection image, expressed as the product of a pixel and a corresponding pixel between the brightness image and the reflection image of the image, and image enhancement can be achieved by reducing the influence of the brightness image on the reflection image purpose.
  • Step 3 Perform feature image extraction on the key image region set of the table to obtain a feature table image set.
  • feature image extraction is performed on the key image region set of the table through a residual block neural network.
  • the residual block neural network includes an input layer, a hidden layer and an output layer.
  • the present application inputs the table key image area set into the residual block neural network input layer, and uses the hidden layer of the residual block neural network to perform the convolution operation on the table key image area set, Obtain the feature map set of the table key image area set, and output the feature map set through the output layer of the residual block neural network, thereby obtaining the feature table image set.
  • the embodiment of the present application also includes adding a shortcut connection to the residual block neural network, the shortcut connection is a direct connection or a shortcut connection, that is, the F(x)+x function of the residual block neural network is substituted The original H(x) function to achieve fast connection.
  • Step 4 Use the pre-built table text filtering model to perform text position detection on the feature table image set. If the position of the text in the feature table image is detected, filter the text and save the feature table image. If there is no The position of the text in the characteristic table image is detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  • the table text filtering model includes a text detection network.
  • the text position detection includes: generating a geometric diagram in the feature table image set, scaling the geometric diagram according to a preset ratio, and inputting the zoomed geometric diagram into the table text filtering model After training, the zoomed geometric figure loss L g is obtained ; the class-balanced cross-entropy is used to calculate the text loss L s in the zoomed geometric figure; the zoomed geometric figure loss and text loss are input to the preview
  • the loss function value is obtained from the loss function, and the text position detection is performed on the feature table image set according to the loss function value.
  • the loss function value is less than the preset threshold, the position of the text in the feature table image is detected, and the text is filtered and the feature table image is saved. If the loss function value is greater than or equal to the preset When the threshold is used, the position of the text in the characteristic table image is not detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  • the preset threshold in this application is 0.01.
  • the loss function includes:
  • L represents the loss function value
  • L s and L g represent text loss and geometric graph loss, respectively
  • ⁇ g represents the importance level value between the two losses.
  • inputting the zoomed geometric figure into the table text filtering model for training to obtain the zoomed geometric figure loss L g includes: inputting the zoomed geometric figure into In the input layer of the tabular text filtering model, feature merging is performed on the zoomed geometric map through the hidden layer of the tabular text filtering model to obtain a feature map, and the output layer of the tabular text filtering model compares all features.
  • the feature map performs frame regression, thereby outputting the loss L g of the geometric map.
  • the hidden layer includes a convolutional layer and a pooling layer.
  • the table text filtering program may also be divided into one or more modules, and the one or more modules are stored in the memory 11 and run by one or more processors (in this embodiment, The processor 12) is executed to complete the application.
  • the form text intelligent filtering device can be divided into an image preprocessing module 10, an enhancement processing module 20, The feature extraction module 30 and the filtering module 40 exemplarily:
  • the image preprocessing module 10 is configured to obtain a document-based form image set, and perform a preprocessing operation on the form image set to obtain a standard form image set.
  • the enhancement processing module 20 is configured to perform enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set.
  • the feature extraction module 30 is configured to perform feature image extraction on the table key image area set to obtain a feature table image set.
  • the filtering module 40 is configured to: use a pre-built table text filtering model to perform text position detection on the characteristic table image set, and if the position of the text in the characteristic table image of the characteristic table image set is detected, the After the text is filtered, the characteristic table image is saved. If the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  • an embodiment of the present application also proposes a computer-readable storage medium having a table text filtering program stored on the computer-readable storage medium, and the table text filtering program can be executed by one or more processors to achieve the following operating:
  • the computer-readable storage medium may be non-volatile or volatile.

Landscapes

  • Engineering & Computer Science (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Multimedia (AREA)
  • Computer Graphics (AREA)
  • Geometry (AREA)
  • Character Input (AREA)
  • Image Processing (AREA)

Abstract

The present application relates to artificial intelligence technology, and disclosed therein is a method for intelligently filtering table text, comprising: acquiring a document-based table image set, and performing a preprocessing operation on the table image set to obtain a standard table image set; enhancing the standard table image set by using an image enhancement algorithm so as to obtain a table key image region set; performing feature image extraction on the table key image region set so as to obtain a feature table image set; performing text position detection on the feature table image set by using a pre-constructed table text filtering model; if the position of the text is detected, filtering the text and then storing a corresponding feature table image; and if the position of the text is not detected, directly saving a corresponding feature table image to thereby complete the text filtering of the table image set. Further proposed in the present application are an apparatus for intelligently filtering table text and a computer-readable storage medium. The present application achieves the intelligent filtering of table text.

Description

表格文本智能过滤方法、装置及计算机可读存储介质Table text intelligent filtering method, device and computer readable storage medium
本申请要求于2019年10月11日提交中国专利局、申请号为201910965807.1,发明名称为“表格文本智能过滤方法、装置及计算机可读存储介质”的中国专利申请的优先权,其全部内容通过引用结合在本申请中。This application claims the priority of a Chinese patent application filed with the Chinese Patent Office on October 11, 2019, the application number is 201910965807.1, and the invention title is "Intelligent Form Text Filtering Method, Device, and Computer-readable Storage Medium". The entire content of the Chinese patent application is approved. The reference is incorporated in this application.
技术领域Technical field
本申请涉及人工智能技术领域,尤其涉及一种表格文本智能过滤方法、装置及计算机可读存储介质。This application relates to the field of artificial intelligence technology, and in particular to a method, device and computer-readable storage medium for intelligently filtering table text.
背景技术Background technique
目前市场上存在各式各样的分类器,但大多数公司都是采用传统的如KNN,SVM,BP神经网络等分类算法。发明人意识到,这些传统的分类器通常在表格文本过滤任务中不够有效,分类准确率也一直都达不到很高的水平,尤其对于保险行业的票据表格文本过滤处理而言是个很大的问题。There are various classifiers on the market, but most companies use traditional classification algorithms such as KNN, SVM, and BP neural network. The inventor realized that these traditional classifiers are usually not effective in table text filtering tasks, and the classification accuracy rate has not reached a very high level, especially for the bill table text filtering processing in the insurance industry. problem.
发明内容Summary of the invention
本申请提供的一种表格文本智能过滤方法,包括:An intelligent filtering method for table text provided in this application includes:
获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
本申请还提供一种电子设备,该电子设备包括存储器和处理器,所述存储器中存储有可在所述处理器上运行的表格文本过滤程序,所述表格文本过滤程序被所述处理器执行时实现如下步骤:The present application also provides an electronic device that includes a memory and a processor, the memory stores a table text filtering program that can be run on the processor, and the table text filtering program is executed by the processor When implementing the following steps:
获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
本申请还提供一种计算机可读存储介质,所述计算机可读存储介质上存储有表格文本过滤程序,所述表格文本过滤程序可被一个或者多个处理器执行,以实现如下所述的表格文本智能过滤方法的步骤:获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;The present application also provides a computer-readable storage medium having a table text filtering program stored on the computer-readable storage medium, and the table text filtering program can be executed by one or more processors to realize the following table The steps of the text intelligent filtering method: obtaining a document-based table image set, and preprocessing the table image set to obtain a standard table image set;
利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
本申请还提供一种表格文本智能过滤装置,所述装置包括:This application also provides an intelligent filtering device for form text, which includes:
图像预处理模块,用于获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;The image preprocessing module is used to obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set;
增强处理模块,用于利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;An enhancement processing module, configured to perform enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
特征提取模块,用于对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;The feature extraction module is used to perform feature image extraction on the table key image area set to obtain a feature table image set;
过滤模块,用于利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。The filtering module is configured to use a pre-built table text filtering model to perform text position detection on the feature table image set, and filter the text if the position of the text in the feature table image of the feature table image set is detected After that, the characteristic table image is saved, and if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
附图说明Description of the drawings
图1为本申请一实施例提供的表格文本智能过滤方法的流程示意图;FIG. 1 is a schematic flowchart of a method for intelligently filtering table text provided by an embodiment of this application;
图2为本申请一实施例提供的电子设备的内部结构示意图;2 is a schematic diagram of the internal structure of an electronic device provided by an embodiment of the application;
图3为本申请一实施例提供的表格文本智能过滤装置的模块示意图。FIG. 3 is a schematic diagram of modules of a table text intelligent filtering device provided by an embodiment of the application.
本申请目的的实现、功能特点及优点将结合实施例,参照附图做进一步说明。The realization, functional characteristics, and advantages of the purpose of this application will be further described in conjunction with the embodiments and with reference to the accompanying drawings.
具体实施方式Detailed ways
应当理解,此处所描述的具体实施例仅仅用以解释本申请,并不用于限定本申请。It should be understood that the specific embodiments described here are only used to explain the application, and not used to limit the application.
本申请提供一种表格文本智能过滤方法。参照图1所示,为本申请一实施例提供的表格文本智能过滤方法的流程示意图。该方法可以由一个装置执行,该装置可以由软件和/或硬件实现。This application provides an intelligent filtering method for table text. Referring to FIG. 1, it is a schematic flowchart of a method for intelligently filtering table text provided by an embodiment of this application. The method can be executed by a device, and the device can be implemented by software and/or hardware.
在本实施例中,表格文本智能过滤方法包括:In this embodiment, the intelligent filtering method for table text includes:
S1、获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集。S1. Obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set.
本申请较佳实施例中,所述文档包括word文档。其中,在所述word文档中,会包含大量的以表格形式出现的文本内容,较佳地,本申请中通过对所述以表格形式出现的文本内容进行扫描,得到表格图像,根据所述表格图像组合形成表格图像集。In a preferred embodiment of the present application, the document includes a word document. Wherein, the word document will contain a large amount of text content in the form of a table. Preferably, in this application, the text content in the form of a table is scanned to obtain a table image, according to the table The image combination forms a table image set.
进一步地,本申请通过以下两种方式获取所述word文档:方式一、利用关键字词从各大搜索引擎中下载得到;方式二、通过从各大专业学术网站中进行下载得到,例如,中国知网。Further, this application obtains the word document in the following two ways: method one, downloading it from major search engines using keyword terms; method two, downloading it from major professional academic websites, for example, China HowNet.
较佳地,本申请较佳实施例中,所述预处理操作包括:根据各比例法对所述表格图像集进行图像灰度化处理后得到灰度表格图像集,利用对比度拉伸方式对所述灰度表格图像集进行对比度增强,将对比度增强后的所述灰度表格图像集进行图像阈值化操作后得到所述标准表格图像集。详细地,所述预处理操作如下所示:Preferably, in a preferred embodiment of the present application, the preprocessing operation includes: performing image gray-scale processing on the table image set according to various scale methods to obtain a gray-scale table image set, and use the contrast stretching method to perform the gray-scale table image set. Contrast enhancement is performed on the gray-scale table image set, and the standard table image set is obtained by performing an image thresholding operation on the gray-scale table image set after the contrast enhancement. In detail, the preprocessing operation is as follows:
图像灰度化处理:Image grayscale processing:
所述图像灰度化处理是将彩色图像转换为灰度图像。灰度图像的亮度信息完全能够表达图像的整体和局部的特征,并且对图像进行灰度化处理之后可以大大降低后续工作的计算量。The image gray-scale processing is to convert a color image into a gray-scale image. The brightness information of the grayscale image can fully express the overall and local characteristics of the image, and the grayscale processing of the image can greatly reduce the amount of calculation for subsequent work.
本申请较佳实施例通过各比例法所述表格图像集转换为灰度表格图像集,所述各比例 法实施步骤为:将所述表格图像集中像素点的R、G、B分量转换为YUV的颜色空间的Y分量,即亮度值,所述Y分量的计算方法如下式所示:In a preferred embodiment of the present application, the table image set in each scale method is converted into a gray scale table image set, and the steps of implementing each scale method are: converting the R, G, and B components of the pixels in the table image set into YUV The Y component of the color space is the brightness value, and the calculation method of the Y component is shown in the following formula:
Y=0.3R+0.59G+0.11BY=0.3R+0.59G+0.11B
其中R、G、B分别是RGB色彩模式中图像像素点的R、G、B值。Among them, R, G, and B are the R, G, and B values of the image pixel in the RGB color mode, respectively.
对比度增强:Contrast enhancement:
所述对比度指的是成像系统中亮度最大值与最小值之间的对比,其中,对比度低会使图像处理难度增大。本申请较佳实施例中采用的是对比度拉伸方法,利用提高灰度级动态范围的方式,达到图像对比度增强的目的。所述对比度拉伸也叫作灰度拉伸,是目前常用的灰度变换方式。The contrast refers to the contrast between the maximum value and the minimum value of the brightness in the imaging system, where low contrast makes image processing more difficult. In the preferred embodiment of the present application, a contrast stretching method is adopted, which uses a method of increasing the dynamic range of gray levels to achieve the purpose of image contrast enhancement. The contrast stretching is also called gray-scale stretching, which is a commonly used gray-scale transformation method at present.
进一步地,本申请根据所述对比度拉伸方法中的分段线性变换函数对特定区域进行灰度拉伸,进一步提高输出图像的对比度。当进行对比度拉伸时,本质上是实现灰度值变换。本申请通过线性拉伸实现灰度值变换,所述线性拉伸指的是输入与输出的灰度值之间为线性关系的像素级运算,灰度变换公式如下所示:Further, the present application performs gray scale stretching on a specific area according to the piecewise linear transformation function in the contrast stretching method, so as to further improve the contrast of the output image. When performing contrast stretching, it essentially realizes gray value conversion. This application implements the gray value transformation through linear stretching. The linear stretching refers to a pixel-level operation with a linear relationship between the input and output gray values. The gray conversion formula is as follows:
D b=f(D a)=a*D a+b D b =f(D a )=a*D a +b
其中a为线性斜率,b为在Y轴上的截距。当a>1时,此时输出的图像对比度相比原图像是增强的。当a<1时,此时输出的图像对比度相比原图像是削弱的,其中D a代表输入图像灰度值,D b代表输出图像灰度值。 Where a is the linear slope and b is the intercept on the Y axis. When a>1, the contrast of the output image at this time is enhanced compared to the original image. When a<1, the contrast of the output image is weaker than the original image at this time, where D a represents the gray value of the input image, and D b represents the gray value of the output image.
c.图像阈值化操作:c. Image thresholding operation:
所述图像阈值化处理通过OTSU算法将对比度增强后的所述灰度表格图像集进行二值化的高效算法。本申请较佳实施例预设灰度t为灰度图像的前景与背景的分割阈值,并假设前景点数占图像比例为w 0,平均灰度为u 0;背景点数占图像比例为w 1,平均灰度为u 1,则灰度图像的总平均灰度为: The image thresholding process is an efficient algorithm for binarizing the contrast-enhanced gray-scale table image set through the OTSU algorithm. The preferred embodiment of the present application presets the gray level t to be the segmentation threshold of the foreground and background of the gray image, and assumes that the proportion of the number of front spots in the image is w 0 , the average gray level is u 0 , and the proportion of background points in the image is w 1 , The average gray level is u 1 , then the total average gray level of the gray image is:
u=w 0*u 0+w 1*u 1u=w 0 *u 0 +w 1 *u 1 ,
灰度图像的前景和背景图象的方差为:The variance of the foreground and background image of the grayscale image is:
g=w 0*(u 0-u)*(u 0-u)+w 1*(u 1-u)*(u 1-u)=w 0*w 1*(u 0-u 1)*(u 0-u 1), g=w 0 *(u 0 -u)*(u 0 -u)+w 1 *(u 1 -u)*(u 1 -u)=w 0 *w 1 *(u 0 -u 1 )* (u 0 -u 1 ),
其中,当方差g最大时,则此时前景和背景差异最大,此时的灰度t为最佳阈值,并将对比度增强后的所述灰度图像中大于所述灰度t的灰度值设置为255,小于所述灰度t的灰度值设置为0,从而得到所述标准表格图像集。Among them, when the variance g is the largest, the difference between the foreground and the background is the largest at this time, the gray level t at this time is the optimal threshold, and the gray level value of the gray level image after the contrast enhancement is greater than the gray level t It is set to 255, and the gray value smaller than the gray t is set to 0, thereby obtaining the standard table image set.
S2、利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集。S2. Use an image enhancement algorithm to perform enhancement processing on the standard table image set to obtain a table key image area set.
本申请较佳实施例中,所述图像增强算法包括阈值分割法和Retinex算法。优先的,本申请通过阈值分割法对所述标准表格图像集中的前景文字和背景图案进行分割。所述阈值分割法的核心思想是通过设置一个阈值T,遍历图像中的每个像素点,当像素点的灰度值大于T时,认为是前景文字,否则认为是背景图案。进一步地,对于分割后的所述标准表格图像集中的特殊文字,本申请采用区域增长法进行分割处理。其中,所述特殊文字包含字符,符号等。所述区域增长法的核心思想是根据事先定义的准则将像素或者子区域聚合成更大的区域,从一组生长点开始(生长点可以是单个像素或者一个小区域),将与生长点性质相似的相邻像素或者区域与生长点合并,形成新的生长点,重复此过程直到不能生长为止。In a preferred embodiment of the present application, the image enhancement algorithm includes a threshold segmentation method and a Retinex algorithm. Preferably, this application uses a threshold segmentation method to segment the foreground text and background pattern in the standard table image set. The core idea of the threshold segmentation method is to traverse each pixel in the image by setting a threshold T. When the gray value of the pixel is greater than T, it is considered as foreground text, otherwise it is considered as background pattern. Further, for the special characters in the standard table image set after segmentation, this application adopts the region growth method to perform segmentation processing. Wherein, the special text includes characters, symbols, etc. The core idea of the area growth method is to aggregate pixels or sub-regions into a larger area according to pre-defined criteria, starting from a set of growth points (the growth point can be a single pixel or a small area), and will be related to the nature of the growth point. Similar adjacent pixels or regions are merged with the growth point to form a new growth point, and this process is repeated until no growth is possible.
较佳地,本申请中利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:Preferably, in this application, the Retinex algorithm is used to calculate the key information image regions in the standard table image set after segmentation, to obtain the table key image regions, so as to combine to form the table key image region set, wherein the Retinex algorithm include:
S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。所述Retinex算法的核心思想为:图像是由亮度图像和反射图像组成,表示为亮度图像和图像反射图像之间像素与对应像素的乘积,通过降低亮度图像对反射图像的影响可以达到图像增强的目的。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area. The core idea of the Retinex algorithm is: an image is composed of a brightness image and a reflection image, expressed as the product of a pixel and a corresponding pixel between the brightness image and the reflection image of the image, and image enhancement can be achieved by reducing the influence of the brightness image on the reflection image purpose.
S3、对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集。S3. Perform feature image extraction on the table key image area set to obtain a feature table image set.
本申请较佳实施例通过残差块神经网络对所述表格关键图像区域集进行特征图像提取。其中,所述残差块神经网络包括输入层、隐藏层以及输出层。较佳地,本申请通过将所述表格关键图像区域集输入至残差块神经网络输入层中,利用所述残差块神经网络的隐藏层对所述表格关键图像区域集进行卷积操作,得到表格关键图像区域集的特征图谱集,并通过所述残差块神经网络的输出层输出所述特征图谱集,从而得到所述特征表格图像集。In a preferred embodiment of the present application, feature image extraction is performed on the key image region set of the table through a residual block neural network. Wherein, the residual block neural network includes an input layer, a hidden layer and an output layer. Preferably, the present application inputs the table key image area set into the residual block neural network input layer, and uses the hidden layer of the residual block neural network to perform a convolution operation on the table key image area set, Obtain the feature map set of the table key image area set, and output the feature map set through the output layer of the residual block neural network, thereby obtaining the feature table image set.
进一步地,本申请实施例中还包括将shortcut连接加入残差块神经网络中,所述shortcut连接即直连或捷径连接,即以所述残差块神经网络的F(x)+x函数替代原本的H(x)函数,从而达到快速连接。Further, the embodiment of the present application also includes adding a shortcut connection to the residual block neural network, the shortcut connection is a direct connection or a shortcut connection, that is, the F(x)+x function of the residual block neural network is substituted The original H(x) function to achieve fast connection.
S4、利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出特征表格图像中文本的位置,将所述文本进行过滤后保存所述特征表格图像,若没有检测出特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。S4. Use the pre-built table text filtering model to perform text position detection on the feature table image set, if the position of the text in the feature table image is detected, filter the text and save the feature table image, if not detected The position of the text in the characteristic table image is extracted, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
本申请较佳实施例中,所述表格文本过滤模型包括文本检测网络。所述文本位置检测包括:在所述特征表格图像集中生成一个几何图,并将所述几何图按照预设的比例进行缩放,将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g;利用类平衡交叉熵计算缩放后的所述几何图中的文本损失L s;将缩放后的所述几何图损失和文本损失输入至预设的损失函数中得到损失函数值,根据所述损失函数值对所述特征表格图像集进行文本位置检测。若所述损失函数值小于预设的阈值时,检测出特征表格图像中文本的位置,并将所述文本进行过滤后保存所述特征表格图像,若所述损失函数值大于或等于预设的阈值时,没有检测出特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。 In a preferred embodiment of the present application, the table text filtering model includes a text detection network. The text position detection includes: generating a geometric diagram in the feature table image set, scaling the geometric diagram according to a preset ratio, and inputting the zoomed geometric diagram into the table text filtering model After training, the zoomed geometric figure loss L g is obtained ; the class-balanced cross-entropy is used to calculate the text loss L s in the zoomed geometric figure; the zoomed geometric figure loss and text loss are input to the preview The loss function value is obtained from the loss function, and the text position detection is performed on the feature table image set according to the loss function value. If the loss function value is less than the preset threshold, the position of the text in the feature table image is detected, and the text is filtered and the feature table image is saved. If the loss function value is greater than or equal to the preset When the threshold is used, the position of the text in the characteristic table image is not detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
优选地,本申请中所述预设的阈值为0.01。其中,所述损失函数包括:Preferably, the preset threshold in this application is 0.01. Wherein, the loss function includes:
L=L sgL g L=L sg L g
其中,L表示损失函数值,L s和L g分别表示文本损失和几何图损失,λ g表示两个损失之间的重要等级值。 Among them, L represents the loss function value, L s and L g represent text loss and geometric graph loss, respectively, and λ g represents the importance level value between the two losses.
进一步,本申请中所述将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g包括:将缩放后的所述几何图输入到所述表格文本过滤模型的输入层中,通过所述表格文本过滤模型的隐藏层对缩放后的所述几何图进行特征合并,得到特征图,并通过所述表格文本过滤模型的输出层对所述特征图进行边框回归,从而输出所述几何图的损失L g。其中,所述隐藏层包含卷积层和池化层。 Further, in this application, inputting the zoomed geometric figure into the table text filtering model for training to obtain the zoomed geometric figure loss L g includes: inputting the zoomed geometric figure into In the input layer of the tabular text filtering model, feature merging is performed on the zoomed geometric map through the hidden layer of the tabular text filtering model to obtain a feature map, and the output layer of the tabular text filtering model compares all features. The feature map performs frame regression, thereby outputting the loss L g of the geometric map. Wherein, the hidden layer includes a convolutional layer and a pooling layer.
发明还提供一种表格文本智能过滤装置。参照图2所示,为本申请一实施例提供的电子设备的内部结构示意图。The invention also provides an intelligent filtering device for table text. Referring to FIG. 2, it is a schematic diagram of the internal structure of an electronic device provided by an embodiment of this application.
在本实施例中,所述电子设备1可以是PC(PersonalComputer,个人电脑),或者是智能手机、平板电脑、便携计算机等终端设备,也可以是一种服务器等。该电子设备1至少包括存储器11、处理器12,通信总线13,以及网络接口14。In this embodiment, the electronic device 1 may be a PC (Personal Computer, personal computer), or a terminal device such as a smart phone, a tablet computer, or a portable computer, or a server. The electronic device 1 at least includes a memory 11, a processor 12, a communication bus 13, and a network interface 14.
其中,存储器11至少包括一种类型的可读存储介质,所述可读存储介质包括闪存、硬 盘、多媒体卡、卡型存储器(例如,SD或DX存储器等)、磁性存储器、磁盘、光盘等。存储器11在一些实施例中可以是电子设备1的内部存储单元,例如该电子设备1的硬盘。存储器11在另一些实施例中也可以是电子设备1的外部存储设备,例如电子设备1上配备的插接式硬盘,智能存储卡(SmartMediaCard,SMC),安全数字(SecureDigital,SD)卡,闪存卡(FlashCard)等。进一步地,存储器11还可以既包括电子设备1的内部存储单元也包括外部存储设备。存储器11不仅可以用于存储安装于电子设备1的应用软件及各类数据,例如表格文本过滤程序01的代码等,还可以用于暂时地存储已经输出或者将要输出的数据。The memory 11 includes at least one type of readable storage medium. The readable storage medium includes flash memory, hard disk, multimedia card, card-type memory (for example, SD or DX memory, etc.), magnetic memory, magnetic disk, optical disk, and the like. The memory 11 may be an internal storage unit of the electronic device 1 in some embodiments, such as a hard disk of the electronic device 1. In other embodiments, the memory 11 may also be an external storage device of the electronic device 1, such as a plug-in hard disk equipped on the electronic device 1, a smart memory card (SmartMediaCard, SMC), a Secure Digital (SD) card, and a flash memory. Card (FlashCard) etc. Further, the memory 11 may also include both an internal storage unit of the electronic device 1 and an external storage device. The memory 11 can be used not only to store application software and various data installed in the electronic device 1, such as the code of the table text filtering program 01, etc., but also to temporarily store data that has been output or will be output.
处理器12在一些实施例中可以是一中央处理器(CentralProcessingUnit,CPU)、控制器、微控制器、微处理器或其他数据处理芯片,用于运行存储器11中存储的程序代码或处理数据,例如执行表格文本过滤程序01等。In some embodiments, the processor 12 may be a central processing unit (CPU), controller, microcontroller, microprocessor, or other data processing chip, for running program codes or processing data stored in the memory 11, For example, execute the form text filtering program 01 and so on.
通信总线13用于实现这些组件之间的连接通信。The communication bus 13 is used to realize the connection and communication between these components.
网络接口14可选的可以包括标准的有线接口、无线接口(如WI-FI接口),通常用于在该装置1与其他电子设备之间建立通信连接。The network interface 14 may optionally include a standard wired interface and a wireless interface (such as a WI-FI interface), and is usually used to establish a communication connection between the device 1 and other electronic devices.
可选地,该装置1还可以包括用户接口,用户接口可以包括显示器(Display)、输入单元比如键盘(Keyboard),可选的用户接口还可以包括标准的有线接口、无线接口。可选地,在一些实施例中,显示器可以是LED显示器、液晶显示器、触控式液晶显示器以及OLED(OrganicLight-EmittingDiode,有机发光二极管)触摸器等。其中,显示器也可以适当的称为显示屏或显示单元,用于显示在电子设备1中处理的信息以及用于显示可视化的用户界面。Optionally, the device 1 may also include a user interface. The user interface may include a display (Display) and an input unit such as a keyboard (Keyboard). The optional user interface may also include a standard wired interface and a wireless interface. Optionally, in some embodiments, the display may be an LED display, a liquid crystal display, a touch liquid crystal display, an OLED (Organic Light-Emitting Diode, organic light emitting diode) touch device, etc. Among them, the display can also be appropriately called a display screen or a display unit, which is used to display the information processed in the electronic device 1 and to display a visualized user interface.
图2仅示出了具有组件11-14以及表格文本过滤程序01的电子设备1,本领域技术人员可以理解的是,图1示出的结构并不构成对电子设备1的限定,可以包括比图示更少或者更多的部件,或者组合某些部件,或者不同的部件布置。Figure 2 only shows the electronic device 1 with the components 11-14 and the form text filtering program 01. Those skilled in the art can understand that the structure shown in Figure 1 does not constitute a limitation on the electronic device 1, and may include comparisons. Fewer or more parts are shown, or some parts are combined, or different parts are arranged.
在图2所示的装置1实施例中,存储器11中存储有表格文本过滤程序01;处理器12执行存储器11中存储的表格文本过滤程序01时实现如下步骤:In the embodiment of the device 1 shown in FIG. 2, the table text filtering program 01 is stored in the memory 11; when the processor 12 executes the table text filtering program 01 stored in the memory 11, the following steps are implemented:
步骤一、获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集。Step 1: Obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set.
本申请较佳实施例中,所述文档包括word文档。其中,在所述word文档中,会包含大量的以表格形式出现的文本内容,较佳地,本申请中通过对所述以表格形式出现的文本内容进行扫描,得到表格图像,根据所述表格图像组合形成表格图像集。In a preferred embodiment of the present application, the document includes a word document. Wherein, the word document will contain a large amount of text content in the form of a table. Preferably, in this application, the text content in the form of a table is scanned to obtain a table image, according to the table The image combination forms a table image set.
进一步地,本申请通过以下两种方式获取所述word文档:方式一、利用关键字词从各大搜索引擎中下载得到;方式二、通过从各大专业学术网站中进行下载得到,例如,中国知网。Further, this application obtains the word document in the following two ways: method one, downloading it from major search engines using keyword terms; method two, downloading it from major professional academic websites, for example, China HowNet.
较佳地,本申请较佳实施例中,所述预处理操作包括:根据各比例法对所述表格图像集进行图像灰度化处理后得到灰度表格图像集,利用对比度拉伸方式对所述灰度表格图像集进行对比度增强,将对比度增强后的所述灰度表格图像集进行图像阈值化操作后得到所述标准表格图像集。详细地,所述预处理操作如下所示:Preferably, in a preferred embodiment of the present application, the preprocessing operation includes: performing image gray-scale processing on the table image set according to various scale methods to obtain a gray-scale table image set, and use the contrast stretching method to perform the gray-scale table image set. Contrast enhancement is performed on the gray-scale table image set, and the standard table image set is obtained by performing an image thresholding operation on the gray-scale table image set after the contrast enhancement. In detail, the preprocessing operation is as follows:
图像灰度化处理:Image grayscale processing:
所述图像灰度化处理是将彩色图像转换为灰度图像。灰度图像的亮度信息完全能够表达图像的整体和局部的特征,并且对图像进行灰度化处理之后可以大大降低后续工作的计算量。The image gray-scale processing is to convert a color image into a gray-scale image. The brightness information of the grayscale image can fully express the overall and local characteristics of the image, and the grayscale processing of the image can greatly reduce the amount of calculation for subsequent work.
本申请较佳实施例通过各比例法所述表格图像集转换为灰度表格图像集,所述各比例法实施步骤为:将所述表格图像集中像素点的R、G、B分量转换为YUV的颜色空间的Y分量,即亮度值,所述Y分量的计算方法如下式所示:In a preferred embodiment of the present application, the table image set in each scale method is converted into a gray scale table image set, and the steps of implementing each scale method are: converting the R, G, and B components of the pixels in the table image set into YUV The Y component of the color space is the brightness value, and the calculation method of the Y component is shown in the following formula:
Y=0.3R+0.59G+0.11BY=0.3R+0.59G+0.11B
其中R、G、B分别是RGB色彩模式中图像像素点的R、G、B值。Among them, R, G, and B are the R, G, and B values of the image pixel in the RGB color mode, respectively.
对比度增强:Contrast enhancement:
所述对比度指的是成像系统中亮度最大值与最小值之间的对比,其中,对比度低会使图像处理难度增大。本申请较佳实施例中采用的是对比度拉伸方法,利用提高灰度级动态范围的方式,达到图像对比度增强的目的。所述对比度拉伸也叫作灰度拉伸,是目前常用的灰度变换方式。The contrast refers to the contrast between the maximum value and the minimum value of the brightness in the imaging system, where low contrast makes image processing more difficult. In the preferred embodiment of the present application, a contrast stretching method is adopted, which uses a method of increasing the dynamic range of gray levels to achieve the purpose of image contrast enhancement. The contrast stretching is also called gray-scale stretching, which is a commonly used gray-scale transformation method at present.
进一步地,本申请根据所述对比度拉伸方法中的分段线性变换函数对特定区域进行灰度拉伸,进一步提高输出图像的对比度。当进行对比度拉伸时,本质上是实现灰度值变换。本申请通过线性拉伸实现灰度值变换,所述线性拉伸指的是输入与输出的灰度值之间为线性关系的像素级运算,灰度变换公式如下所示:Further, the present application performs gray scale stretching on a specific area according to the piecewise linear transformation function in the contrast stretching method, so as to further improve the contrast of the output image. When performing contrast stretching, it essentially realizes gray value conversion. This application implements the gray value transformation through linear stretching. The linear stretching refers to a pixel-level operation with a linear relationship between the input and output gray values. The gray conversion formula is as follows:
D b=f(D a)=a*D a+b D b =f(D a )=a*D a +b
其中a为线性斜率,b为在Y轴上的截距。当a>1时,此时输出的图像对比度相比原图像是增强的。当a<1时,此时输出的图像对比度相比原图像是削弱的,其中D a代表输入图像灰度值,D b代表输出图像灰度值。 Where a is the linear slope and b is the intercept on the Y axis. When a>1, the contrast of the output image at this time is enhanced compared to the original image. When a<1, the contrast of the output image is weaker than the original image at this time, where D a represents the gray value of the input image, and D b represents the gray value of the output image.
c.图像阈值化操作:c. Image thresholding operation:
所述图像阈值化处理通过OTSU算法将对比度增强后的所述灰度表格图像集进行二值化的高效算法。本申请较佳实施例预设灰度t为灰度图像的前景与背景的分割阈值,并假设前景点数占图像比例为w 0,平均灰度为u 0;背景点数占图像比例为w 1,平均灰度为u 1,则灰度图像的总平均灰度为: The image thresholding process is an efficient algorithm for binarizing the contrast-enhanced gray-scale table image set through the OTSU algorithm. The preferred embodiment of the present application presets the gray level t to be the segmentation threshold of the foreground and background of the gray image, and assumes that the proportion of the number of front spots in the image is w 0 , the average gray level is u 0 , and the proportion of background points in the image is w 1 , The average gray level is u 1 , then the total average gray level of the gray image is:
u=w 0*u 0+w 1*u 1u=w 0 *u 0 +w 1 *u 1 ,
灰度图像的前景和背景图象的方差为:The variance of the foreground and background image of the grayscale image is:
g=w 0*(u 0-u)*(u 0-u)+w 1*(u 1-u)*(u 1-u)=w 0*w 1*(u 0-u 1)*(u 0-u 1), g=w 0 *(u 0 -u)*(u 0 -u)+w 1 *(u 1 -u)*(u 1 -u)=w 0 *w 1 *(u 0 -u 1 )* (u 0 -u 1 ),
其中,当方差g最大时,则此时前景和背景差异最大,此时的灰度t为最佳阈值,并将对比度增强后的所述灰度图像中大于所述灰度t的灰度值设置为255,小于所述灰度t的灰度值设置为0,从而得到所述标准表格图像集。Among them, when the variance g is the largest, the difference between the foreground and the background is the largest at this time, the gray level t at this time is the optimal threshold, and the gray level value of the gray level image after the contrast enhancement is greater than the gray level t It is set to 255, and the gray value smaller than the gray t is set to 0, thereby obtaining the standard table image set.
步骤二、利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集。Step 2: Using an image enhancement algorithm to perform enhancement processing on the standard table image set to obtain a table key image area set.
本申请较佳实施例中,所述图像增强算法包括阈值分割法和Retinex算法。优先的,本申请通过阈值分割法对所述标准表格图像集中的前景文字和背景图案进行分割。所述阈值分割法的核心思想是通过设置一个阈值T,遍历图像中的每个像素点,当像素点的灰度值大于T时,认为是前景文字,否则认为是背景图案。进一步地,对于分割后的所述标准表格图像集中的特殊文字,本申请采用区域增长法进行分割处理。其中,所述特殊文字包含字符,符号等。所述区域增长法的核心思想是根据事先定义的准则将像素或者子区域聚合成更大的区域,从一组生长点开始(生长点可以是单个像素或者一个小区域),将与生长点性质相似的相邻像素或者区域与生长点合并,形成新的生长点,重复此过程直到不能生长为止。In a preferred embodiment of the present application, the image enhancement algorithm includes a threshold segmentation method and a Retinex algorithm. Preferably, this application uses a threshold segmentation method to segment the foreground text and background pattern in the standard table image set. The core idea of the threshold segmentation method is to traverse each pixel in the image by setting a threshold T. When the gray value of the pixel is greater than T, it is considered as foreground text, otherwise it is considered as background pattern. Further, for the special characters in the standard table image set after segmentation, this application adopts the region growth method to perform segmentation processing. Wherein, the special text includes characters, symbols, etc. The core idea of the area growth method is to aggregate pixels or sub-regions into a larger area according to pre-defined criteria, starting from a set of growth points (the growth point can be a single pixel or a small area), and will be related to the nature of the growth point. Similar adjacent pixels or regions are merged with the growth point to form a new growth point, and this process is repeated until no growth is possible.
较佳地,本申请中利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:Preferably, in this application, the Retinex algorithm is used to calculate the key information image regions in the standard table image set after segmentation, to obtain the table key image regions, so as to combine to form the table key image region set, wherein the Retinex algorithm include:
S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。所述Retinex算法的核心思想为:图像是由亮度图像和反射图像组成,表示为亮度图像和图像反射图像 之间像素与对应像素的乘积,通过降低亮度图像对反射图像的影响可以达到图像增强的目的。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area. The core idea of the Retinex algorithm is: an image is composed of a brightness image and a reflection image, expressed as the product of a pixel and a corresponding pixel between the brightness image and the reflection image of the image, and image enhancement can be achieved by reducing the influence of the brightness image on the reflection image purpose.
步骤三、对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集。Step 3: Perform feature image extraction on the key image region set of the table to obtain a feature table image set.
本申请较佳实施例通过残差块神经网络对所述表格关键图像区域集进行特征图像提取。其中,所述残差块神经网络包括输入层、隐藏层以及输出层。较佳地,本申请通过将所述表格关键图像区域集输入至残差块神经网络输入层中,利用所述残差块神经网络的隐藏层对所述表格关键图像区域集进行卷积操作,得到表格关键图像区域集的特征图谱集,并通过所述残差块神经网络的输出层输出所述特征图谱集,从而得到所述特征表格图像集。In a preferred embodiment of the present application, feature image extraction is performed on the key image region set of the table through a residual block neural network. Wherein, the residual block neural network includes an input layer, a hidden layer and an output layer. Preferably, the present application inputs the table key image area set into the residual block neural network input layer, and uses the hidden layer of the residual block neural network to perform the convolution operation on the table key image area set, Obtain the feature map set of the table key image area set, and output the feature map set through the output layer of the residual block neural network, thereby obtaining the feature table image set.
进一步地,本申请实施例中还包括将shortcut连接加入残差块神经网络中,所述shortcut连接即直连或捷径连接,即以所述残差块神经网络的F(x)+x函数替代原本的H(x)函数,从而达到快速连接。Further, the embodiment of the present application also includes adding a shortcut connection to the residual block neural network, the shortcut connection is a direct connection or a shortcut connection, that is, the F(x)+x function of the residual block neural network is substituted The original H(x) function to achieve fast connection.
步骤四、利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出特征表格图像中文本的位置,将所述文本进行过滤后保存所述特征表格图像,若没有检测出特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Step 4. Use the pre-built table text filtering model to perform text position detection on the feature table image set. If the position of the text in the feature table image is detected, filter the text and save the feature table image. If there is no The position of the text in the characteristic table image is detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
本申请较佳实施例中,所述表格文本过滤模型包括文本检测网络。所述文本位置检测包括:在所述特征表格图像集中生成一个几何图,并将所述几何图按照预设的比例进行缩放,将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g;利用类平衡交叉熵计算缩放后的所述几何图中的文本损失L s;将缩放后的所述几何图损失和文本损失输入至预设的损失函数中得到损失函数值,根据所述损失函数值对所述特征表格图像集进行文本位置检测。若所述损失函数值小于预设的阈值时,检测出特征表格图像中文本的位置,并将所述文本进行过滤后保存所述特征表格图像,若所述损失函数值大于或等于预设的阈值时,没有检测出特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。 In a preferred embodiment of the present application, the table text filtering model includes a text detection network. The text position detection includes: generating a geometric diagram in the feature table image set, scaling the geometric diagram according to a preset ratio, and inputting the zoomed geometric diagram into the table text filtering model After training, the zoomed geometric figure loss L g is obtained ; the class-balanced cross-entropy is used to calculate the text loss L s in the zoomed geometric figure; the zoomed geometric figure loss and text loss are input to the preview The loss function value is obtained from the loss function, and the text position detection is performed on the feature table image set according to the loss function value. If the loss function value is less than the preset threshold, the position of the text in the feature table image is detected, and the text is filtered and the feature table image is saved. If the loss function value is greater than or equal to the preset When the threshold is used, the position of the text in the characteristic table image is not detected, and the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
优选地,本申请中所述预设的阈值为0.01。其中,所述损失函数包括:Preferably, the preset threshold in this application is 0.01. Wherein, the loss function includes:
L=L sgL g L=L sg L g
其中,L表示损失函数值,L s和L g分别表示文本损失和几何图损失,λ g表示两个损失之间的重要等级值。 Among them, L represents the loss function value, L s and L g represent text loss and geometric graph loss, respectively, and λ g represents the importance level value between the two losses.
进一步,本申请中所述将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g包括:将缩放后的所述几何图输入到所述表格文本过滤模型的输入层中,通过所述表格文本过滤模型的隐藏层对缩放后的所述几何图进行特征合并,得到特征图,并通过所述表格文本过滤模型的输出层对所述特征图进行边框回归,从而输出所述几何图的损失L g。其中,所述隐藏层包含卷积层和池化层。 Further, in this application, inputting the zoomed geometric figure into the table text filtering model for training to obtain the zoomed geometric figure loss L g includes: inputting the zoomed geometric figure into In the input layer of the tabular text filtering model, feature merging is performed on the zoomed geometric map through the hidden layer of the tabular text filtering model to obtain a feature map, and the output layer of the tabular text filtering model compares all features. The feature map performs frame regression, thereby outputting the loss L g of the geometric map. Wherein, the hidden layer includes a convolutional layer and a pooling layer.
可选地,在其他实施例中,表格文本过滤程序还可以被分割为一个或者多个模块,一个或者多个模块被存储于存储器11中,并由一个或多个处理器(本实施例为处理器12)所执行以完成本申请。Optionally, in other embodiments, the table text filtering program may also be divided into one or more modules, and the one or more modules are stored in the memory 11 and run by one or more processors (in this embodiment, The processor 12) is executed to complete the application.
参照图3所示,为本申请表格文本智能过滤装置一实施例中的程序模块示意图,该实施例中,所述表格文本智能过滤装置可以被分割为图像预处理模块10、增强处理模块20、特征提取模块30以及过滤模块40,示例性地:3, which is a schematic diagram of program modules in an embodiment of the form text intelligent filtering device of this application. In this embodiment, the form text intelligent filtering device can be divided into an image preprocessing module 10, an enhancement processing module 20, The feature extraction module 30 and the filtering module 40 exemplarily:
所述图像预处理模块10用于:获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集。The image preprocessing module 10 is configured to obtain a document-based form image set, and perform a preprocessing operation on the form image set to obtain a standard form image set.
所述增强处理模块20用于:利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集。The enhancement processing module 20 is configured to perform enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set.
所述特征提取模块30用于:对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集。The feature extraction module 30 is configured to perform feature image extraction on the table key image area set to obtain a feature table image set.
所述过滤模块40用于:利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。The filtering module 40 is configured to: use a pre-built table text filtering model to perform text position detection on the characteristic table image set, and if the position of the text in the characteristic table image of the characteristic table image set is detected, the After the text is filtered, the characteristic table image is saved. If the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
上述图像预处理模块10、增强处理模块20、特征提取模块30以及过滤模块40等程序模块被执行时所实现的功能或操作步骤与上述实施例大体相同,在此不再赘述。The functions or operation steps implemented by the program modules such as the image preprocessing module 10, the enhancement processing module 20, the feature extraction module 30, and the filtering module 40 when executed are substantially the same as those in the foregoing embodiment, and will not be repeated here.
此外,本申请实施例还提出一种计算机可读存储介质,所述计算机可读存储介质上存储有表格文本过滤程序,所述表格文本过滤程序可被一个或多个处理器执行,以实现如下操作:In addition, an embodiment of the present application also proposes a computer-readable storage medium having a table text filtering program stored on the computer-readable storage medium, and the table text filtering program can be executed by one or more processors to achieve the following operating:
获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
所述计算机可读存储介质可以是非易失性,也可以是易失性。The computer-readable storage medium may be non-volatile or volatile.
本申请计算机可读存储介质具体实施方式与上述表格文本智能过滤装置和方法各实施例基本相同,在此不作累述。The specific implementation of the computer-readable storage medium of the present application is basically the same as the foregoing embodiments of the table text intelligent filtering device and method, and will not be repeated here.
需要说明的是,上述本申请实施例序号仅仅为了描述,不代表实施例的优劣。并且本文中的术语“包括”、“包含”或者其任何其他变体意在涵盖非排他性的包含,从而使得包括一系列要素的过程、装置、物品或者方法不仅包括那些要素,而且还包括没有明确列出的其他要素,或者是还包括为这种过程、装置、物品或者方法所固有的要素。在没有更多限制的情况下,由语句“包括一个……”限定的要素,并不排除在包括该要素的过程、装置、物品或者方法中还存在另外的相同要素。It should be noted that the serial numbers of the above-mentioned embodiments of the present application are only for description, and do not represent the superiority or inferiority of the embodiments. And the terms "include", "include" or any other variants thereof in this article are intended to cover non-exclusive inclusion, so that a process, device, article or method including a series of elements not only includes those elements, but also includes those elements that are not explicitly included. The other elements listed may also include elements inherent to the process, device, article, or method. If there are no more restrictions, the element defined by the sentence "including a..." does not exclude the existence of other identical elements in the process, device, article, or method that includes the element.
通过以上的实施方式的描述,本领域的技术人员可以清楚地了解到上述实施例方法可借助软件加必需的通用硬件平台的方式来实现,当然也可以通过硬件,但很多情况下前者是更佳的实施方式。基于这样的理解,本申请的技术方案本质上或者说对现有技术做出贡献的部分可以以软件产品的形式体现出来,该计算机软件产品存储在如上所述的一个存储介质(如ROM/RAM、磁碟、光盘)中,包括若干指令用以使得一台终端设备(可以是手机,计算机,服务器,或者网络设备等)执行本申请各个实施例所述的方法。Through the description of the above implementation manners, those skilled in the art can clearly understand that the above-mentioned embodiment method can be implemented by means of software plus the necessary general hardware platform, of course, it can also be implemented by hardware, but in many cases the former is better.的实施方式。 Based on this understanding, the technical solution of this application essentially or the part that contributes to the existing technology can be embodied in the form of a software product, and the computer software product is stored in a storage medium (such as ROM/RAM) as described above. , Magnetic disk, optical disk), including a number of instructions to make a terminal device (which can be a mobile phone, a computer, a server, or a network device, etc.) execute the method described in each embodiment of the present application.
以上仅为本申请的优选实施例,并非因此限制本申请的专利范围,凡是利用本申请说明书及附图内容所作的等效结构或等效流程变换,或直接或间接运用在其他相关的技术领域,均同理包括在本申请的专利保护范围内。The above are only the preferred embodiments of the application, and do not limit the scope of the patent for this application. Any equivalent structure or equivalent process transformation made using the content of the description and drawings of the application, or directly or indirectly applied to other related technical fields , The same reason is included in the scope of patent protection of this application.

Claims (20)

  1. 一种表格文本智能过滤方法,其中,所述方法包括:An intelligent filtering method for table text, wherein the method includes:
    获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
    利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
    对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
    利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  2. 如权利要求1所述的表格文本智能过滤方法,其中,所述将所述表格图像集进行预处理操作,得到标准表格图像集,包括:8. The method for intelligently filtering table text according to claim 1, wherein the preprocessing operation of the table image set to obtain a standard table image set comprises:
    根据各比例法对所述表格图像集进行图像灰度化处理后得到灰度表格图像集,利用对比度拉伸方式对所述灰度表格图像集进行对比度增强,将对比度增强后的所述灰度表格图像集进行图像阈值化操作后得到所述标准表格图像集。Perform image gray-scale processing on the table image set according to each ratio method to obtain a gray-scale table image set, and use a contrast stretching method to perform contrast enhancement on the gray-scale table image set, and the grayscale after the contrast enhancement The standard table image set is obtained after the image thresholding operation is performed on the table image set.
  3. 如权利要求1所述的表格文本智能过滤方法,其中,所述利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集,包括:8. The intelligent filtering method for form text according to claim 1, wherein said using an image enhancement algorithm to perform enhancement processing on said standard form image set to obtain a form key image area set comprises:
    通过阈值分割法将所述标准表格图像集中的图像前景文字和图像背景图案进行分割;Segmenting the image foreground text and image background pattern in the standard table image set by a threshold segmentation method;
    利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:The Retinex algorithm is used to calculate the key information image area in the standard table image set after segmentation to obtain the table key image area, thereby combining to form the table key image area set, wherein the Retinex algorithm includes:
    S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
    其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area.
  4. 如权利要求1所述的表格文本智能过滤方法,其中,所述对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集,包括:The intelligent filtering method for form text according to claim 1, wherein said extracting the feature image from the key image region set of the form to obtain the feature form image set comprises:
    将所述表格关键图像区域集输入至残差块神经网络输入层中,利用所述残差块神经网络的隐藏层对所述表格关键图像区域集进行卷积操作,得到所述表格关键图像区域集的特征图谱集,通过所述残差块神经网络的输出层输出所述特征图谱集,从而得到所述特征表格图像集。The table key image area set is input into the residual block neural network input layer, and the hidden layer of the residual block neural network is used to perform a convolution operation on the table key image area set to obtain the table key image area The feature atlas set of the set is output through the output layer of the residual block neural network to obtain the feature table image set.
  5. 如权利要求1至4中任意一项所述的表格文本智能过滤方法,其中,所述利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,包括:The intelligent filtering method for form text according to any one of claims 1 to 4, wherein said using a pre-built form text filtering model to perform text position detection on said feature form image set comprises:
    在所述特征表格图像集中生成一个几何图,并将所述几何图按照预设的比例进行缩放,将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L gGenerate a geometric diagram in the feature table image set, and scale the geometric diagram according to a preset ratio, and input the scaled geometric diagram into the table text filtering model for training to obtain the scaled The geometric figure loss L g ;
    利用类平衡交叉熵计算缩放后的所述几何图中的文本损失L s Calculate the text loss L s in the zoomed geometric graph by using class balance cross entropy;
    将缩放后的所述几何图损失和文本损失输入至预设的损失函数中得到损失函数值,根据所述损失函数值对所述特征表格图像集进行文本位置检测。Inputting the scaled geometric graph loss and text loss into a preset loss function to obtain a loss function value, and performing text position detection on the feature table image set according to the loss function value.
  6. 如权利要求5所述的表格文本智能过滤方法,其中,所述将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g,包括: The intelligent filtering method for table text according to claim 5, wherein said inputting the scaled geometric figure into the table text filtering model for training to obtain the scaled geometric figure loss L g includes :
    将缩放后的所述几何图输入到所述表格文本过滤模型的输入层中;Input the zoomed geometric figure into the input layer of the table text filtering model;
    通过所述表格文本过滤模型的隐藏层对缩放后的所述几何图进行特征合并,得到特征图,其中,所述隐藏层包含卷积层和池化层;Performing feature merging on the zoomed geometric map through the hidden layer of the table text filtering model to obtain a feature map, wherein the hidden layer includes a convolutional layer and a pooling layer;
    通过所述表格文本过滤模型的输出层对所述特征图进行边框回归,从而输出所述几何图的损失L gPerform frame regression on the feature map through the output layer of the table text filtering model, thereby outputting the loss L g of the geometric map.
  7. 一种电子设备,其中,所述电子设备包括存储器和处理器,所述存储器上存储有可在所述处理器上运行的表格文本过滤程序,所述表格文本过滤程序被所述处理器执行时实现如下步骤:An electronic device, wherein the electronic device includes a memory and a processor, the memory stores a table text filtering program that can be run on the processor, and when the table text filtering program is executed by the processor To achieve the following steps:
    获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
    利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
    对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
    利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  8. 如权利要求7所述的电子设备,其中,所述将所述表格图像集进行预处理操作,得到标准表格图像集,包括:8. The electronic device according to claim 7, wherein said performing a preprocessing operation on said form image set to obtain a standard form image set comprises:
    根据各比例法对所述表格图像集进行图像灰度化处理后得到灰度表格图像集,利用对比度拉伸方式对所述灰度表格图像集进行对比度增强,将对比度增强后的所述灰度表格图像集进行图像阈值化操作后得到所述标准表格图像集。Perform image gray-scale processing on the table image set according to each ratio method to obtain a gray-scale table image set, and use a contrast stretching method to perform contrast enhancement on the gray-scale table image set, and the grayscale after the contrast enhancement The standard table image set is obtained after the image thresholding operation is performed on the table image set.
  9. 如权利要求7所述的电子设备,其中,所述利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集,包括:8. The electronic device according to claim 7, wherein said using an image enhancement algorithm to perform enhancement processing on said standard table image set to obtain a table key image area set comprises:
    通过阈值分割法将所述标准表格图像集中的图像前景文字和图像背景图案进行分割;Segmenting the image foreground text and image background pattern in the standard table image set by a threshold segmentation method;
    利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:The Retinex algorithm is used to calculate the key information image area in the standard table image set after segmentation to obtain the table key image area, thereby combining to form the table key image area set, wherein the Retinex algorithm includes:
    S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
    其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area.
  10. 如权利要求7所述的电子设备,其中,所述对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集,包括:8. The electronic device according to claim 7, wherein said performing feature image extraction on said table key image area set to obtain a feature table image set comprises:
    将所述表格关键图像区域集输入至残差块神经网络输入层中,利用所述残差块神经网络的隐藏层对所述表格关键图像区域集进行卷积操作,得到所述表格关键图像区域集的特征图谱集,通过所述残差块神经网络的输出层输出所述特征图谱集,从而得到所述特征表格图像集。The table key image area set is input into the residual block neural network input layer, and the hidden layer of the residual block neural network is used to perform a convolution operation on the table key image area set to obtain the table key image area The feature atlas set of the set is output through the output layer of the residual block neural network to obtain the feature table image set.
  11. 如权利要求7至10中任意一项所述的电子设备,其中,所述利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,包括:8. The electronic device according to any one of claims 7 to 10, wherein said using a pre-built form text filtering model to perform text position detection on said feature form image set comprises:
    在所述特征表格图像集中生成一个几何图,并将所述几何图按照预设的比例进行缩放,将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L gGenerate a geometric diagram in the feature table image set, and scale the geometric diagram according to a preset ratio, and input the scaled geometric diagram into the table text filtering model for training to obtain the scaled The geometric figure loss L g ;
    利用类平衡交叉熵计算缩放后的所述几何图中的文本损失L s Calculate the text loss L s in the zoomed geometric graph by using class balance cross entropy;
    将缩放后的所述几何图损失和文本损失输入至预设的损失函数中得到损失函数值,根据所述损失函数值对所述特征表格图像集进行文本位置检测。Inputting the scaled geometric graph loss and text loss into a preset loss function to obtain a loss function value, and performing text position detection on the feature table image set according to the loss function value.
  12. 如权利要求11所述的电子设备,其中,所述将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g,包括: The electronic device as claimed in claim 11, wherein the geometry of the scaled text input to the filter table of the geometric loss L g obtained after the scaled training model, comprising:
    将缩放后的所述几何图输入到所述表格文本过滤模型的输入层中;Input the zoomed geometric figure into the input layer of the table text filtering model;
    通过所述表格文本过滤模型的隐藏层对缩放后的所述几何图进行特征合并,得到特征图,其中,所述隐藏层包含卷积层和池化层;Performing feature merging on the zoomed geometric map through the hidden layer of the table text filtering model to obtain a feature map, wherein the hidden layer includes a convolutional layer and a pooling layer;
    通过所述表格文本过滤模型的输出层对所述特征图进行边框回归,从而输出所述几何图的损失L gPerform frame regression on the feature map through the output layer of the table text filtering model, thereby outputting the loss L g of the geometric map.
  13. 一种计算机可读存储介质,其中,所述计算机可读存储介质上存储有表格文本过滤程序,所述表格文本过滤程序可被一个或者多个处理器执行,以实现如下所述的表格文本智能过滤方法的步骤:A computer-readable storage medium, wherein a table text filtering program is stored on the computer-readable storage medium, and the table text filtering program can be executed by one or more processors to realize the table text intelligence as described below Steps of the filtering method:
    获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;Acquiring a document-based form image set, and performing a preprocessing operation on the form image set to obtain a standard form image set;
    利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;Performing enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
    对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;Performing feature image extraction on the table key image area set to obtain a feature table image set;
    利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。Use the pre-built table text filtering model to detect the text position of the feature table image set. If the position of the text in the feature table image of the feature table image set is detected, filter the text and save the feature For the table image, if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  14. 如权利要求13所述的计算机可读存储介质,其中,所述将所述表格图像集进行预处理操作,得到标准表格图像集,包括:15. The computer-readable storage medium of claim 13, wherein the preprocessing operation on the form image set to obtain a standard form image set comprises:
    根据各比例法对所述表格图像集进行图像灰度化处理后得到灰度表格图像集,利用对比度拉伸方式对所述灰度表格图像集进行对比度增强,将对比度增强后的所述灰度表格图像集进行图像阈值化操作后得到所述标准表格图像集。Perform image gray-scale processing on the table image set according to each ratio method to obtain a gray-scale table image set, and use a contrast stretching method to perform contrast enhancement on the gray-scale table image set, and the grayscale after the contrast enhancement The standard table image set is obtained after the image thresholding operation is performed on the table image set.
  15. 如权利要求13所述的计算机可读存储介质,其中,所述利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集,包括:15. The computer-readable storage medium according to claim 13, wherein said using an image enhancement algorithm to perform enhancement processing on said standard table image set to obtain a table key image area set comprises:
    通过阈值分割法将所述标准表格图像集中的图像前景文字和图像背景图案进行分割;Segmenting the image foreground text and image background pattern in the standard table image set by a threshold segmentation method;
    利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:The Retinex algorithm is used to calculate the key information image area in the standard table image set after segmentation to obtain the table key image area, thereby combining to form the table key image area set, wherein the Retinex algorithm includes:
    S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
    其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area.
  16. 如权利要求13所述的计算机可读存储介质,其中,所述对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集,包括:15. The computer-readable storage medium according to claim 13, wherein said performing feature image extraction on said table key image area set to obtain a feature table image set comprises:
    将所述表格关键图像区域集输入至残差块神经网络输入层中,利用所述残差块神经网络的隐藏层对所述表格关键图像区域集进行卷积操作,得到所述表格关键图像区域集的特征图谱集,通过所述残差块神经网络的输出层输出所述特征图谱集,从而得到所述特征表格图像集。The table key image area set is input into the residual block neural network input layer, and the hidden layer of the residual block neural network is used to perform a convolution operation on the table key image area set to obtain the table key image area The feature atlas set of the set is output through the output layer of the residual block neural network to obtain the feature table image set.
  17. 如权利要求13至16中任意一项所述的计算机可读存储介质,其中,所述利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,包括:15. The computer-readable storage medium according to any one of claims 13 to 16, wherein said using a pre-built table text filtering model to perform text position detection on said feature table image set comprises:
    在所述特征表格图像集中生成一个几何图,并将所述几何图按照预设的比例进行缩放,将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L gGenerate a geometric diagram in the feature table image set, and scale the geometric diagram according to a preset ratio, and input the scaled geometric diagram into the table text filtering model for training to obtain the scaled The geometric figure loss L g ;
    利用类平衡交叉熵计算缩放后的所述几何图中的文本损失L s Calculate the text loss L s in the zoomed geometric graph by using class balance cross entropy;
    将缩放后的所述几何图损失和文本损失输入至预设的损失函数中得到损失函数值,根据所述损失函数值对所述特征表格图像集进行文本位置检测。Inputting the scaled geometric graph loss and text loss into a preset loss function to obtain a loss function value, and performing text position detection on the feature table image set according to the loss function value.
  18. 如权利要求17所述的计算机可读存储介质,其中,所述将缩放后的所述几何图输入至所述表格文本过滤模型中进行训练后得到缩放后的所述几何图损失L g,包括: The computer-readable storage medium according to claim 17, wherein the input of the zoomed geometric figure into the table text filtering model for training to obtain the zoomed geometric figure loss L g comprises :
    将缩放后的所述几何图输入到所述表格文本过滤模型的输入层中;Input the zoomed geometric figure into the input layer of the table text filtering model;
    通过所述表格文本过滤模型的隐藏层对缩放后的所述几何图进行特征合并,得到特征图,其中,所述隐藏层包含卷积层和池化层;Performing feature merging on the zoomed geometric map through the hidden layer of the table text filtering model to obtain a feature map, wherein the hidden layer includes a convolutional layer and a pooling layer;
    通过所述表格文本过滤模型的输出层对所述特征图进行边框回归,从而输出所述几何图的损失L gPerform frame regression on the feature map through the output layer of the table text filtering model, thereby outputting the loss L g of the geometric map.
  19. 一种表格文本智能过滤装置,其中,所述装置包括:A table text intelligent filtering device, wherein the device includes:
    图像预处理模块,用于获取基于文档的表格图像集,将所述表格图像集进行预处理操作,得到标准表格图像集;The image preprocessing module is used to obtain a document-based table image set, and perform a preprocessing operation on the table image set to obtain a standard table image set;
    增强处理模块,用于利用图像增强算法对所述标准表格图像集进行增强处理,得到表格关键图像区域集;An enhancement processing module, configured to perform enhancement processing on the standard table image set by using an image enhancement algorithm to obtain a table key image area set;
    特征提取模块,用于对所述表格关键图像区域集进行特征图像提取,得到特征表格图像集;The feature extraction module is used to perform feature image extraction on the table key image area set to obtain a feature table image set;
    过滤模块,用于利用预先构建的表格文本过滤模型对所述特征表格图像集进行文本位置检测,若检测出所述特征表格图像集的特征表格图像中文本的位置,则将所述文本进行过滤后保存所述特征表格图像,若没有检测出所述特征表格图像集的特征表格图像中文本的位置,直接保存所述特征表格图像,从而完成所述表格图像集的文本过滤。The filtering module is configured to use a pre-built table text filtering model to perform text position detection on the feature table image set, and filter the text if the position of the text in the feature table image of the feature table image set is detected After that, the characteristic table image is saved, and if the position of the text in the characteristic table image of the characteristic table image set is not detected, the characteristic table image is directly saved, thereby completing the text filtering of the table image set.
  20. 如权利要求19所述的表格文本智能过滤装置,其中,所述增强处理模块包括:The table text intelligent filtering device according to claim 19, wherein the enhanced processing module comprises:
    分割模块,用于通过阈值分割法将所述标准表格图像集中的图像前景文字和图像背景图案进行分割;A segmentation module, configured to segment the image foreground text and image background pattern in the standard table image set by a threshold segmentation method;
    计算模块,用于利用Retinex算法计算出分割后的所述标准表格图像集中的关键信息图像区域,得到表格关键图像区域,从而组合形成所述表格关键图像区域集,其中,所述Retinex算法包括:The calculation module is configured to use the Retinex algorithm to calculate the key information image area in the standard table image set after segmentation to obtain the table key image area, thereby combining to form the table key image area set, wherein the Retinex algorithm includes:
    S(x,y)=R(x,y)×L(x,y)S(x,y)=R(x,y)×L(x,y)
    其中,S(x,y)表示表格关键图像区域,R(x,y)表示反射光图像,L(x,y)代表光亮度图像,x表示表格关键图像区域的横坐标,y表示表格关键图像区域的纵坐标。Among them, S(x,y) represents the table key image area, R(x,y) represents the reflected light image, L(x,y) represents the brightness image, x represents the abscissa of the table key image area, and y represents the table key The ordinate of the image area.
PCT/CN2020/112334 2019-10-11 2020-08-30 Method and apparatus for intelligently filtering table text, and computer-readable storage medium WO2021068682A1 (en)

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201910965807.1 2019-10-11
CN201910965807.1A CN110929561B (en) 2019-10-11 2019-10-11 Intelligent form text filtering method and device and computer readable storage medium

Publications (1)

Publication Number Publication Date
WO2021068682A1 true WO2021068682A1 (en) 2021-04-15

Family

ID=69848874

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2020/112334 WO2021068682A1 (en) 2019-10-11 2020-08-30 Method and apparatus for intelligently filtering table text, and computer-readable storage medium

Country Status (2)

Country Link
CN (1) CN110929561B (en)
WO (1) WO2021068682A1 (en)

Families Citing this family (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110929561B (en) * 2019-10-11 2024-04-12 平安科技(深圳)有限公司 Intelligent form text filtering method and device and computer readable storage medium

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106897690A (en) * 2017-02-22 2017-06-27 南京述酷信息技术有限公司 PDF table extracting methods
CN110163198A (en) * 2018-09-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of Table recognition method for reconstructing, device and storage medium
US20190310868A1 (en) * 2017-01-26 2019-10-10 Nice Ltd. Method and system for accessing table content in a digital image of the table
CN110929561A (en) * 2019-10-11 2020-03-27 平安科技(深圳)有限公司 Intelligent filtering method and device for table text and computer readable storage medium

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9715748B2 (en) * 2014-12-04 2017-07-25 The United States Of America As Represented By The Secretary Of The Air Force Method and apparatus for graphical data interaction and vizualization of graphs via paths

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20190310868A1 (en) * 2017-01-26 2019-10-10 Nice Ltd. Method and system for accessing table content in a digital image of the table
CN106897690A (en) * 2017-02-22 2017-06-27 南京述酷信息技术有限公司 PDF table extracting methods
CN110163198A (en) * 2018-09-27 2019-08-23 腾讯科技(深圳)有限公司 A kind of Table recognition method for reconstructing, device and storage medium
CN110929561A (en) * 2019-10-11 2020-03-27 平安科技(深圳)有限公司 Intelligent filtering method and device for table text and computer readable storage medium

Also Published As

Publication number Publication date
CN110929561B (en) 2024-04-12
CN110929561A (en) 2020-03-27

Similar Documents

Publication Publication Date Title
CN110853047B (en) Intelligent image segmentation and classification method, device and computer readable storage medium
CN110738203B (en) Field structured output method, device and computer readable storage medium
US20200342277A1 (en) Colorspace encoding multimedia data on a physical page
US9930218B2 (en) Content aware improvement of captured document images
CN109657582B (en) Face emotion recognition method and device, computer equipment and storage medium
US9262690B2 (en) Method and device for detecting glare pixels of image
CN109117846B (en) Image processing method and device, electronic equipment and computer readable medium
US10198628B2 (en) Method and apparatus for determining a document suitability for server-based optical character recognition (OCR) processing
WO2021012494A1 (en) Deep learning-based face recognition method and apparatus, and computer-readable storage medium
Yang et al. Improved local binary pattern for real scene optical character recognition
US11886492B2 (en) Method of matching image and apparatus thereof, device, medium and program product
JP2006285944A (en) Device and method for detecting structural element of subject
CN110717497A (en) Image similarity matching method and device and computer readable storage medium
US9235757B1 (en) Fast text detection
WO2021012493A1 (en) Short video keyword extraction method and apparatus, and storage medium
CN107330387B (en) Pedestrian detection method based on image data
US20090169108A1 (en) System and method for recognizing smiling faces captured by a mobile electronic device
CN110335206B (en) Intelligent filter method, device and computer readable storage medium
CN114519858B (en) Document image recognition method and device, storage medium and electronic equipment
WO2021068682A1 (en) Method and apparatus for intelligently filtering table text, and computer-readable storage medium
CN113222921A (en) Image processing method and system
CN110210467B (en) Formula positioning method of text image, image processing device and storage medium
RU2633182C1 (en) Determination of text line orientation
CN110705547B (en) Method and device for recognizing text in image and computer readable storage medium
US8971669B2 (en) Method and apparatus for image processing

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 20874859

Country of ref document: EP

Kind code of ref document: A1

NENP Non-entry into the national phase

Ref country code: DE

122 Ep: pct application non-entry in european phase

Ref document number: 20874859

Country of ref document: EP

Kind code of ref document: A1