CN113239932A - Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device) - Google Patents

Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device) Download PDF

Info

Publication number
CN113239932A
CN113239932A CN202110560479.4A CN202110560479A CN113239932A CN 113239932 A CN113239932 A CN 113239932A CN 202110560479 A CN202110560479 A CN 202110560479A CN 113239932 A CN113239932 A CN 113239932A
Authority
CN
China
Prior art keywords
tesseract
ocr
file
pfd
data set
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110560479.4A
Other languages
Chinese (zh)
Inventor
赵亮
王迪
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xian University of Architecture and Technology
Original Assignee
Xian University of Architecture and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xian University of Architecture and Technology filed Critical Xian University of Architecture and Technology
Priority to CN202110560479.4A priority Critical patent/CN113239932A/en
Publication of CN113239932A publication Critical patent/CN113239932A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/148Segmentation of character regions
    • G06V30/153Segmentation of character regions using recognition of characters or words
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/044Recurrent networks, e.g. Hopfield networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V2201/00Indexing scheme relating to image or video recognition or understanding
    • G06V2201/02Recognising information on displays, dials, clocks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • General Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)

Abstract

The invention discloses a Tesseract-OCR-based method for identifying a hollow speed scale in a PFD (pulse frequency device) of an aircraft main display, and belongs to the field of instrument identification. The identification method of the air speed scale in the aircraft main display PFD based on Tesseract-OCR comprises the steps of preprocessing a picture, extracting and storing the characteristics of incomplete characters by utilizing a Tesseract-OCR targeted training data set, and solving the problems that the identification rate is low and even the identification cannot be realized when the incomplete characters exist in roller type digital display of the air speed scale in an aircraft instrument; and then, by combining the LSTM neural network model, the incomplete character features can be effectively trained, and the recognition rate of incomplete characters is improved.

Description

Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device)
Technical Field
The invention belongs to the field of instrument recognition, and particularly relates to a method for recognizing a hollow speed scale in a PFD (pulse frequency device) of an aircraft main display based on Tesseract-OCR (optical character recognition).
Background
In instrument Recognition, the Optical Character Recognition (OCR) algorithm is a common instrument Recognition method, and in 1985, the HP laboratory developed an OCR engine of Tesseract, whose basic theory is a process of determining a shape by detecting dark and light patterns and then translating the shape into a computer text by a Character Recognition method. The defects of the existing Tesseract OCR algorithm are shown as follows: (1) when incomplete characters appear in the roller type digital display of an airspeed scale in an aeronautical instrument, the existing data set cannot play a role, and the situation that the characters cannot be identified can occur. (2) When a targeted data set is trained, the recognition rate is not high.
Disclosure of Invention
The invention aims to overcome the defects of the prior art and provides a method for identifying an air speed scale in an aircraft main display PFD based on Tesseract-OCR.
In order to achieve the purpose, the invention adopts the following technical scheme to realize the purpose:
a method for identifying an air speed scale in an aircraft main display PFD based on Tesseract-OCR comprises the following steps:
(1) sequentially carrying out graying and binarization preprocessing on the collected picture to obtain a preprocessed picture;
(2) carrying out character segmentation based on the preprocessed picture to obtain complete characters and incomplete characters;
(3) carrying out targeted training on the incomplete characters by using Tesseract-OCR to obtain a data set;
(4) integrating an LSTM neural network model by using the data set to obtain a new data set;
(5) and carrying out graying and binarization preprocessing on the picture to be recognized, and calling and combining the new data set to recognize on Tesseract-OCR.
Further, the step (1) utilizes a conversion formula from RGB to a gray scale map to perform graying:
GARY=RED*0.299+GREEN*0.588+BLUE*0.133。
further, the step (1) divides the image into two parts, namely a foreground and a background, by setting a preset threshold value for binarization.
Further, binarization is performed by adopting the following formula:
Figure BDA0003078766990000021
wherein f (x, y) is an original image; t is a gray threshold; g (x, y) is a binary image obtained by threshold value operation.
Further, in the step (2), character segmentation is carried out on the preprocessed picture by adopting a vertical projection method.
Further, the specific process of the step (3) is as follows:
training the incomplete character set by a jessboxeditor;
correcting the Box files of the incomplete characters one by one, and defining a font characteristic file after the correction is finished;
and creating a batch preprocessing file under the directory where the sample picture is located, executing the batch preprocessing file to obtain a finally generated language file, and copying the language file into a tessdata file of the program to obtain a trained database.
Further, the step (4) specifically comprises:
(401) extracting lstmf files from tif and box files generated from the data set for training of an LSTM neural network model;
(402) extracting an LSTM file from the traineddata file to obtain an LSTM neural network model;
(403) beginning training when the stage file eng.lstm is generated, and ending training when the error rate of the LSTM neural network model is lower than 0.01;
(404) and generating a checkpoint file after training is finished, combining the language file generated by the data set and the checkpoint file to generate a new language file, and placing the new language file in a tessdata folder to obtain a new data set.
Further, the number of training times in step (403) is 6000.
Compared with the prior art, the invention has the following beneficial effects:
the identification method of the air speed scale in the aircraft main display PFD based on Tesseract-OCR comprises the steps of preprocessing a picture, extracting and storing the characteristics of incomplete characters by utilizing a Tesseract-OCR targeted training data set, and solving the problems that the identification rate is low and even the identification cannot be realized when the incomplete characters exist in roller type digital display of the air speed scale in an aircraft instrument; and then, by combining the LSTM neural network model, incomplete character features can be effectively trained, and the recognition rate of incomplete characters is improved.
Drawings
FIG. 1 is a schematic flow diagram of the present invention;
FIG. 2 is a picture after graying of an embodiment;
FIG. 3 is a picture after binarization according to the embodiment;
fig. 4 is a picture in an embodiment, wherein fig. 4(a), fig. 4(b), fig. 4(c), and fig. 4(d) are an original image, a binarized image, a histogram of character projection, and a divided character image, respectively;
FIG. 5 is a partially cut character set according to an embodiment;
FIG. 6 is a diagram illustrating the effect of the embodiment; wherein, fig. 6(a) is an effect graph of recognition using a self-contained data set, fig. 6(b) is an effect graph of recognition using a trained data set, and fig. 6(c) is an effect graph of database recognition combined with an LSTM neural network;
FIG. 7 is a graph comparing the recognition rates of examples.
Detailed Description
In order to make the technical solutions of the present invention better understood by those skilled in the art, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be obtained by a person skilled in the art without any inventive step based on the embodiments of the present invention, shall fall within the scope of the present invention.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are capable of operation in sequences other than those illustrated or described herein. Furthermore, the terms "comprises," "comprising," and any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
In consideration of the complexity of the aerometer, the invention carries out targeted data set training aiming at the condition that incomplete characters exist in the roller type digital display of the airspeed scale in the aerometer, solves the problem that incomplete characters cannot be identified, and improves the identification rate when incomplete characters exist by combining with an LSTM neural network model.
The invention is described in further detail below with reference to the accompanying drawings:
referring to fig. 1, fig. 1 is a schematic flow chart of the present invention, and a method for identifying an air velocity scale in a PFD of an aircraft main display based on Tesseract-OCR includes the following steps:
(1) pretreatment: carrying out graying and binaryzation on the collected picture so that the picture can be better identified;
(2) character segmentation: carrying out character segmentation on the preprocessed picture, dividing the preprocessed picture into complete characters and incomplete characters, and carrying out targeted character set training on the incomplete characters so as to carry out subsequent recognition;
(3) training data set: training the incomplete characters obtained after segmentation to obtain a data set;
(4) training an LSTM neural network model: extracting an lstmf file from a file generated by a previously trained sample data set for lstm training, then extracting an lstm file from a traineddata file, starting training from a generated stage file, training 6000 times to enable the error rate to be lower than 0.01, combining a checkpoint file generated after the training and a language file generated by the previously trained data set to generate a new language file, and placing the new language file in a tessdata folder;
(5) character recognition: and preprocessing the picture to be recognized, and calling the trained data set through codes to finish recognition.
In summary, the recognition method of the air speed scale in the aircraft main display PFD based on Tesseract-OCR provided by the invention firstly preprocesses the acquired picture to enable the picture to be better recognized, can improve the algorithm recognition rate, needs to perform targeted character set training because the main display PFD number in the aircraft cockpit is displayed in a rolling mode, adopts jessboxer training to extract and store the characteristics of incomplete characters, and then combines Tesseract OCR with a neural network model to enable the obtained incomplete character characteristics to be effectively trained, and improves the recognition rate.
In the embodiment of the present invention, step (1) specifically includes the following steps:
the collected picture is preprocessed, wherein the preprocessing comprises graying and binarization, so that subsequent preprocessing is facilitated, as the human eyes have the strongest green sensitivity and the weakest blue, the green channel weight is the largest, the blue channel weight is the smallest, and the conversion formula from RGB to a gray scale map is as follows:
GARY=RED*0.299+GREEN*0.588+BLUE*0.133
in the formula, GARY represents gray; RED represents the RED component; GREEN represents the GREEN component; BLUE represents the BLUE component.
After the image to be recognized is grayed, some interference factors need to be removed for better recognition, so that binarization preprocessing is performed, the preprocessed image is a black-and-white image, and the foreground can be highlighted and separated from the background.
The binarization preprocessing is as follows:
Figure BDA0003078766990000061
wherein f (x, y) is an original image; t is a gray threshold; g (x, y) is a binary image obtained by threshold value operation.
Referring to fig. 2 and fig. 3, fig. 2 is a picture after graying, and fig. 3 is a picture after binarization. In the embodiment of the present invention, the threshold value 194 is selected as a segmentation value for better preprocessing.
In step (2) of the embodiment of the present invention, the preprocessed graph is cut to cut out a portion to be recognized, and then a vertical projection method is used to perform character segmentation, wherein the specific vertical projection segmentation method is as follows:
the character cutting adopts a vertical projection method in OpenCV, the principle is to analyze a distribution histogram of pixels of a binary image so as to find out boundary points of adjacent characters for segmentation, and when an image only consists of an object and a background, a gray level histogram becomes an obvious double-peak value.
At this time, it is necessary to find a valley value to perform segmentation, find a first peak value and a second peak value, and then find a valley value between the first peak value and the second peak value, thereby determining a threshold value of the valley value. OTSU, Chinese translation is the maximum inter-class variance method, and is proposed by Japanese scientists at the end of the 20 th 70 s, and the method is used for researching the gray level characteristics of an image and enabling the image to have a foreground part and a background part according to the gray level distribution of the image. The value of the inter-class method between the foreground image and the background image is calculated. The higher the value, the larger the difference between the foreground image and the background image. For this reason, in the erroneous classification, the background image also takes the contents of the target image, or some contents in the background image are listed in the target image. Thus, the variance becomes small. Therefore, if the dispersion of the object from the background image is large, the error rate will be reduced.
The principle of the OTSU algorithm is as follows: when the foreground background image is segmented, setting the threshold value as t and setting the ratio of the total number of the foreground image pixel points to the total image pixel points as w0The average gray level of the foreground image is u0Similarly, the value of the number of background image pixels to the total image pixels is set as w1The corresponding order is u1And u0The total average gray scale is as follows.
u=w0u0=w1u1
The variance of the foreground and background images is:
g=w0(u0-u)2+(u1-u)2
in actual calculations, an equivalent formula can be used:
σ2=w0u1(u0±u1)2
when the variance of the foreground and background images is large, i.e. g in the formula is large, t is the optimal segmentation threshold of the image. The method is used for operating the instrument image according to the characteristics of the aircraft instrument image. The character cutting pattern is shown in fig. 4.
In the embodiment of the invention, in the step (3), 1000 sample pictures are collected and preprocessed, the preprocessed pictures are subjected to character segmentation by using a vertical projection method, the obtained incomplete character set is trained by using a jessboarder, and the partially segmented character set is shown in fig. 5.
After Tesseract-OCR is downloaded, Java JDK is downloaded firstly, Jtessboxeditors are installed after downloading is completed, so that Box files need to be corrected one by one, the more data is, the larger the workload is, after correction is good, font feature files are defined, a batch preprocessing file is created under a directory where sample pictures are located, the batch preprocessing file is executed to obtain a finally generated language file, and the finally generated language file is copied into a tessdata file of a program, namely a trained database can be used.
Step (4) of the embodiment of the present invention specifically includes:
(401) extracting lstmf files from tif and box files generated from the data set in the step (3) for LSTM neural network model training;
(402) extracting lstm files from traineddata files;
(403) starting training from the generated stage file eng.lstm, and ending the training until the error rate is lower than 0.01, wherein the training is carried out for 6000 times;
(404) and generating a checkpoint file after training, combining the checkpoint file with a new language file generated by the previously trained sample data set to generate a new language file, placing the synthesized file in a tessdata folder, and identifying through code calling to test the identification rate.
In the embodiment of the invention, after the picture to be recognized is preprocessed, the data set of the self-contained library, the trained data set and the trained data set are respectively called by codes to be recognized by combining with an LSTM model, and the recognition rate is tested.
Referring to fig. 6, fig. 6(a) is a graph showing the effect of recognition using a self-contained data set, fig. 6(b) is a graph showing the effect of recognition using a self-trained data set, and fig. 6(c) is a graph showing the effect of database recognition combined with an LSTM neural network, and it can be seen that the recognition effect of fig. 6(c) is the best. Referring to fig. 7, it can be seen that the recognition rate of the present invention is greatly improved.
According to the identification method of the air velocity scale in the aircraft main display PFD based on Tesseract-OCR, the collected picture is preprocessed through OpenCV so as to be convenient for subsequent preprocessing, due to the complexity of an aircraft instrument, targeted character set training is carried out on the aircraft instrument, the characteristics of incomplete characters are extracted and stored, the problem that the incomplete characters cannot be identified is solved, the characteristics of the character set are effectively utilized by combining an LSTM neural network model, and the identification rate of the incomplete characters is improved.
The above-mentioned contents are only for illustrating the technical idea of the present invention, and the protection scope of the present invention is not limited thereby, and any modification made on the basis of the technical solution according to the technical idea proposed by the present invention falls within the protection scope of the claims of the present invention.

Claims (8)

1. A method for identifying an air speed scale in an aircraft main display PFD based on Tesseract-OCR is characterized by comprising the following steps:
(1) sequentially carrying out graying and binaryzation pretreatment on the collected picture to obtain a pretreated picture;
(2) carrying out character segmentation based on the preprocessed picture to obtain complete characters and incomplete characters;
(3) carrying out targeted training on the incomplete characters by using Tesseract-OCR to obtain a data set;
(4) integrating an LSTM neural network model by using the data set to obtain a new data set;
(5) and carrying out graying and binarization preprocessing on the picture to be recognized, and calling and combining the new data set to recognize on Tesseract-OCR.
2. The Tesseract-OCR based recognition method of the air velocity scale in the PFD of the main display of the airplane as claimed in claim 1, wherein the step (1) grays by using the conversion formula from RGB to gray scale map:
GARY=RED*0.299+GREEN*0.588+BLUE*0.133。
3. the Tesseract-OCR based identification method of the air velocity scale in the aircraft main display PFD according to claim 1, wherein the step (1) is performed by binarizing the image by dividing it into two parts-foreground and background by setting a preset threshold.
4. The Tesseract-OCR-based identification method of the air velocity scale in the PFD of the main display of the aircraft according to claim 3, characterized in that binarization is performed by using the following formula:
Figure FDA0003078766980000011
wherein f (x, y) is an original image; t is a gray threshold; g (x, y) is a binary image obtained by threshold value operation.
5. The Tesseract-OCR-based identification method of the air speed scale in the PFD of the main display of the airplane as claimed in claim 1, wherein the step (2) uses a vertical projection method to perform character segmentation on the preprocessed picture.
6. The Tesseract-OCR-based identification method of the air speed scale in the PFD of the main display of the airplane as claimed in claim 1, wherein the specific process of the step (3) is as follows:
training the incomplete character set by a jessboxeditor;
correcting the Box files of the incomplete characters one by one, and defining a font characteristic file after the correction is finished;
and creating a batch preprocessing file under the directory where the sample pictures are located, executing the batch preprocessing file to obtain a finally generated language file, and copying the language file into a tessdata file of the program to obtain the trained database.
7. The Tesseract-OCR-based identification method of the air speed scale in the aircraft main display PFD according to claim 6, wherein the step (4) specifically comprises:
(401) extracting lstmf files from tif and box files generated from the data set for training of an LSTM neural network model;
(402) extracting an LSTM file from the traineddata file to obtain an LSTM neural network model;
(403) beginning training when the stage file eng.lstm is generated, and ending training when the error rate of the LSTM neural network model is lower than 0.01;
(404) and generating a checkpoint file after training is finished, combining the language file generated by the data set and the checkpoint file to generate a new language file, and placing the new language file in a tessdata folder to obtain a new data set.
8. The Tesseract-OCR based recognition method for air velocity scales in PFD main display of airplane as claimed in claim 7, wherein the number of training times in step (403) is 6000.
CN202110560479.4A 2021-05-21 2021-05-21 Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device) Pending CN113239932A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110560479.4A CN113239932A (en) 2021-05-21 2021-05-21 Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device)

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110560479.4A CN113239932A (en) 2021-05-21 2021-05-21 Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device)

Publications (1)

Publication Number Publication Date
CN113239932A true CN113239932A (en) 2021-08-10

Family

ID=77138124

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110560479.4A Pending CN113239932A (en) 2021-05-21 2021-05-21 Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device)

Country Status (1)

Country Link
CN (1) CN113239932A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115439861A (en) * 2022-09-30 2022-12-06 北京中盛益华科技有限公司 Water gauge recognition method based on OCR
CN118014072A (en) * 2024-04-10 2024-05-10 中国电建集团昆明勘测设计研究院有限公司 Construction method and system of knowledge graph for hydraulic and hydroelectric engineering

Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014040169A1 (en) * 2012-09-14 2014-03-20 Broadbandtv, Corp. Intelligent supplemental search engine optimization
CN110363095A (en) * 2019-06-20 2019-10-22 华南农业大学 A kind of recognition methods for table font
WO2020248513A1 (en) * 2019-06-11 2020-12-17 苏州玖物互通智能科技有限公司 Ocr method for comprehensive performance test

Patent Citations (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2014040169A1 (en) * 2012-09-14 2014-03-20 Broadbandtv, Corp. Intelligent supplemental search engine optimization
WO2020248513A1 (en) * 2019-06-11 2020-12-17 苏州玖物互通智能科技有限公司 Ocr method for comprehensive performance test
CN110363095A (en) * 2019-06-20 2019-10-22 华南农业大学 A kind of recognition methods for table font

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
PAK LAOWATANACHAI等: "Optical-based Limit Order Book Modelling using Deep Neural Networks", 《2020 8TH INTERNATIONAL ELECTRICAL ENGINEERING CONGRESS (IEECON)》 *
从你的全世界路过: "Tesseract-OCR 4.0LSTM训练流程(Windows环境下)", 《HTTPS://ZHUANLAN.ZHIHU.COM/P/77013854》 *
苗成强: "基于卷积神经网络的燃气表码字识别系统的设计与实现", 《中国优秀硕士学位论文全文数据库信息科技辑》 *
赵国生等: "《Python网络爬虫技术与实战》", 31 January 2021 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115439861A (en) * 2022-09-30 2022-12-06 北京中盛益华科技有限公司 Water gauge recognition method based on OCR
CN118014072A (en) * 2024-04-10 2024-05-10 中国电建集团昆明勘测设计研究院有限公司 Construction method and system of knowledge graph for hydraulic and hydroelectric engineering

Similar Documents

Publication Publication Date Title
CN108764074B (en) Subjective item intelligently reading method, system and storage medium based on deep learning
CN103824053B (en) The sex mask method and face gender detection method of a kind of facial image
CN103824091B (en) A kind of licence plate recognition method for intelligent transportation system
CN113239932A (en) Tesseract-OCR-based identification method for air velocity scale in PFD (flight display device)
CN104778470A (en) Character detection and recognition method based on component tree and Hough forest
CN109189965A (en) Pictograph search method and system
CN105260428A (en) Picture processing method and apparatus
CN110008912B (en) Social platform matching method and system based on plant identification
CN112749696A (en) Text detection method and device
Brisinello et al. Optical Character Recognition on images with colorful background
JP2008251029A (en) Character recognition device and license plate recognition system
CN115240210A (en) System and method for auxiliary exercise of handwritten Chinese characters
De Nardin et al. Few-shot pixel-precise document layout segmentation via dynamic instance generation and local thresholding
CN107729863B (en) Human finger vein recognition method
CN113011431A (en) Chinese character stroke segmentation and extraction method and system based on MaskRCNN
CN110766001B (en) Bank card number positioning and end-to-end identification method based on CNN and RNN
Bala et al. Image simulation for automatic license plate recognition
CN109558875A (en) Method, apparatus, terminal and storage medium based on image automatic identification
CN113421256B (en) Dot matrix text line character projection segmentation method and device
Munir et al. Automatic character extraction from handwritten scanned documents to build large scale database
CN115359562A (en) Sign language letter spelling recognition method based on convolutional neural network
CN113505784A (en) Automatic nail annotation analysis method and device, electronic equipment and storage medium
Rani et al. Object Detection in Natural Scene Images Using Thresholding Techniques
CN112712080A (en) Character recognition processing method for acquiring image by moving character screen
CN110298350A (en) A kind of efficient block letter Uighur words partitioning algorithm

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20210810

RJ01 Rejection of invention patent application after publication