CN111274863A - Text prediction method based on text peak probability density - Google Patents

Text prediction method based on text peak probability density Download PDF

Info

Publication number
CN111274863A
CN111274863A CN201911237840.9A CN201911237840A CN111274863A CN 111274863 A CN111274863 A CN 111274863A CN 201911237840 A CN201911237840 A CN 201911237840A CN 111274863 A CN111274863 A CN 111274863A
Authority
CN
China
Prior art keywords
text
predicted
mask
center
data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN201911237840.9A
Other languages
Chinese (zh)
Other versions
CN111274863B (en
Inventor
张发恩
孙天齐
袁智超
陆强
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Innovation Qizhi (chengdu) Technology Co Ltd
Original Assignee
Innovation Qizhi (chengdu) Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Innovation Qizhi (chengdu) Technology Co Ltd filed Critical Innovation Qizhi (chengdu) Technology Co Ltd
Priority to CN201911237840.9A priority Critical patent/CN111274863B/en
Publication of CN111274863A publication Critical patent/CN111274863A/en
Application granted granted Critical
Publication of CN111274863B publication Critical patent/CN111274863B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/40Document-oriented image-based pattern recognition
    • G06V30/41Analysis of document content
    • G06V30/413Classification of content, e.g. text, photographs or tables
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F17/00Digital computing or data processing equipment or methods, specially adapted for specific functions
    • G06F17/10Complex mathematical operations
    • G06F17/18Complex mathematical operations for evaluating statistical data, e.g. average values, frequency distributions, probability functions, regression analysis
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/22Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Data Mining & Analysis (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Pure & Applied Mathematics (AREA)
  • Mathematical Optimization (AREA)
  • Mathematical Analysis (AREA)
  • Computational Mathematics (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Operations Research (AREA)
  • Probability & Statistics with Applications (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Algebra (AREA)
  • Artificial Intelligence (AREA)
  • Databases & Information Systems (AREA)
  • Software Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Image Analysis (AREA)

Abstract

The invention relates to a text positioning method based on text peak probability density, which comprises the steps of preprocessing data, generating a required text mask and a text peak probability graph by disclosing the existing marking information of a data set; performing data enhancement on the processed training set, transmitting the data into a convolutional neural network for training, and continuously optimizing network parameters through a combined loss function to obtain an optimal model; firstly, loading the trained model parameters, transmitting the pictures to be predicted to the deep learning network to obtain a predicted text prediction result, and performing post-processing on prediction information to obtain an accurate text positioning result. According to the technical scheme, the learning of the neural network on the probability information of the text center is increased, the segmentation mass center of the text prediction image is obtained by matching with a post-processing algorithm, the center line of the curved text can be obtained, the boundary of the curved text is further obtained, and the long, multidirectional and curved text can be well processed.

Description

Text prediction method based on text peak probability density
Technical Field
The invention relates to the technical field of image processing, in particular to a text prediction method based on text peak probability density.
Background
Characters are visible anywhere in life, with the wide popularization of intelligent terminals and digital products, people can shoot scenes in which people are interested anytime and anywhere, the scenes contain a large amount of useful text information, such as road sign information, shop names, propaganda slogans and the like in natural scenes, and the text information can provide decision input for intelligent transportation, unmanned driving, data intelligence and the like.
In an industrial scene, laser engraving numbers, printing numbers, spot welding numbers and the like on parts need to be systematically input by workers to trace the source of products so as to accurately position to a production link, but human errors can exist when the workers observe, record and input the codes into a computer by eyes.
The intelligent, fast and efficient life style is created, and the current people are oriented, in different scenes, the background is complex and various, the surrounding environment can generate illumination or shading influence on character positioning, different characters and fonts can also bring great interference to a text positioning algorithm, and therefore the robust character positioning algorithm is the basis of character recognition.
The existing text positioning method mainly comprises the following two types:
based on the segmentation idea, the whole image is segmented, then the segmented image is sent to a trained text detector to judge whether characters exist or not, then the segmented image with the characters is combined into an image and is sent to a text detection model to be predicted, and text positioning information is obtained.
Based on the deep learning idea, feature extraction is carried out on the image, at least one text preselection area is selected by using an area recommendation network, the candidate areas are adjusted by a convolutional neural network, and finally the preselection areas are combined to obtain the final text area position.
The method carries out feature extraction and image combination by stages through the two models, is not an end-to-end process, has long positioning time, and easily separates character parts by image segmentation parts, thereby easily causing classification errors. The method has the advantages that the method is difficult to distinguish adjacent texts from bent texts, cannot effectively position characters with different sizes in a digital image, cannot identify the characters, and is poor in text positioning effect under a complex background.
Disclosure of Invention
The invention provides a text positioning method based on text peak probability density, designs a multi-scale text detection network, and solves a central line by adopting a segmentation centroid method so as to obtain an accurate prediction result.
A text positioning method based on text peak probability density comprises the following four steps:
s1) image preprocessing: performing data preprocessing, and generating a required text mask and a text peak probability chart by disclosing the existing labeling information of a data set;
s2) data enhancement: performing data enhancement on the processed training set, wherein the data enhancement comprises at least one random transformation of brightness, contrast, hue, saturation, rotation, mirror image, cutting, scaling and the like;
s3) network training: transmitting the data into a convolutional neural network for training, and continuously optimizing network parameters through a combined loss function to obtain an optimal model;
s4) prediction process: firstly, loading the trained model parameters, transmitting the pictures to be predicted to the deep learning network to obtain a predicted text prediction result, and performing post-processing on prediction information to obtain an accurate text positioning result.
In the above technical solution, step S1) specifically includes:
s11) data preprocessing: setting a feature map of a text line center area to be composed of 0-255 gray levels, wherein the numerical value of the center line of the text line is larger as the text line is closer, the center of the text is whiter, the edge is blacker, and a gradual change effect is presented;
s12) text mask generation: generating a required text mask and a text line center mask by disclosing the existing labeling information of the data set, wherein the text mask generates a text rectangular mask by text block positioning information in the data set, generally 4 corner point coordinates of a text block;
s13) text peak probability map: and obtaining a text block central line according to the 4 corner point coordinates marked by the text, wherein the central line is the peak highest point with the value of 255, the central line gradually decreases from the central line to the edge of the text block, and the value of the edge is 0, so that a text peak probability graph is obtained.
In the above technical solution, the convolutional neural network base module in step S3) includes a residual network module, a concat module, an upsampling module, and the like.
In the foregoing technical solution, the post-processing of the prediction information in step S4) is specifically:
s41) firstly, calculating the connected domain of the predicted text mask and the text peak probability graph and eliminating the smaller connected domain;
s42) calculating coordinates of four corner points of the predicted text box, dividing the text box, and calculating mass center coordinates of each divided block by using the prediction information;
s43) fitting the circle center and the radius of all centroids calculated by the text box by using a least square method, carrying out non-calculation on the longest edge of the predicted text box which is less than 2 times of the broadside, carrying out deviation on a centroid connecting line to obtain accurate text prediction information, carrying out polar coordinate transformation on a curved text box to finally obtain a horizontal text, and directly carrying out text cutting on the predicted box which does not accord with the segmentation condition.
The invention has the beneficial effects that:
1. different from the traditional text detection or segmentation mode, the learning of the neural network on the text center probability information is increased. The text peak probability map is just like a mountain (the peak is the center and the foot is the boundary), and the top of the mountain can more easily separate text instances (through semantic segmentation). Although text masks exist, the boundary between the boundary and the center is not accurate enough, and the network can be well helped to distinguish text positioning information by fully utilizing the text peak probability map. And (3) solving the segmentation mass center of the text prediction image by matching with a post-processing algorithm, obtaining a curved text center line and further obtaining a curved text boundary, and well processing long, multi-directional and curved texts.
2. And a coordinate learning mode of a text region in the past is abandoned, and a text mask and a text peak probability map are generated by the labeled data, so that the convolutional neural network can learn the mask and peak probabilities at the same time.
3. A multi-scale text detection network is designed.
4. And (3) using high-pass filtering on the probability map obtained by prediction to inhibit pixels with lower probability, further separating adjacent texts, eliminating a characteristic region with a small number of elements by using a connected domain method, and reducing prediction errors.
5. A center line is solved by a segmentation centroid method so as to obtain an accurate prediction result.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a schematic diagram of a convolutional neural network architecture design of a text localization method based on text peak probability density.
Fig. 2 is a schematic diagram of a training process of a text positioning method based on text peak probability density according to the present invention.
Fig. 3 is a schematic diagram of a prediction flow of a text positioning method based on text peak probability density according to the present invention.
FIG. 4 is a schematic diagram of the post-processing steps of the prediction information of the text positioning method based on the text peak probability density.
FIG. 5 is a schematic diagram of centroid calculation of a text localization method based on text peak probability density according to the present invention.
Fig. 6 shows a picture to be predicted according to an embodiment of the method of the present invention.
FIG. 7 is a text mask predicted by the embodiment of FIG. 6.
FIG. 8 is a graph of the text peak probability predicted by the embodiment of FIG. 6.
FIG. 9 is a diagram of the embodiment of FIG. 6 calculating the centroid of a text mask within a text block to obtain a text centerline.
Fig. 10 is an image obtained after the embodiment in fig. 6 is fitted.
Fig. 11 is a text prediction result image obtained by the embodiment in fig. 6.
FIG. 12 is a final image of the warped text of FIG. 6 after correction and segmentation.
Detailed Description
The technical scheme of the invention is further described by combining the attached drawings of the invention:
a text positioning method based on text peak probability density comprises the following four steps:
s1) image preprocessing: performing data preprocessing, and generating a required text mask and a text peak probability chart by disclosing the existing labeling information of a data set;
s2) data enhancement: performing data enhancement on the processed training set, wherein the data enhancement comprises at least one random transformation of brightness, contrast, hue, saturation, rotation, mirror image, cutting, scaling and the like;
s3) network training: transmitting the data into a convolutional neural network for training, and continuously optimizing network parameters through a combined loss function to obtain an optimal model;
s4) prediction process: firstly, loading the trained model parameters, transmitting the pictures to be predicted to the deep learning network to obtain a predicted text prediction result, and performing post-processing on prediction information to obtain an accurate text positioning result.
In the above technical solution, step S1) specifically includes:
s11) data preprocessing: setting a feature map of a text line center area to be composed of 0-255 gray levels, wherein the numerical value of the center line of the text line is larger as the text line is closer, the center of the text is whiter, the edge is blacker, and a gradual change effect is presented;
s12) text mask generation: generating a required text mask and a text line center mask by disclosing the existing labeling information of the data set, wherein the text mask generates a text rectangular mask by text block positioning information in the data set, generally 4 corner point coordinates of a text block;
s13) text peak probability map: and obtaining a text block central line according to the 4 corner point coordinates marked by the text, wherein the central line is the peak highest point with the value of 255, the central line gradually decreases from the central line to the edge of the text block, and the value of the edge is 0, so that a text peak probability graph is obtained.
In the above technical solution, the convolutional neural network basic module in step S3) includes a residual network module, a concat module, an upsampling module, and the like, and the network architecture is designed as shown in fig. 1, where Conv2d _ block is 1 layer of convolution, BN layer, and Relu layer, Res _ block is a residual module, upsample is an upsampling layer, Conv2d is a convolutional layer, h is an image height, and w is an image width. The network structure of the invention can obviously improve the multi-scale text detection and the distinguishing of adjacent text lines in the image.
In the above technical solution, step S3) is a schematic diagram of a network training process shown in fig. 2.
In the above technical solution, a schematic diagram of the prediction process in step S4) is shown in fig. 3.
In the above technical solution, the post-processing step of the prediction information in step S4) is as shown in fig. 4, and specifically includes:
s41) firstly, calculating the connected domain of the predicted text mask and the text peak probability graph and eliminating the smaller connected domain;
s42) calculating coordinates of four corner points of the predicted text box, dividing the text box, and calculating mass center coordinates of each divided block by using the prediction information; the principle is shown in fig. 5, wherein the small circle is a local centroid coordinate calculated by a text prediction mask of each segment;
s43) fitting the circle center and the radius of all centroids calculated by the text box by using a least square method, carrying out non-calculation on the longest edge of the predicted text box which is less than 2 times of the broadside, carrying out deviation on a centroid connecting line to obtain accurate text prediction information, carrying out polar coordinate transformation on a curved text box to finally obtain a horizontal text, and directly carrying out text cutting on the predicted box which does not accord with the segmentation condition.
The following describes in detail the implementation steps of the text localization method based on the text peak probability density according to the present invention with specific examples according to the attached drawings.
As shown in fig. 2, the image preprocessing, data enhancement and network training steps described in the method of the present invention are specifically described as follows:
firstly, carrying out data preprocessing on a public data set, decoding labeling information, generating a text mask and a text center mountain peak probability graph, respectively putting an original image and a preprocessed image into three folders, and manufacturing a training set and a test set;
performing data enhancement on the three images of the training set according to batches, amplifying the training set, increasing the data density and avoiding overfitting;
constructing a convolutional neural network;
establishing a combined Loss function, using an L1 distance to constrain the predicted text mask (tp) and text peak probability (tcp), and superposing the two parts of Loss to form a comprehensive Loss function Loss, wherein the Loss is tp _ Loss + lambda tcp _ Loss, and 5 is assigned through a large number of experiments lambda, so that the Loss function can be reduced most quickly;
and training the model, and storing the optimal model parameters obtained by each iterative training.
As shown in FIG. 3, the prediction process of the present invention is described as follows:
loading optimal model parameters;
predicting the model, carrying out scaling processing on the image to be predicted as shown in FIG. 6, and transmitting the image to be predicted into the model;
predicting to obtain a text mask and a text peak probability graph, as shown in fig. 7 and 8, eliminating a text region with a smaller probability through threshold filtering, and performing gaussian filtering on predicted information to enable a connection position to be smoother;
generating connected domains, eliminating smaller connected domains, performing digital marking on the connected domains, traversing the number of the connected domains, and calculating the minimum external moment of each connected domain to obtain 4 coordinates of the text box;
after a text circumscribed rectangle is obtained, carrying out blocking processing on the rectangle according to the perpendicular direction of the longest side;
calculating the centroid of the text mask in the text block, calculating and connecting each region divided in the rectangle to obtain the text center line, as shown in fig. 9;
obtaining a plurality of centroid points in the rectangular frame, fitting curvature and radius by using the plurality of centroid points by using a least square method, setting a threshold value, and when the radius is larger than the width or the height of the image, not needing to perform least square fitting on the text frame; the circle centers after fitting are marked by circular points, as shown in FIG. 10;
performing two-side displacement on the fitted text center line in the direction of the perpendicular line of the long side of the rectangle, wherein the distance is half of the width of the rectangle, and connecting the two ends to finish the detection of the bent text;
according to the curvature and the radius of the detected bent text, performing polar coordinate transformation on the bent text, and for the non-bent text, directly using the minimum external moment to perform segmentation;
the text prediction result and the result of the corrected and segmented warped text shown in fig. 11 are finally obtained, as shown in fig. 12.
The foregoing shows and describes the general principles, essential features, and advantages of the invention. It will be understood by those skilled in the art that the present invention is not limited to the embodiments described above, which are described in the specification and illustrated only to illustrate the principle of the present invention, but that various changes and modifications may be made therein without departing from the spirit and scope of the present invention, which fall within the scope of the invention as claimed. The scope of the invention is defined by the appended claims and equivalents thereof.

Claims (4)

1. A text positioning method based on text peak probability density is characterized by comprising the following four steps:
s1) image preprocessing: performing data preprocessing, and generating a required text mask and a text peak probability chart by disclosing the existing labeling information of a data set;
s2) data enhancement: performing data enhancement on the processed training set, wherein the data enhancement comprises at least one random transformation of brightness, contrast, hue, saturation, rotation, mirror image, cutting, scaling and the like;
s3) network training: transmitting the data into a convolutional neural network for training, and continuously optimizing network parameters through a combined loss function to obtain an optimal model;
s4) prediction process: firstly, loading the trained model parameters, transmitting the pictures to be predicted to the deep learning network to obtain a predicted text prediction result, and performing post-processing on prediction information to obtain an accurate text positioning result.
2. The method as claimed in claim 1, wherein the step S1) specifically includes:
s11) data preprocessing: setting a feature map of a text line center area to be composed of 0-255 gray levels, wherein the numerical value of the center line of the text line is larger as the text line is closer, the center of the text is whiter, the edge is blacker, and a gradual change effect is presented;
s12) text mask generation: generating a required text mask and a text line center mask by disclosing the existing labeling information of the data set, wherein the text mask generates a text rectangular mask through 4 corner point coordinates of text blocks in the data set;
s13) text peak probability map: and obtaining a center line of the text block according to the coordinates of the 4 corner points marked by the text, wherein the center line is the peak highest point, the value of the peak highest point is 255, the center line gradually decreases towards the edge of the text block, and the value of the edge is 0, so that a text peak probability map is obtained.
3. The method of claim 1, wherein the text positioning method based on the text peak probability density comprises:
the convolutional neural network basic module in the step S3) comprises a residual error network module, a concat module and an up-sampling module.
4. The method as claimed in claim 1, wherein the post-processing of the prediction information in step S4) is specifically:
s41) firstly, calculating the connected domain of the predicted text mask and the text peak probability graph and eliminating the smaller connected domain;
s42) calculating coordinates of four corner points of the predicted text box, dividing the text box, and calculating mass center coordinates of each divided block by using the prediction information;
s43) fitting the circle center and the radius of all centroids calculated by the text box by using a least square method, carrying out non-calculation on the longest edge of the predicted text box which is less than 2 times of the broadside, carrying out deviation on a centroid connecting line to obtain accurate text prediction information, carrying out polar coordinate transformation on a curved text box to finally obtain a horizontal text, and directly carrying out text cutting on the predicted box which does not accord with the segmentation condition.
CN201911237840.9A 2019-12-06 2019-12-06 Text prediction method based on text mountain probability density Active CN111274863B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911237840.9A CN111274863B (en) 2019-12-06 2019-12-06 Text prediction method based on text mountain probability density

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911237840.9A CN111274863B (en) 2019-12-06 2019-12-06 Text prediction method based on text mountain probability density

Publications (2)

Publication Number Publication Date
CN111274863A true CN111274863A (en) 2020-06-12
CN111274863B CN111274863B (en) 2023-07-07

Family

ID=71001506

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911237840.9A Active CN111274863B (en) 2019-12-06 2019-12-06 Text prediction method based on text mountain probability density

Country Status (1)

Country Link
CN (1) CN111274863B (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112446356A (en) * 2020-12-15 2021-03-05 西北工业大学 Method for detecting text with any shape in natural scene based on multiple polar coordinates
CN113034415A (en) * 2021-03-23 2021-06-25 哈尔滨市科佳通用机电股份有限公司 Method for amplifying small parts of railway locomotive image

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107862834A (en) * 2017-12-08 2018-03-30 梁金凤 A kind of mountain flood monitoring and warning system based on cloud framework
RU2016142229A (en) * 2016-10-26 2018-04-26 Федеральное государственное казенное военное образовательное учреждение высшего образования "Военный учебно-научный центр Военно-воздушных сил "Военно-воздушная академия имени профессора Н.Е. Жуковского и Ю.А. Гагарина" (г. Воронеж) Министерства обороны Российской Федерации METHOD FOR DETECTING AND LOCALIZING TEXT FORMS IN IMAGES
CN109147010A (en) * 2018-08-22 2019-01-04 广东工业大学 Band attribute Face image synthesis method, apparatus, system and readable storage medium storing program for executing
CN109801438A (en) * 2019-01-18 2019-05-24 创新奇智(南京)科技有限公司 A kind of intelligent sales counter based on recognition of face and interactive voice
CN109903358A (en) * 2019-01-23 2019-06-18 贵州师范大学 A kind of world natural heritage declares the method for drafting of graph

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
RU2016142229A (en) * 2016-10-26 2018-04-26 Федеральное государственное казенное военное образовательное учреждение высшего образования "Военный учебно-научный центр Военно-воздушных сил "Военно-воздушная академия имени профессора Н.Е. Жуковского и Ю.А. Гагарина" (г. Воронеж) Министерства обороны Российской Федерации METHOD FOR DETECTING AND LOCALIZING TEXT FORMS IN IMAGES
CN107862834A (en) * 2017-12-08 2018-03-30 梁金凤 A kind of mountain flood monitoring and warning system based on cloud framework
CN109147010A (en) * 2018-08-22 2019-01-04 广东工业大学 Band attribute Face image synthesis method, apparatus, system and readable storage medium storing program for executing
CN109801438A (en) * 2019-01-18 2019-05-24 创新奇智(南京)科技有限公司 A kind of intelligent sales counter based on recognition of face and interactive voice
CN109903358A (en) * 2019-01-23 2019-06-18 贵州师范大学 A kind of world natural heritage declares the method for drafting of graph

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
ZHU, YX, ET AL: "TextMountain:Accurate scene text detection via instance segmentation", 《PATTERN RECOGNITION》 *
ZHU, YX, ET AL: "TextMountain:Accurate scene text detection via instance segmentation", 《PATTERN RECOGNITION》, vol. 110, 28 February 2021 (2021-02-28) *
孙红星等: "文本区域字符颜色极性判断方法", 《东北大学学报(自然科学版)》 *
孙红星等: "文本区域字符颜色极性判断方法", 《东北大学学报(自然科学版)》, no. 03, 28 March 2007 (2007-03-28) *
岳颖等: "基于态度意义的语篇组织文本可视化", 《外语学刊》 *
岳颖等: "基于态度意义的语篇组织文本可视化", 《外语学刊》, no. 02, 30 April 2017 (2017-04-30) *

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112446356A (en) * 2020-12-15 2021-03-05 西北工业大学 Method for detecting text with any shape in natural scene based on multiple polar coordinates
CN112446356B (en) * 2020-12-15 2024-05-24 西北工业大学 Method for detecting text with arbitrary shape in natural scene based on multiple polar coordinates
CN113034415A (en) * 2021-03-23 2021-06-25 哈尔滨市科佳通用机电股份有限公司 Method for amplifying small parts of railway locomotive image
CN113034415B (en) * 2021-03-23 2021-09-14 哈尔滨市科佳通用机电股份有限公司 Method for amplifying small parts of railway locomotive image

Also Published As

Publication number Publication date
CN111274863B (en) 2023-07-07

Similar Documents

Publication Publication Date Title
CN111160352B (en) Workpiece metal surface character recognition method and system based on image segmentation
EP3309703B1 (en) Method and system for decoding qr code based on weighted average grey method
CN107273502B (en) Image geographic labeling method based on spatial cognitive learning
CN110766008A (en) Text detection method facing any direction and shape
CN110647829A (en) Bill text recognition method and system
CN112418216B (en) Text detection method in complex natural scene image
CN111340028A (en) Text positioning method and device, electronic equipment and storage medium
CN112990205B (en) Method and device for generating handwritten character sample, electronic equipment and storage medium
CN111259878A (en) Method and equipment for detecting text
CN111242024A (en) Method and system for recognizing legends and characters in drawings based on machine learning
CN114155527A (en) Scene text recognition method and device
CN111523622B (en) Method for simulating handwriting by mechanical arm based on characteristic image self-learning
CN110443235B (en) Intelligent paper test paper total score identification method and system
CN111626292B (en) Text recognition method of building indication mark based on deep learning technology
CN112446259A (en) Image processing method, device, terminal and computer readable storage medium
CN111680690A (en) Character recognition method and device
CN110598698A (en) Natural scene text detection method and system based on adaptive regional suggestion network
CN109741273A (en) A kind of mobile phone photograph low-quality images automatically process and methods of marking
CN111126266B (en) Text processing method, text processing system, equipment and medium
CN111274863A (en) Text prediction method based on text peak probability density
CN116597466A (en) Engineering drawing text detection and recognition method and system based on improved YOLOv5s
CN114758341A (en) Intelligent contract image identification and contract element extraction method and device
CN114155540B (en) Character recognition method, device, equipment and storage medium based on deep learning
CN116030472A (en) Text coordinate determining method and device
CN114187445A (en) Method and device for recognizing text in image, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant