CN110472632B - Character segmentation method and device based on character features and computer storage medium - Google Patents

Character segmentation method and device based on character features and computer storage medium Download PDF

Info

Publication number
CN110472632B
CN110472632B CN201910702665.XA CN201910702665A CN110472632B CN 110472632 B CN110472632 B CN 110472632B CN 201910702665 A CN201910702665 A CN 201910702665A CN 110472632 B CN110472632 B CN 110472632B
Authority
CN
China
Prior art keywords
character
image
features
segmentation
feature
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910702665.XA
Other languages
Chinese (zh)
Other versions
CN110472632A (en
Inventor
刘晋
高珍喻
李云辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Maritime University
Original Assignee
Shanghai Maritime University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Maritime University filed Critical Shanghai Maritime University
Priority to CN201910702665.XA priority Critical patent/CN110472632B/en
Publication of CN110472632A publication Critical patent/CN110472632A/en
Application granted granted Critical
Publication of CN110472632B publication Critical patent/CN110472632B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/26Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion
    • G06V10/267Segmentation of patterns in the image field; Cutting or merging of image elements to establish the pattern region, e.g. clustering-based techniques; Detection of occlusion by performing operations on regions, e.g. growing, shrinking or watersheds

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Multimedia (AREA)
  • Character Input (AREA)
  • Character Discrimination (AREA)
  • Image Analysis (AREA)

Abstract

The invention provides a character segmentation method based on character features, which is applied to the technical field of image processing and comprises the following steps: acquiring an image to be processed; carrying out binarization processing on the image to be processed to obtain a binarized image; extracting the features of the binary image by adopting a basic feature extraction network; performing feature extraction on the forms of the characters according to the extracted features to acquire first features, and performing feature extraction on the number of the characters to acquire second features; fusing the first features and the second features by adopting a semantic segmentation network so as to generate a semantic segmentation graph; and determining the segmentation position of the character according to the semantic segmentation graph. In addition, the invention also discloses a character segmentation device based on character characteristics and a computer storage medium.

Description

Character segmentation method and device based on character features and computer storage medium
Technical Field
The present invention relates to the field of image processing technologies, and in particular, to a character segmentation method and apparatus based on character features, and a computer storage medium.
Background
Character segmentation is the basis and precondition of image character information extraction, and reasonable and correct segmentation must be performed on characters.
For character segmentation, methods proposed earlier are projection-based segmentation and Connected domain-based segmentation (Vertical projection, Connected domain). The two methods do not aim at the condition of conglutination characters when being proposed, so that each character cannot be well divided aiming at the conditions of unclear, missing and conglutination of more strokes in the actual scene picture. Meanwhile, due to the morphological characteristics of the upper-lower structure and the left-right structure of the Chinese character, a plurality of characters are divided into a plurality of parts. Although morphological characteristics of characters are considered in the Water drop and Clustering-based segmentation method (Water drop and Clustering), local characteristics of the characters are only optimized, the adhered strokes are simply divided in a gravity-like or Clustering mode, improvement is achieved to a certain extent, and the result of division of the complicated adhered strokes is not ideal.
Therefore, an effective character cutting method is lacking in the prior art.
Disclosure of Invention
In view of the above-mentioned shortcomings of the prior art, it is an object of the present invention to provide a character segmentation method and apparatus based on character features, which can solve the shortcomings of the conventional method by adding character information such as word width and word number as the basis of segmentation. Meanwhile, the convolutional neural network is adopted to extract character information, and the full convolutional network is applied to carry out semantic segmentation, so that the problems of stroke missing and stroke adhesion character segmentation can be effectively solved.
To achieve the above and other related objects, the present invention provides a character segmentation method based on character features, the method comprising:
acquiring an image to be processed;
carrying out binarization processing on the image to be processed to obtain a binarized image;
extracting the features of the binary image by adopting a basic feature extraction network;
performing feature extraction on the forms of the characters according to the extracted features to acquire first features, and performing feature extraction on the number of the characters to acquire second features;
fusing the first characteristic and the second characteristic by adopting a semantic segmentation network to generate a semantic segmentation graph;
and determining the segmentation position of the character according to the semantic segmentation graph.
In one implementation manner, the step of performing binarization processing on the image to be processed to obtain a binarized image includes:
generating a gray level histogram according to the image to be processed;
acquiring a foreground peak and a background peak corresponding to the gray level histogram;
acquiring gray values corresponding to the troughs of the foreground peak and the background peak;
the acquired gray value is taken as a binarization threshold value.
In one implementation manner, the step of performing feature extraction on the binarized image by using a basic feature extraction network includes:
and carrying out feature extraction on the binary image by adopting a Convolutional Neural Network (CNN).
In one implementation, the step of fusing the first feature and the second feature by using a semantic segmentation network to generate a semantic segmentation map includes:
receiving the first feature and the second feature, and restoring the size of the data through deconvolution and upsampling operations until the size of the image to be processed is reached;
and carrying out Softmax classification on the restored image, and taking the classified image as a semantic segmentation map.
In one implementation, the step of performing Softmax classification on the restored image and using the classified image as a semantic segmentation map includes:
classifying each pixel point in the restored image;
acquiring the probability of each pixel point corresponding to the character class;
and performing segmentation according to the obtained probability.
In one implementation, the step of training the convolutional neural network CNN includes:
constructing a training data set, wherein the data set comprises 30000 pictures with adhesion conditions prepared by adopting 3755 Chinese characters specified in GB2312 primary national standard, the size of the pictures is 512 x 512, the size of characters in the pictures is between [70px and 80px ], and the number of the characters is between [2 and 5 ];
randomly adding white noise and interference textures to the data set to obtain an enhanced image;
and performing Convolutional Neural Network (CNN) training based on the enhanced image.
The invention also discloses a character segmentation device based on character characteristics, which comprises a processor and a memory connected with the processor through a communication bus; wherein, the first and the second end of the pipe are connected with each other,
the memory is used for storing a character segmentation program based on character features;
the processor is configured to execute the character feature-based character segmentation program to implement any one of the character feature-based character segmentation steps.
And a computer storage medium storing one or more programs executable by one or more processors to cause the one or more processors to perform any of the character feature based character segmentation steps is also disclosed.
As described above, embodiments of the present invention provide a character segmentation method, apparatus and computer-readable storage medium based on character features, which can solve the shortcomings of the conventional methods by adding character information such as word width and number of words as the basis of segmentation. Meanwhile, the convolutional neural network is adopted to extract character information, and the full convolutional network is applied to carry out semantic segmentation, so that the problems of stroke missing and stroke adhesion character segmentation can be effectively solved.
Drawings
Fig. 1 is a schematic flow chart of a character segmentation method based on character features according to an embodiment of the present invention.
Fig. 2 is a schematic diagram of an application of a character segmentation method based on character features according to an embodiment of the present invention.
Fig. 3 is a schematic diagram of an application of a character segmentation method based on character features according to an embodiment of the present invention.
Fig. 4 is a schematic diagram of an application of a character segmentation method based on character features according to an embodiment of the present invention.
Detailed Description
The embodiments of the present invention are described below with reference to specific embodiments, and other advantages and effects of the present invention will be easily understood by those skilled in the art from the disclosure of the present specification. The invention is capable of other and different embodiments and of being practiced or of being carried out in various ways, and its several details are capable of modification in various respects, all without departing from the spirit and scope of the present invention.
Please refer to fig. 1-4. It should be noted that the drawings provided in the present embodiment are only for illustrating the basic idea of the present invention, and the components related to the present invention are only shown in the drawings rather than drawn according to the number, shape and size of the components in actual implementation, and the type, quantity and proportion of the components in actual implementation may be changed freely, and the layout of the components may be more complicated.
As shown in fig. 1, an embodiment of the present invention provides a character segmentation method based on character features, where the method includes:
and S101, acquiring an image to be processed.
The image to be processed is an image including text characters and requiring segmentation processing.
And S102, carrying out binarization processing on the image to be processed to obtain a binarized image.
On a gray level histogram generated by the image, two peaks respectively represent a target image and a background image, and the trough position between the two peaks is selected as a binarization threshold value T.
Figure BDA0002151266550000041
Wherein f (x, y) is the gray scale value of the gray scale image, and g (x, y) is the gray scale image after binarization.
Specifically, other methods for generating the binarized image by selecting the threshold value may be used, such as a P parameter method, a maximum entropy threshold method, a maximum inter-class variance method, and the like.
And S103, extracting the features of the binary image by adopting a basic feature extraction network.
Specifically, the basic feature extraction network is a Convolutional Neural Network (CNN), and the trained CNN is required to be used for feature billiards.
In the CNN training process, firstly, a training set picture meeting requirements is adopted, and 30000 pictures with adhesion conditions are formulated by adopting 3755 Chinese characters specified in GB2312 first-level national standard in a data set. The size of the picture adopted in the experiment is 512 x 512, the size of the characters in the picture is between [70px and 80px ], and the number of the characters is between [2 and 5 ]. White noise and interference textures are artificially and randomly added for data enhancement when the character picture is constructed. And generating a character information label, namely character width and character number information, by performing semantic marking manually or semi-manually according to the generated sample image. The key parameter configuration table in the specific process of constructing the data set is shown in fig. 2. The data set portion picture pattern is shown, for example, in fig. 3.
In one embodiment of the present invention, the preprocessed image is subjected to a plurality of convolution and pooling operations, and the basic structure of each convolution operation unit includes 3 convolution layers, 3 active layers and 1 pooling layer. Through a plurality of convolution and pooling processes, multi-dimensional data is converted into one-dimensional data through the flat layer, so that feature fusion is facilitated, forward propagation data volume is reduced through the Dropout layer, and a final output result is obtained through the full connection layer.
Or converting multidimensional data into one-dimensional data through a flat layer through multiple convolution and pooling processing so as to facilitate feature fusion, performing concatenate operation through a fusion layer to fuse character morphological features and character number features, reducing forward propagation data amount through a Drapout layer, and obtaining a final output result through a full connection layer.
For layer I neurons in a neural network, the output of the neuron is represented as y l . For the ith neuron in the l +1 layer neural network, the method is used
Figure BDA0002151266550000051
Express their corresponding weights by
Figure BDA0002151266550000052
Indicating its corresponding bias. The common two-dimensional convolutional neural network calculation formula is as follows:
Figure BDA0002151266550000053
Figure BDA0002151266550000054
wherein the content of the first and second substances,
Figure BDA0002151266550000055
represents the calculated value of the ith neuron in the l +1 layer neural network,
Figure BDA0002151266550000056
represents the output result corresponding to the neuron after the calculated value is processed by the activation function f ().
The neural network calculation formula adopting the Dropout mechanism is as follows:
r i l =Bernoulli(p) i
Figure BDA0002151266550000061
Figure BDA0002151266550000062
Figure BDA0002151266550000063
wherein, Bernoulli (p) i Means for randomly generating a function for the ith neuron in the l-th neural network with a probability p
Figure BDA0002151266550000064
The 0, 1 vectors represented. Then processing the output of the ith neuron in the original l-th layer neural network through the generated vector
Figure BDA0002151266550000065
For results use
Figure BDA0002151266550000066
To indicate.
After the neural network training adopting the Dropout mechanism is completed, in a stage of performing prediction processing by using the neural network, a calculation formula of a prediction result of an ith neuron in the l +1 layer neural network is as follows:
Figure BDA0002151266550000067
Figure BDA0002151266550000068
s104, extracting the features of the characters according to the extracted features to obtain first features, and extracting the features of the number of the characters to obtain second features.
In one embodiment of the invention, the image to be semantically segmented is preprocessed, 1) the sub-network of character morphological feature extraction: and (3) marking a feature diagram obtained by passing the result of the preprocessing through 1 convolutional layer and 1 pooling layer as Fa 1_1, a feature diagram obtained by passing Fa 1_1 through 3 convolutional layers and 1 pooling layer as Fa 1_2, a feature diagram obtained by passing Fa 1_2 through 3 convolutional layers and 1 pooling layer as Fa 1_3, a feature diagram obtained by passing Fa 1_3 through 1 convolutional layer and 1 tiled layer as Fa 1_4, and an output result of the character morphological feature extraction network obtained by passing Fa 1_4 through 4 dense layers and 3 dropout layers, wherein the segmentation result is shown in FIG. 4.
2) Character number feature extraction sub-network: and marking a feature diagram obtained by passing the result of the preprocessing through 1 convolutional layer and 1 pooling layer as Fa 2_1, marking a feature diagram obtained by passing Fa 2_1 through 3 convolutional layers and 1 pooling layer as Fa 2_2, marking a feature diagram obtained by passing Fa 2_2 through 3 convolutional layers and 1 pooling layer as Fa 2_3, marking a feature diagram obtained by passing Fa 2_3 through 1 convolutional layer and 1 tiling layer as Fa 2_4, processing Fa 2_4 and Fa 1_4 through one fusion layer, and passing the fusion result through 4 dense layers and 3 drop layers to obtain an output result of the character number feature extraction network.
In a specific embodiment of the present invention, the basic feature extraction network and the character information feature extraction network are used as the basic network, the results output by the two networks are processed through 1 fusion layer, the fusion result is processed through 3 deconvolution and upsampling operations, each operation unit includes 3 deconvolution layers and one upsampling layer, and finally, the output result of the network is obtained through 4 deconvolution layers.
The neural network training described above uses adam (adaptive motion estimation) as an optimizer. The loss weight set to the neural network output of the character information extraction section is 0.5, and the loss weight set to the neural network output of the semantic segmentation section is 1.0. In addition, the neural network output of the character information extraction part adopts a loss calculation method of 'geographic _ crosssensory' and the neural network output of the semantic segmentation part adopts a loss calculation method of 'binary _ crosssensory'.
The initial learning rate learning _ rate in the Adam parameter is set to 0.0001(1e-4), the exponential decay rate of the first order moment estimate beta _1 is set to 0.9, and the exponential decay rate of the second order moment estimate beta _2 is set to 0.999. Also to prevent division by zero in the calculation, epsilon is set to 1e-08 while decade is set to 0.0.
And S105, fusing the first characteristic and the second characteristic by adopting a semantic segmentation network to generate a semantic segmentation graph.
Multiple deconvolution and upsampling operations are then employed to reduce the data size to the same size as the input image. And classifying each pixel point by using the obtained data and adopting a softmax classification function, wherein the value of each pixel point represents the probability of the pixel point being classified into the character class, and generating a semantic segmentation graph.
softmax function:
Figure BDA0002151266550000071
the K-dimensional real vector of the obtained image data is mapped to a K-dimensional real vector σ (z) whose components are 0 to 1.
And S106, determining the segmentation position of the character according to the semantic segmentation graph.
The semantic segmentation graph can be subjected to graying and binarization processing, the processing result is subjected to opening and closing operation, a minimum closure area is found, a maximum rectangle is found in the area, and the coordinates of the rectangle are obtained. And determining the segmentation position of the character according to the obtained word width and the number of the characters in S140 and S150.
A convolutional neural network: convolutional Neural Networks (CNNs) were proposed inspired by the visual neuroscience. The structure mainly comprises a convolution layer and a pooling layer. The earliest convolutional neural network model was the LeNet-5 model proposed by LeCun Y in 1998. In the model, the original image is converted into several feature maps by a convolution layer and a sampling layer. These feature maps map low-level local region features onto higher-level global features through convolution operations using the action of convolution kernels. Since then, there are continuous methods based on convolutional neural network infrastructure improvement to get good results in ImageNet competitions.
Based on projection segmentation: and scanning the image transversely and longitudinally, counting the number of pixel black points in two directions, wherein the columns and the rows without the pixel black points are the segmentation positions.
Based on clustering segmentation: according to the characteristics of the gray scale, the color, the texture, the shape and the like of the image, the image is divided into a plurality of non-overlapping regions, the characteristics are similar in the same region, and obvious difference exists between different regions. I.e. a clustering algorithm is used to group pixels to achieve image segmentation.
The full convolution neural network (FCN) is a neural network structure proposed by Long Jonathan et al in 2015, and mainly aims at the segmentation task of images. The method converts a full connection layer in a traditional Convolutional Neural Network (CNN) into a convolutional layer, and classifies an image at a pixel level in an end-to-end mode, so that the problem of semantic level image segmentation is solved.
In the field of engineering application, the task of character recognition has a bottleneck that the recognition accuracy rate is difficult to improve. This is because it is difficult to obtain accurate segmentation for characters in an image. Some of the problems that are difficult to deal with are, for example, that a single character is often cut apart for a left-right structure of a chinese character, that a single character is cut into a plurality of parts for a missing stroke of a chinese character, and that a plurality of characters are cut into a character area for a more common adhesion of strokes of a chinese character. Therefore, the method of adding character information such as word width and number of words as a basis for division can solve the disadvantages of the conventional method. Meanwhile, the convolutional neural network is adopted to extract character information, and the full convolutional network is applied to carry out semantic segmentation, so that the problems of stroke missing and stroke adhesion character segmentation can be effectively solved.
The invention also discloses a character segmentation device based on character characteristics, which comprises a processor and a memory connected with the processor through a communication bus; wherein the content of the first and second substances,
the memory is used for storing a character segmentation program based on character features;
the processor is configured to execute the character feature-based character segmentation program to implement any one of the character feature-based character segmentation steps.
And a computer storage medium storing one or more programs executable by one or more processors to cause the one or more processors to perform any of the character feature based character segmentation steps is also disclosed.
The foregoing embodiments are merely illustrative of the principles and utilities of the present invention and are not intended to limit the invention. Those skilled in the art can modify or change the above-described embodiments without departing from the spirit and scope of the present invention. Accordingly, it is intended that all equivalent modifications or changes which can be made by those skilled in the art without departing from the spirit and technical spirit of the present invention be covered by the claims of the present invention.

Claims (8)

1. A character segmentation method based on character features is characterized by comprising the following steps:
acquiring an image to be processed;
carrying out binarization processing on the image to be processed to obtain a binarized image;
extracting the features of the binary image by adopting a basic feature extraction network;
performing feature extraction on the forms of the characters according to the extracted features to acquire first features, and performing feature extraction on the number of the characters to acquire second features;
fusing the first characteristic and the second characteristic by adopting a semantic segmentation network to generate a semantic segmentation graph;
and determining the segmentation position of the character according to the semantic segmentation graph.
2. The character feature-based character segmentation method according to claim 1, wherein the step of performing binarization processing on the image to be processed to obtain a binarized image comprises:
generating a gray level histogram according to the image to be processed;
acquiring a foreground peak and a background peak corresponding to the gray level histogram;
acquiring gray values corresponding to the troughs of the foreground peak and the background peak;
the acquired gray value is taken as a binarization threshold value.
3. The character feature-based character segmentation method according to claim 1 or 2, wherein the step of performing feature extraction on the binarized image by using a basic feature extraction network comprises:
and carrying out feature extraction on the binary image by adopting a Convolutional Neural Network (CNN).
4. The character segmentation method based on character features according to claim 3, wherein the step of fusing the first features and the second features by using a semantic segmentation network to generate a semantic segmentation graph comprises:
receiving the first feature and the second feature, and restoring the size of the data through deconvolution and upsampling operations until the size of the image to be processed is reached;
and carrying out Softmax classification on the restored image, and taking the classified image as a semantic segmentation map.
5. The character segmentation method based on character features according to claim 4, wherein the step of performing Softmax classification on the restored image and taking the classified image as a semantic segmentation map comprises the steps of:
classifying each pixel point in the restored image;
acquiring the probability of each pixel point corresponding to the character class;
and performing segmentation according to the obtained probability.
6. The character feature-based character segmentation method according to claim 5, wherein the step of training the convolutional neural network CNN comprises:
constructing a training data set, wherein the data set comprises 30000 pictures with adhesion conditions prepared by adopting 3755 Chinese characters specified in GB2312 primary national standard, the size of the pictures is 512 x 512, the size of characters in the pictures is between [70px and 80px ], and the number of the characters is between [2 and 5 ];
randomly adding white noise and interference textures to the data set to obtain an enhanced image;
and performing Convolutional Neural Network (CNN) training based on the enhanced image.
7. An apparatus for character segmentation based on character features, the apparatus comprising a processor, and a memory coupled to the processor via a communication bus; wherein the content of the first and second substances,
the memory is used for storing a character segmentation program based on character features;
the processor, configured to execute the character feature-based character segmentation program to implement the steps of the character feature-based character segmentation method according to any one of claims 1 to 6.
8. A computer storage medium storing one or more programs, the one or more programs being executable by one or more processors to cause the one or more processors to perform the steps of the character feature based character segmentation method as claimed in any one of claims 1 to 6.
CN201910702665.XA 2019-07-31 2019-07-31 Character segmentation method and device based on character features and computer storage medium Active CN110472632B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910702665.XA CN110472632B (en) 2019-07-31 2019-07-31 Character segmentation method and device based on character features and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910702665.XA CN110472632B (en) 2019-07-31 2019-07-31 Character segmentation method and device based on character features and computer storage medium

Publications (2)

Publication Number Publication Date
CN110472632A CN110472632A (en) 2019-11-19
CN110472632B true CN110472632B (en) 2022-09-30

Family

ID=68509348

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910702665.XA Active CN110472632B (en) 2019-07-31 2019-07-31 Character segmentation method and device based on character features and computer storage medium

Country Status (1)

Country Link
CN (1) CN110472632B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111626283B (en) * 2020-05-20 2022-12-13 北京字节跳动网络技术有限公司 Character extraction method and device and electronic equipment
CN111723815B (en) * 2020-06-23 2023-06-30 中国工商银行股份有限公司 Model training method, image processing device, computer system and medium
CN112270370B (en) * 2020-11-06 2023-06-02 北京环境特性研究所 Vehicle apparent damage assessment method

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650721A (en) * 2016-12-28 2017-05-10 吴晓军 Industrial character identification method based on convolution neural network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5875637B2 (en) * 2013-12-19 2016-03-02 キヤノン株式会社 Image processing apparatus and image processing method

Patent Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106650721A (en) * 2016-12-28 2017-05-10 吴晓军 Industrial character identification method based on convolution neural network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
一种通用的基于图像分割的验证码识别方法;白培瑞等;《山东科技大学学报(自然科学版)》;20180509(第03期);全文 *

Also Published As

Publication number Publication date
CN110472632A (en) 2019-11-19

Similar Documents

Publication Publication Date Title
CN110119728B (en) Remote sensing image cloud detection method based on multi-scale fusion semantic segmentation network
CN110176027B (en) Video target tracking method, device, equipment and storage medium
CN108509978B (en) Multi-class target detection method and model based on CNN (CNN) multi-level feature fusion
Zhao et al. Pixelated semantic colorization
Wu et al. Semantic segmentation of high-resolution remote sensing images using fully convolutional network with adaptive threshold
Zhuo et al. Cloud classification of ground-based images using texture–structure features
CN103049763B (en) Context-constraint-based target identification method
CN110472632B (en) Character segmentation method and device based on character features and computer storage medium
CN110598600A (en) Remote sensing image cloud detection method based on UNET neural network
CN111652317B (en) Super-parameter image segmentation method based on Bayes deep learning
CN108510499B (en) Image threshold segmentation method and device based on fuzzy set and Otsu
CN109800817B (en) Image classification method based on fusion semantic neural network
CN109886271B (en) Image accurate segmentation method integrating deep learning network and improving edge detection
CN113705371B (en) Water visual scene segmentation method and device
CN110399895A (en) The method and apparatus of image recognition
CN110136162B (en) Unmanned aerial vehicle visual angle remote sensing target tracking method and device
CN110991257B (en) Polarized SAR oil spill detection method based on feature fusion and SVM
Mukherjee et al. Enhancement of image resolution by binarization
CN110827312A (en) Learning method based on cooperative visual attention neural network
US11367206B2 (en) Edge-guided ranking loss for monocular depth prediction
CN111507337A (en) License plate recognition method based on hybrid neural network
CN114511576A (en) Image segmentation method and system for scale self-adaptive feature enhanced deep neural network
CN115713632A (en) Feature extraction method and device based on multi-scale attention mechanism
CN112149526A (en) Lane line detection method and system based on long-distance information fusion
CN115033721A (en) Image retrieval method based on big data

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information
CB03 Change of inventor or designer information

Inventor after: Liu Jin

Inventor after: Gao Zhenyu

Inventor after: Li Yunhui

Inventor before: Liu Jin

Inventor before: Gao Zhenyu

GR01 Patent grant
GR01 Patent grant