CN112364726B - Part code-spraying character positioning method based on improved EAST - Google Patents
Part code-spraying character positioning method based on improved EAST Download PDFInfo
- Publication number
- CN112364726B CN112364726B CN202011163480.5A CN202011163480A CN112364726B CN 112364726 B CN112364726 B CN 112364726B CN 202011163480 A CN202011163480 A CN 202011163480A CN 112364726 B CN112364726 B CN 112364726B
- Authority
- CN
- China
- Prior art keywords
- candidate
- improved
- east
- area
- loss function
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 29
- 238000005507 spraying Methods 0.000 title abstract description 8
- 230000006870 function Effects 0.000 claims abstract description 31
- 238000012545 processing Methods 0.000 claims abstract description 13
- 238000012805 post-processing Methods 0.000 claims abstract description 5
- 230000008447 perception Effects 0.000 claims abstract description 4
- 238000012549 training Methods 0.000 claims description 17
- 230000035939 shock Effects 0.000 claims description 12
- 238000011176 pooling Methods 0.000 claims description 6
- 230000009467 reduction Effects 0.000 claims description 5
- 238000012163 sequencing technique Methods 0.000 claims description 2
- 239000007921 spray Substances 0.000 claims description 2
- 238000012216 screening Methods 0.000 claims 1
- 238000013135 deep learning Methods 0.000 abstract description 3
- 230000000694 effects Effects 0.000 description 10
- 238000001514 detection method Methods 0.000 description 9
- 230000008569 process Effects 0.000 description 6
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 2
- 238000010586 diagram Methods 0.000 description 2
- 238000005516 engineering process Methods 0.000 description 2
- 230000006872 improvement Effects 0.000 description 2
- 238000002372 labelling Methods 0.000 description 2
- 238000011160 research Methods 0.000 description 2
- 238000004891 communication Methods 0.000 description 1
- 238000013527 convolutional neural network Methods 0.000 description 1
- 230000010485 coping Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000004807 localization Effects 0.000 description 1
- 239000011159 matrix material Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000005070 sampling Methods 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000001629 suppression Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V10/00—Arrangements for image or video recognition or understanding
- G06V10/20—Image preprocessing
- G06V10/22—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition
- G06V10/225—Image preprocessing by selection of a specific region containing or referencing a pattern; Locating or processing of specific regions to guide the detection or recognition based on a marking or identifier characterising the area
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06V—IMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
- G06V30/00—Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
- G06V30/40—Document-oriented image-based pattern recognition
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Artificial Intelligence (AREA)
- Computational Linguistics (AREA)
- Molecular Biology (AREA)
- Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Biophysics (AREA)
- Multimedia (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Life Sciences & Earth Sciences (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Character Input (AREA)
- Image Analysis (AREA)
Abstract
The application provides a part code-spraying character positioning method based on improved EAST, which is based on the existing EAST network structure, improves the network structure, label generation, loss function and candidate area processing of the existing EAST network, obtains the improved network structure, improved label generation, improved loss function and candidate area post-processing, and uses the improved EAST network to position characters. The application optimizes the network structure, label generation, loss function and candidate area post-treatment based on the existing EAST algorithm, and the improved deep learning network can effectively enlarge the perception field of view, improve the recognition efficiency of long texts and realize the accurate positioning of part code-spraying characters.
Description
Technical Field
The invention relates to the technical field of character recognition, in particular to a part code-spraying character positioning method based on improved EAST.
Background
With the development of information science and technology, image processing and machine vision techniques have been widely used in the industry. Character text detection has been the most widely used machine vision technology, and its industrial application has been the focus and difficulty of research. The first step of character text detection is to locate the character text region. The EAST algorithm is one of excellent natural scene text localization algorithms, and has excellent performance in a plurality of public data sets. But limited by the receptive field size, training process setting and the like, the EAST algorithm is not ideal in character positioning, and the detection effect on long texts is required to be improved.
Therefore, there is a need for an end-to-end character recognition method that has little dependency on character segmentation and high communication.
Disclosure of Invention
In view of this, the present invention provides a method for improved EAST-based part spray symbol positioning, characterized by: the method is based on the existing EAST network structure, improves the network structure, label generation, loss function and candidate area processing of the existing EAST network, obtains improved network structure, improved label generation, improved loss function and candidate area post-processing, and locates characters by using the improved EAST network;
The network structure for improving the existing EAST network specifically comprises:
constructing a VGG16 model, wherein the VGG16 model comprises 5 convolution layers and 5 pooling layers, the convolution kernel size is 3 multiplied by 3, and the step length is 1;
Improving a VGG16 model, and replacing convolution operation of the last stage in the VGG16 network structure with mixed cavity convolution HDC with the cavity coefficient of [1,2,5 ];
the improved label generation specifically comprises the following steps: narrowing a quadrangle formed by a character area of the target training image in a mode of distinguishing long and short sides;
The improved loss function specifically includes: replacing a balance cross entropy loss function in an EAST algorithm by using a dice loss function;
The processing of the improved candidate region specifically includes: a pre-ordering based local sense NMS algorithm is used to order all candidate boxes according to their upper left corner coordinates.
Further, the improved label generation specifically includes the following steps:
narrowing the quadrangle formed by the character area of the target training image in a mode of treating the long side and the short side respectively;
the short side of the quadrangle is reduced by 0.3 times on the basis of the length of the short side of the original label remark;
the long side of the quadrangle is reduced by shrunk _rate times on the basis of the length of the long side of the original label remark:
wherein shrunk _rate represents a reduction multiple of the long side, length s represents the length of the short side of the quadrangle, and length l represents the length of the long side of the quadrangle.
Further, the dice loss function is determined by the following method:
L=Ls+λgLg=Ls+λg(LIoU+λθLθ) (2)
where L represents the modified dice loss function, L s represents the score map loss, L g represents the geometric loss, λ g represents the coefficient of geometric loss, L IoU represents the overlap area loss, λ θ represents the coefficient of angular loss, and L θ represents the angular loss.
Further, the processing of the improved candidate region specifically includes a pre-ordered based local sense NMS algorithm, and the pre-ordered based local sense NMS algorithm specifically includes the following steps:
s1: initializing a set box, an overlapping degree threshold lambda and a shock threshold omega of all candidate rectangular frames;
s2: extracting the left upper corner coordinates (x 1,y1) of all rectangular frames in the boxes, respectively arranging according to the sequence from small to large of x 1 and y 1, respectively calculating and comparing the shock times of x 1 and y 1 according to omega, and arranging the boxes according to the sequence from small to large of the shock times;
S3: let s be the empty set, p be the empty set, for g e boxes, where g represents an element in the set boxes of all candidate rectangular boxes, i.e. a candidate rectangular box, and sequentially execute, in order from small to large, the number of times of shock of all rectangular boxes in boxes: if p is not empty and the overlapping area ratio of p and g is larger than lambda, p is a rectangular frame obtained by combining p and g; if p is not null and the overlapping area ratio of p and g is not greater than lambda, then p is put into s; if p is null, p=g;
s4: if p is not null, then put p into s;
s5: and outputting a set S, wherein the S represents the filtered rectangular frame set.
Further, the processing of improving the candidate region further includes candidate box merging, where the candidate box merging is determined by the following method:
S(a)=ηgS(g)+ηpS(p) (3)
ai=ηgS(g)gi+ηpS(p)pi (4)
Wherein a is a combined rectangular frame, g and p represent candidate rectangular frames meeting an overlapping condition, i represents parameter indexes in the rectangular frames, i=1, 2,3 …,8 coordinates are shared by 4 vertexes, 1 angle value is obtained, S represents a score value of the rectangular frame, eta g represents an area coefficient of the candidate rectangular frame g, eta p represents an area coefficient of the candidate rectangular frame p;
the area coefficients eta g and eta p are set by the following method:
Whether the area of the candidate rectangular frame g is larger than that of the candidate rectangular frame p is judged, if yes, η g=1.1,ηp =1, and if not, η g=1,ηp =1.1.
The beneficial technical effects of the application are as follows: the application optimizes the network structure, label generation, loss function and candidate area post-treatment based on the existing EAST algorithm, and the improved deep learning network can effectively enlarge the perception field of view, improve the recognition efficiency of long texts and realize the accurate positioning of part code-spraying characters.
Drawings
The invention is further described below with reference to the accompanying drawings and examples:
Fig. 1 is a schematic diagram of an improved EAST algorithm network architecture proposed by the present invention.
Fig. 2 is a schematic diagram of the label generation process of the present invention.
FIG. 3 is a graph comparing the detection effect of the present invention with the detection effect of the original algorithm.
Detailed Description
The invention is further described below with reference to the accompanying drawings of the specification:
The invention provides a part code-spraying character positioning method based on improved EAST, which is characterized in that: the method is based on the existing EAST network structure, improves the network structure, label generation, loss function and candidate area processing of the existing EAST network, obtains improved network structure, improved label generation, improved loss function and candidate area post-processing, and locates characters by using the improved EAST network;
The network structure for improving the existing EAST network specifically comprises:
constructing a VGG16 model, wherein the VGG16 model comprises 5 convolution layers and 5 pooling layers, the convolution kernel size is 3 multiplied by 3, and the step length is 1;
Improving a VGG16 model, and replacing convolution operation of the last stage in the VGG16 network structure with mixed cavity convolution HDC with the cavity coefficient of [1,2,5 ];
the improved label generation specifically comprises the following steps: narrowing a quadrangle formed by a character area of the target training image in a mode of distinguishing long and short sides;
The improved loss function specifically includes: replacing a balance cross entropy loss function in an EAST algorithm by using a dice loss function;
The processing of the improved candidate region specifically includes: a pre-ordering based local sense NMS algorithm is used to order all candidate boxes according to their upper left corner coordinates.
In the convolutional neural network, the size of the receptive field is determined by parameters such as a convolutional kernel, a convolutional step length and the like. In the standard VGG16 model, the receptive field expansion is achieved by the superposition and pooling operations of the convolution layers. However, the method for expanding the receptive field is limited by the number of convolution layers, so that the effect is limited, the size of the convolution kernel is expanded, parameters are excessive, and the receptive field is increased through a pooling layer, so that information is lost. These factors limit the size of the VGG16 network receptive field. The cavity convolution is a method that the size of a convolution kernel is unchanged, but a plurality of cavities are added to enlarge the whole convolution matrix and expand a convolution area, so that parameters are not increased and a receptive field is expanded, but grid effect and the phenomenon that a small target cannot be detected can occur. However, a hybrid cavity convolution scheme is proposed, and the above problem can be avoided as long as the following three requirements are met by the convolution coefficient setting in the process of building the neural network, and according to the thought, the convolution operation (Conv_5) of the final stage of VGG16 is replaced, the original standard convolution for 3 times is replaced by the hybrid cavity convolution (HDC) with the cavity coefficient of [1,2,5], so that the original 7×7 receptive field of the part is enlarged to 17×17, and the VGG16 network with larger receptive field is obtained. The new network structure is shown in fig. 1, and the dark blue part is the improvement part of the VGG16 network, wherein Maxpooling represents the maximum pooling operation, up Sampling represents the upsampling operation, and Concat represents the feature map stitching.
According to the technical scheme, the network structure, the label generation, the loss function and the candidate region post-processing are optimized on the basis of the existing EAST algorithm, the improved deep learning network can effectively enlarge the perception field of view, the recognition efficiency of long texts is improved, and the accurate positioning of part code-spraying characters is realized.
In this embodiment, the improved label generation specifically includes the following steps:
narrowing the quadrangle formed by the character area of the target training image in a mode of treating the long side and the short side respectively;
the short side of the quadrangle is reduced by 0.3 times on the basis of the length of the short side of the original label remark;
the long side of the quadrangle is reduced by shrunk _rate times on the basis of the length of the long side of the original label remark:
wherein shrunk _rate represents a reduction multiple of the long side, length s represents the length of the short side of the quadrangle, and length l represents the length of the long side of the quadrangle.
In the network training process, character area labels of all training images are quadrilateral, and coordinates of four points are expressed in a format of (x 1,y1,x2,y2,x3y3,x4,y4). x and y represent the abscissa and ordinate, respectively, of the point. The subscripts "1", "2", "3", "4" denote points at the upper left, upper right, lower right and lower left corners, respectively, i.e., are labeled in a clockwise order. In EAST algorithm training, a form RBOX channels { d 1,d2,d3d4, θ } is adopted to generate a training tag, wherein d i represents distances from points in the training tag to four edges of the tag (text boxes in FIG. 1), and θ represents a rotation angle (angle in FIG. 1). There are another 1 channel to record the likelihood score (score map in fig. 1) that each point in the image is a character point. In order to reduce interference caused by label labeling errors, a score map (score map) in an EAST algorithm is reduced by 0.3 times on the basis of the minimum circumscribed rectangle of the original label labeling. The overall label generation process is shown in fig. 2. The method for reducing the label rectangle can reduce the interference brought by the character edge area and improve the accuracy of character positioning. However, as is apparent from fig. 2, since the long-side cardinality of the character area is large, the 0.3-fold reduction results in a large-area character area not entering the score map. Such training is very bad for long character positioning. Aiming at the problem, the label training of the EAST algorithm is improved, the short sides of the label rectangle are still reduced by 0.3 times by adopting a mode of differentiating the long sides, and the reduction multiple of the long sides is the ratio of the length of the short sides of the rectangle to the length of the long sides of the rectangle.
In this embodiment, the dice loss function is determined by the following method:
L=Ls+λgLg=Ls+λg(LIoU+λθLθ) (2)
where L represents the modified dice loss function, L s represents the score map loss, L g represents the geometric loss, λ g represents the coefficient of geometric loss, L IoU represents the overlap area loss, λ θ represents the coefficient of angular loss, and L θ represents the angular loss.
For neural network training, the loss function is an optimization target of the whole network training, and the quality of the loss function setting directly influences the training result of the network and the detection effect of final parameters. For the FCN network proposed by the present invention, the loss function L is mainly composed of two parts of score map loss L s and geometric loss L g, where geometric loss L g is composed of overlapping area loss L IoU and angle loss L θ. For the score map loss L s, it is essentially a two-class loss function, i.e., the character region is a positive sample, the other regions are negative samples, and the positive and negative samples are very unbalanced, and the area of the character region is much smaller than the other regions. The EAST algorithm adopts a balanced cross entropy loss function to solve the sample imbalance problem, and research in recent years finds that the dice has more excellent performance in coping with the sample imbalance problem. Therefore, the invention adopts the race loss to express the score graph loss, and can better cope with the sample unbalance problem compared with the EAST algorithm.
In this embodiment, the processing of the improved candidate region specifically includes a pre-ordered based local-awareness NMS algorithm, where the pre-ordered based local-awareness NMS algorithm specifically includes the following steps:
s1: initializing a set box, an overlapping degree threshold lambda and a shock threshold omega of all candidate rectangular frames;
s2: extracting the left upper corner coordinates (x 1,y1) of all rectangular frames in the boxes, respectively arranging according to the sequence from small to large of x 1 and y 1, respectively calculating and comparing the shock times of x 1 and y 1 according to omega, and arranging the boxes according to the sequence from small to large of the shock times;
S3: let S be the empty set, p be the empty set, for g e boxes, where g represents an element in the set boxes of all candidate rectangular boxes, i.e. a candidate rectangular box, and sequentially execute, in order from small to large, the number of times of shock of all rectangular boxes in boxes: if p is not empty and the overlapping area ratio of p and g is larger than lambda, p is a rectangular frame obtained by combining p and g; if p is not null and the overlapping area ratio of p and g is not greater than lambda, then p is put into S; if p is null, p=g;
s5: and outputting a set S, wherein the S represents the filtered rectangular frame set.
Typically, when the network is used for character region detection, there may be hundreds or even thousands of candidate frames extracted by FCN. Most of them overlap each other, so combining this huge number of candidate boxes is the last step of the whole character region locating algorithm. Previous subsequent block merging often employed non-maximal suppression, NMS, algorithms, but the time complexity of such algorithms was O (n 2), which is very time consuming. For this case, the EAST algorithm employs a locally aware NMS algorithm. The algorithm only compares two adjacent candidate areas, and if the overlapping degree is larger than a threshold value, the two candidate areas are combined, and if the overlapping degree is smaller than the threshold value, the two candidate areas are reserved. The time complexity depends on the arrangement of the candidate regions, and is preferably O (n) (i.e., the candidate frames that should be merged are all arranged together), and is most preferably O (n 2) (i.e., the candidate frames that do not overlap with each other are alternately arranged). Meanwhile, two candidate frames meeting the overlap condition are not discarded one by one as in the standard NMS, but are combined according to the scoring condition.
It can be seen that the temporal complexity of the locally perceived NMS algorithm depends on the random candidate arrangement. Aiming at the whole problem, the invention improves the local sensing NMS algorithm, provides the local sensing NMS algorithm based on pre-sequencing, sequences all candidate frames according to the values of the left upper corner coordinates of the candidate frames, and can arrange all the candidate frames which are possibly combined together, thereby achieving the effect that the time complexity is O (n) as much as possible. Meanwhile, the arrangement of the candidate frames which are possibly combined is beneficial to improving the combining precision, so that the final text positioning effect is improved.
In this embodiment, the process of improving the candidate region further includes candidate box merging, where the candidate box merging is determined by the following method:
S(a)=ηgS(g)+ηpS(p) (3)
ai=ηgS(g)gi+ηpS(p)pi (4)
Wherein a is a combined rectangular frame, g and p represent candidate rectangular frames meeting an overlapping condition, i represents parameter indexes in the rectangular frames, i=1, 2,3 …,8 coordinates are shared by 4 vertexes, 1 angle value is obtained, S represents a score value of the rectangular frame, eta g represents an area coefficient of the candidate rectangular frame g, eta p represents an area coefficient of the candidate rectangular frame p;
the area coefficients eta g and eta p are set by the following method:
Whether the area of the candidate rectangular frame g is larger than that of the candidate rectangular frame p is judged, if yes, η g=1.1,ηp =1, and if not, η g=1,ηp =1.1.
In addition, in the original local sense NMS algorithm, only the scoring condition is considered when the candidate rectangular frames are combined, the area information of the candidate frames is also important for long texts, and the possibility that the candidate frames with larger areas cover the whole text area is higher. Thus, in the candidate frame merging herein, η, which introduces the larger of the area coefficients η, g and p, is set to 1.05, and the other is 1.
The detection effect before and after the improvement of the comparison algorithm is implemented according to the improved character positioning algorithm based on EAST, as shown in figure 3. It is obvious that the original EAST algorithm has defects of long text detection, the rectangular box on the rightmost side of the first row of characters does not cover all character areas, and the last 1 of the second row of characters is only covered by half. The optimized algorithm of the invention is used for detecting long text, so that the whole character area is well covered, and the condition that the head and the tail of the character area are missed does not occur.
Finally, it is noted that the above embodiments are only for illustrating the technical solution of the present invention and not for limiting the same, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications and equivalents may be made thereto without departing from the spirit and scope of the technical solution of the present invention, which is intended to be covered by the scope of the claims of the present invention.
Claims (1)
1. A method for improving the positioning of part spray characters based on EAST, which is characterized in that: the method is based on the existing EAST network structure, improves the network structure, label generation, loss function and candidate area processing of the existing EAST network, obtains improved network structure, improved label generation, improved loss function and candidate area post-processing, and locates characters by using the improved EAST network;
The network structure for improving the existing EAST network specifically comprises:
constructing a VGG16 model, wherein the VGG16 model comprises 5 convolution layers and 5 pooling layers, the convolution kernel size is 3 multiplied by 3, and the step length is 1;
Improving a VGG16 model, and replacing convolution operation of the last stage in the VGG16 network structure with mixed cavity convolution HDC with the cavity coefficient of [1,2,5 ];
the improved label generation specifically comprises the following steps: narrowing a quadrangle formed by a character area of the target training image in a mode of distinguishing long and short sides;
The improved loss function specifically includes: replacing a balance cross entropy loss function in an EAST algorithm by using a dice loss function;
The processing of the improved candidate region specifically includes: a local perception NMS algorithm based on pre-sequencing is adopted, and all candidate frames are sequenced according to the values of the upper left corner coordinates of the candidate frames;
The improved label generation specifically comprises the following steps:
narrowing the quadrangle formed by the character area of the target training image in a mode of treating the long side and the short side respectively;
the short side of the quadrangle is reduced by 0.3 times on the basis of the length of the short side of the original label remark;
the long side of the quadrangle is reduced by shrunk _rate times on the basis of the length of the long side of the original label remark:
Wherein shrunk _rate represents a reduction multiple of the long side, length s represents the length of the short side of the quadrangle, and length l represents the length of the long side of the quadrangle;
the dice loss function is determined by the following method:
L=Ls+λgLg=Ls+λg(LIoU+λθLθ) (2)
Wherein L represents the modified dice loss function, L s represents the score map loss, L g represents the geometric loss, λ g represents the coefficient of geometric loss, L IoU represents the overlap area loss, λ θ represents the coefficient of angular loss, and L θ represents the angular loss;
the processing of the improved candidate region specifically comprises a pre-ordered based local awareness NMS algorithm, and the pre-ordered based local awareness NMS algorithm specifically comprises the following steps:
s1: initializing a set box, an overlapping degree threshold lambda and a shock threshold omega of all candidate rectangular frames;
S2: extracting the left upper corner coordinates (x 1,y1) of all rectangular frames in the boxes, respectively arranging according to the sequence from small to large of x 1 and y 1, respectively calculating and comparing the shock times of x 1 and y 1 according to omega, and arranging the boxes according to the sequence from small to large of the shock times;
S3: let S be the empty set, p be the empty set, for g e boxes, where g represents an element in the set boxes of all candidate rectangular boxes, i.e. a candidate rectangular box, and sequentially execute, in order from small to large, the number of times of shock of all rectangular boxes in boxes: if p is not empty and the overlapping area ratio of p and g is larger than lambda, p is a rectangular frame obtained by combining p and g; if p is not null and the overlapping area ratio of p and g is not greater than lambda, then p is put into S; if p is null, p=g;
s4: if p is not null, then put p into S;
S5: outputting a set S, wherein the S represents a rectangular frame set after screening;
the processing of the improved candidate region further includes candidate block merging, the candidate block merging determined by:
S(a)=ηgS(g)+ηpS(p) (3)
ai=ηgS(g)gi+ηpS(p)pi (4)
Wherein a is a combined rectangular frame, g and p represent candidate rectangular frames meeting an overlapping condition, i represents parameter indexes in the rectangular frames, i=1, 2,3 …, 8 coordinates are shared by 4 vertexes, 1 angle value is obtained, S represents a score value of the rectangular frame, eta g represents an area coefficient of the candidate rectangular frame g, eta p represents an area coefficient of the candidate rectangular frame p;
the area coefficients eta g and eta p are set by the following method:
Whether the area of the candidate rectangular frame g is larger than that of the candidate rectangular frame p is judged, if yes, η g=1.1,ηp =1, and if not, η g=1,ηp =1.1.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011163480.5A CN112364726B (en) | 2020-10-27 | 2020-10-27 | Part code-spraying character positioning method based on improved EAST |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202011163480.5A CN112364726B (en) | 2020-10-27 | 2020-10-27 | Part code-spraying character positioning method based on improved EAST |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112364726A CN112364726A (en) | 2021-02-12 |
CN112364726B true CN112364726B (en) | 2024-06-04 |
Family
ID=74510665
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202011163480.5A Active CN112364726B (en) | 2020-10-27 | 2020-10-27 | Part code-spraying character positioning method based on improved EAST |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112364726B (en) |
Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1997017988A1 (en) * | 1995-11-13 | 1997-05-22 | Fibrogen, Inc. | Type ix collagen and chimeras |
US7143153B1 (en) * | 2000-11-09 | 2006-11-28 | Ciena Corporation | Internal network device dynamic health monitoring |
CN107133616A (en) * | 2017-04-02 | 2017-09-05 | 南京汇川图像视觉技术有限公司 | A kind of non-division character locating and recognition methods based on deep learning |
CN110287960A (en) * | 2019-07-02 | 2019-09-27 | 中国科学院信息工程研究所 | The detection recognition method of curve text in natural scene image |
CN111310760A (en) * | 2020-02-13 | 2020-06-19 | 辽宁师范大学 | Method for detecting onychomycosis characters by combining local prior characteristics and depth convolution characteristics |
WO2020164281A1 (en) * | 2019-02-13 | 2020-08-20 | 平安科技(深圳)有限公司 | Form parsing method based on character location and recognition, and medium and computer device |
CN111583180A (en) * | 2020-04-03 | 2020-08-25 | 苏宁云计算有限公司 | Image tampering identification method and device, computer equipment and storage medium |
Family Cites Families (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US9780909B2 (en) * | 2014-12-29 | 2017-10-03 | Juniper Networks, Inc. | Network topology optimization with feasible optical paths |
US20180359029A1 (en) * | 2017-01-18 | 2018-12-13 | Ciena Corporation | Distributed measurements and analysis in networks |
-
2020
- 2020-10-27 CN CN202011163480.5A patent/CN112364726B/en active Active
Patent Citations (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO1997017988A1 (en) * | 1995-11-13 | 1997-05-22 | Fibrogen, Inc. | Type ix collagen and chimeras |
US7143153B1 (en) * | 2000-11-09 | 2006-11-28 | Ciena Corporation | Internal network device dynamic health monitoring |
CN107133616A (en) * | 2017-04-02 | 2017-09-05 | 南京汇川图像视觉技术有限公司 | A kind of non-division character locating and recognition methods based on deep learning |
WO2020164281A1 (en) * | 2019-02-13 | 2020-08-20 | 平安科技(深圳)有限公司 | Form parsing method based on character location and recognition, and medium and computer device |
CN110287960A (en) * | 2019-07-02 | 2019-09-27 | 中国科学院信息工程研究所 | The detection recognition method of curve text in natural scene image |
CN111310760A (en) * | 2020-02-13 | 2020-06-19 | 辽宁师范大学 | Method for detecting onychomycosis characters by combining local prior characteristics and depth convolution characteristics |
CN111583180A (en) * | 2020-04-03 | 2020-08-25 | 苏宁云计算有限公司 | Image tampering identification method and device, computer equipment and storage medium |
Non-Patent Citations (3)
Title |
---|
LEW,S.Environmental controls on the abundance of methanotrophs and methanogens in peat bog lakes.SCIENCE OF THE TOTAL ENVIRONMENT.2018,1201-1211. * |
基于复杂网络理论的Java开源系统演化分析;唐倩文;陈良育;;计算机科学(第08期);173-180 * |
基于深度神经网络的汽车刹车片喷码检测识别技术研究;李代杨;万方硕士学位论文;1-7 * |
Also Published As
Publication number | Publication date |
---|---|
CN112364726A (en) | 2021-02-12 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111640125B (en) | Aerial photography graph building detection and segmentation method and device based on Mask R-CNN | |
CN111666938B (en) | Two-place double-license-plate detection and identification method and system based on deep learning | |
CN108334881B (en) | License plate recognition method based on deep learning | |
CN110598690B (en) | End-to-end optical character detection and recognition method and system | |
CN110458172A (en) | A kind of Weakly supervised image, semantic dividing method based on region contrast detection | |
CN102360494B (en) | Interactive image segmentation method for multiple foreground targets | |
CN113435240B (en) | End-to-end form detection and structure identification method and system | |
CN106778757A (en) | Scene text detection method based on text conspicuousness | |
CN110334709B (en) | License plate detection method based on end-to-end multi-task deep learning | |
CN107977592B (en) | Image text detection method and system, user terminal and server | |
CN110008900A (en) | A kind of visible remote sensing image candidate target extracting method by region to target | |
CN113888505B (en) | Natural scene text detection method based on semantic segmentation | |
CN110852327A (en) | Image processing method, image processing device, electronic equipment and storage medium | |
CN111339874A (en) | Single-stage face segmentation method | |
CN112270679A (en) | Image segmentation method for convex polygon outline combining concave points and concave edges | |
CN114511627A (en) | Target fruit positioning and dividing method and system | |
CN114693966A (en) | Target detection method based on deep learning | |
CN112364726B (en) | Part code-spraying character positioning method based on improved EAST | |
CN107832732A (en) | Method for detecting lane lines based on ternary tree traversal | |
Lin et al. | Fast vehicle detector for autonomous driving | |
CN105844641A (en) | Adaptive threshold segmentation method in dynamic environment | |
CN112861860B (en) | Text detection method in natural scene based on upper and lower boundary extraction | |
CN115376022A (en) | Application of small target detection algorithm based on neural network in unmanned aerial vehicle aerial photography | |
CN114724175A (en) | Pedestrian image detection network, detection method, training method, electronic device, and medium | |
CN114220082A (en) | Lane line identification method and device and computer readable storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |