CN108399408A - A kind of deformed characters antidote based on deep space converting network - Google Patents

A kind of deformed characters antidote based on deep space converting network Download PDF

Info

Publication number
CN108399408A
CN108399408A CN201810181595.3A CN201810181595A CN108399408A CN 108399408 A CN108399408 A CN 108399408A CN 201810181595 A CN201810181595 A CN 201810181595A CN 108399408 A CN108399408 A CN 108399408A
Authority
CN
China
Prior art keywords
character
network
space converting
deformed characters
transformation parameter
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810181595.3A
Other languages
Chinese (zh)
Inventor
李子衿
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN201810181595.3A priority Critical patent/CN108399408A/en
Publication of CN108399408A publication Critical patent/CN108399408A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition
    • G06V30/14Image acquisition
    • G06V30/146Aligning or centring of the image pick-up or image-field
    • G06V30/1475Inclination or skew detection or correction of characters or of image to be recognised
    • G06V30/1478Inclination or skew detection or correction of characters or of image to be recognised of characters or characters lines
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)

Abstract

Optical character identification OCR is the classical problem in Pattern recognition and image processing field, and current artificial intelligence field basic research direction of interest.The technology is widely used to the industries such as bank, traffic, customs, public security, logistics, greatly reduces human cost, improves the working efficiency of the mankind.In order to carry out automatic straightening to the character of a variety of affine transformations such as translation, inclination, rotation, dimensional variation, the invention discloses a kind of deformed characters antidotes based on deep space converting network, by the introducing of spatial alternation network, the correction to deformed characters is realized;Wherein the spatial alternation network is character space converting network, specifically includes character locating network, mesh generator and character pixels sampler.The present invention can be suitably used for the various deformation situation of character picture, be reduced to the constraints of input picture, can be suitable for gray level image and RGB image simultaneously, and have stronger noiseproof feature.

Description

A kind of deformed characters antidote based on deep space converting network
Technical field
The invention belongs to character picture processing technology fields, and in particular to a kind of deformation based on deep space converting network Character antidote.
Background technology
Optical character identification (OCR, optical character recognition) is Pattern recognition and image processing neck The classical problem in domain, and current artificial intelligence field basic research direction of interest.The technology be widely used to bank, The industries such as traffic, customs, public security, logistics, greatly reduce human cost, improve the working efficiency of the mankind.But in reality In the character recognition application scenarios on border, due to shooting angle and distance etc., character picture more or less will present out deformation, Such as translation, inclination, rotation, dimensional variation, such deformation can influence the identification accuracy of OCR to a certain extent.Therefore, right It is to promote the important channel of OCR recognition accuracies that deformed characters, which carry out correction,.Deformed characters correction generally passes through image procossing etc. Relevant technological means is realized so that computerized algorithm can preferably be identified character, to further promote character Recognition accuracy.
The existing deformed characters correcting technology based on image procossing mainly has following three kinds:
(1) the character correction algorithm based on boundary rectangle.The depth-width ratio that the program passes through calculating character minimum enclosed rectangle Determine the degree of rotation of character picture, and according to the correspondence of depth-width ratio and character rotation angle come calculating character image need The angle to be corrected.The program is only capable of the rotational correction for solving the problems, such as to accord with image in same scale size condition lower word, and can not Suitable for other deformations of character picture, such as character scaling, inclination, translation.
(2) the character correcting technology based on parallel separation augmented vector approach.The program is asked using low-rank matrix Topic models character correction, and parallel separation method is combined with method of Lagrange multipliers, realizes the extensive of gray level image Multiple correction.The program is designed for gray level image, and is difficult to be applicable in the RGB image being widely present in life.
(3) the character antidote based on Hough transformation.The program is being inclined after detecting angle by using Hough transformation Projection localization is carried out to character on inclined-plane, and corresponding rotational correction is carried out one by one to the character of segmentation.The limitation of this method It is similar to scheme (1), it is only capable of the rotational deformation of processing character, and other deformations can not be suitable for.
Spatial alternation network was suggested to solve input picture in depth convolutional neural networks in 2016 and locally becomes for the first time Adaptive problem is changed, to achieve the purpose that improve classification accuracy.Spatial alternation network is one kind in deep neural network On the basis of improved after can to image local spatial alternation carry out adaptive learning depth network, it can utilize sit The affine transformation of mark matrix goes to realize that the classical of a series of images deforms, such as translation, scaling, rotation, inclination and other geometry Transformation.
The invention discloses a kind of deformed characters antidotes based on deep space converting network, are become using deep space Switching network automatic returning goes out the affine transformation parameter of deformed characters image, to carry out recovery correction to deformed characters.This programme Consider the various affine transformations of character, such as rotation, translation, scale, inclination, therefore can be suitably used for the more of character picture Kind deformation.Due to scheme using deep space converting network go carry out transformation parameter automatic returning so that scheme is to defeated Entering the constraints of image reduces, and can be suitable for gray level image and RGB image simultaneously, and have stronger noiseproof feature.
Invention content
The purpose of the present invention:In order to the character of a variety of affine transformations such as translation, inclination, rotation, dimensional variation into Row automatic straightening, the invention discloses a kind of deformed characters antidotes based on deep space converting network.
Technical scheme of the present invention:
A kind of deformed characters antidote based on deep space converting network, which is characterized in that pass through spatial alternation net The correction to deformed characters is realized in the introducing of network;Wherein the spatial alternation network is character space converting network, specific to wrap Include character locating network, mesh generator and character pixels sampler.
The introducing by spatial alternation network realizes the correction to deformed characters, specifically includes following steps:
The first step, character image data pretreatment:Affine transformation character sample is generated by character picture training set, including Translation, rotation, scaling, inclination conditions, and record corresponding affine transformation parameter, and by the deformed characters sample of generation and former character Sample matches, so as to the training that exercises supervision;
Second step designs character space converting network according to spatial alternation Principles of Network;
Third walks, the training of character space converting network:Utilize the paired sample and paired sample handled well in the first step The affine transformation parameter of middle deformed characters exercises supervision training to the character space converting network of design;
4th step, affine transformation parameter return:Affine transformation parameter is carried out using character locating network to deformed characters to return Return;
5th step, sampling grid generate:Mesh generator calculates output image lattice coordinate by affine transformation parameter θ Point corresponds to mesh coordinate position in the input image;
6th step, character pixels interpolation sampling:According to pair of the input grid being calculated in the 5th step and output grid Coordinate relationship is answered, carries out the interpolation sampling of character pixels in the input image, to determine the tool for exporting each grid in grid Volumetric pixel value completes the straightening of character.
Mnist may be used in character picture training set described in step 1.
The character locating network is designed according to character picture size difficulty or ease situation to be solved, specifically includes convolution Layer, pond layer, nonlinear activation elementary layer, full articulamentum and recurrence layer.
The supervised training of affine transformation parameter described in step 3 is by calculating net regression value and true tag value Averaged Square Error of Multivariate simultaneously carries out reversed gradient propagation and realizes;
The supervised training of paired sample described in step 3 is by calculating former character sample and character sample after correction Average pixel value mean square error simultaneously carries out reversed gradient propagation and realizes;
In supervised training described in step 3, training loss function consists of two parts, i.e. paired sample loss and change Parameter loss is changed, mathematical formulae indicates as follows:
Loss=MSE (Iin, Iout)+MSE(θevl, θgt)
Wherein IinAnd IoutThe correcting image of the deformed characters image and network calculations output of input, θ are indicated respectivelyevlWith θgtIndicate that the affine transformation parameter and true transformation parameter that deep space converting network returns, MSE indicate average square respectively Error;Under the action of the loss function, the parameter value of Optimized model is propagated by reversed gradient so that model reaches more satisfactory State.
Affine transformation parameter described in step 4 returns way of realization:
Regard the convolutional layer of character locating network, pond layer, non-linear unit layer as one group of feature extraction unit, then deforms Character picture will pass through feature extraction unit for several times successively, using recurrence layer is input to after several layers of full articulamentum, to generate 6 transformation parameter θ of affine transformation, can be indicated with following formula:
θ=Floc(Cin)
Character pixels interpolation sampling specific method described in step 6 is:
The bilinearity character pixels interpolation sampling of following form is carried out in the input image:
Wherein H and W indicates the height and width of input character picture respectively,Indicate input character picture at coordinate points (n, m) Pixel value,Indicate the pixel value of i-th of coordinate points of output character image;Particularly, when the input net that required solution obtains Lattice pointIt has been more than the section of (- 1,1), then has meaned that some coordinate points for exporting grid do not have corresponding input coordinate Point is mapped, and Boundary filling strategy nearby can be taken.
Beneficial effects of the present invention:The deformed characters antidote based on deep space converting network designed by the present invention It goes to carry out regression estimates to the affine transformation parameter of deformed characters using deep space converting network, effectively covers character figure The various deformations situation such as translation, inclination, scale, rotation of picture, the character correction solution for comparing traditional have widely suitable The property used.The deep neural network can allow grayscale character image or RGB character pictures to be inputted simultaneously, be based on depth convolution The character locating network of network makes have stronger noiseproof feature.In addition, the program is not necessarily to the feature of engineer's character picture Extraction module, but deep space converting network is allowed to go to excavate the deformation state of image automatically, and the parameter of its deformation is assessed, it is real The deformed characters correction of " end-to-end " pattern is showed, this significantly reduces the cumbersome step that engineer's algorithm goes extraction transformation parameter Suddenly.Finally, the program can go to carry out the adjustment of network structure according to the complexity of task, this also improves character correction The Generalization Capability of technology.
Description of the drawings:
Fig. 1 is the flow chart of the deformed characters antidote based on deep space converting network.
Wherein, dotted arrow part indicates the loss function during model training, the supervised training of implementation model.Solid line Arrows show data flow direction.
Fig. 2 is the correction idiographic flow schematic diagram of one embodiment of the invention.
Specific implementation mode
Shown in reference to the accompanying drawings of the specification, the specific implementation mode of the present invention, in Fig. 1, dotted arrow part table are introduced Loss function in representation model training process, the supervised training of implementation model.Solid arrow indicates data flow direction.
The specific design and correction implementation steps of the present invention is completely as follows:The first step, character image data pretreatment:Pass through Character picture training set generates affine transformation character sample, including translation, rotation, scaling, inclination conditions, and records corresponding affine Transformation parameter, and the deformed characters sample of generation and former character sample are matched, so as to the training that exercises supervision.Wherein character picture The selection of training set is selected according to the target character classification of quasi- correction, and the character of such as quasi- correction belongs to numeric type, then can select Use Mnist as character picture training set.
Second step designs character space converting network according to the principle of spatial alternation network, specifically includes character locating net Network, mesh generator and character pixels sampler.Character locating network is according to character picture size difficulty or ease situation to be solved Design, including convolutional layer, pond layer, nonlinear activation elementary layer, full articulamentum and recurrence layer.
Third walks, the training of character space converting network:Utilize the paired sample and paired sample handled well in the first step The affine transformation parameter of middle deformed characters exercises supervision training to the character space converting network of design.Wherein, affine transformation is joined Several supervised trainings is by calculating the Averaged Square Error of Multivariate of net regression value and true tag value and by carry out reversed gradient propagation It realizes;The supervised training of the paired sample is equal by the average pixel value for calculating former character sample and character sample after correction Square error simultaneously carries out reversed gradient propagation and realizes;Training loss function consists of two parts, i.e. paired sample loss and transformation Parameter is lost, and mathematical formulae indicates as follows:
Loss=MSE (Iin, Iout)+MSE(θevl, θgt)
Wherein IinAnd IoutThe correcting image of the deformed characters image and network calculations output of input, θ are indicated respectivelyevlWith θgtIndicate that the affine transformation parameter and true transformation parameter that deep space converting network returns, MSE indicate average square respectively Error;Under the action of the loss function, the parameter value of Optimized model is propagated by reversed gradient so that model reaches more satisfactory State.
4th step, affine transformation parameter return:Affine transformation parameter is carried out using character locating network to deformed characters to return Return;For example, the convolutional layer of character locating network, pond layer, non-linear unit layer can be regarded as one group of feature extraction unit, then Deformed characters image will pass through feature extraction unit for several times successively, using being input to recurrence layer after several layers of full articulamentum, thus 6 transformation parameter θ for generating affine transformation, can be indicated with following formula:
θ=Floc(Cin)
5th step, sampling grid generate:Mesh generator calculates output image lattice coordinate by affine transformation parameter θ Point corresponds to mesh coordinate position in the input image;In specific implementation process, in order to carry out dimension normalization, first output is schemed In the linear normalizing of mesh coordinate of picture to (- 1,1) range, for each mesh point of output imageUtilize first The parameter θ returned in step solves the mesh point coordinate value of input picture
The grid point coordinates of these input pictures is also to be normalized in (- 1,1) range, and the sampling that will be walked as third Grid.
6th step, character pixels interpolation sampling:According to pair of the input grid being calculated in the 5th step and output grid Coordinate relationship is answered, carries out the interpolation sampling of character pixels in the input image, to determine the tool for exporting each grid in grid Volumetric pixel value, for example, the bilinearity character pixels interpolation sampling of following form can be carried out in the input image:
Wherein H and W indicates the height and width of input character picture respectively,Indicate input character picture at coordinate points (n, m) Pixel value,Indicate the pixel value of i-th of coordinate points of output character image;Particularly, when the input net that required solution obtains Lattice pointIt has been more than the section of (- 1,1), then has meaned that some coordinate points for exporting grid do not have corresponding input coordinate Point is mapped, and can take Boundary filling strategy nearby at this time.
The specific embodiment application of technical solution of the present invention is as follows:
Assuming that word " is removed " in the deformation that input character picture is a width 80*80 resolution ratio, which first passes around character Network is positioned, returns out transformation parameter θ values by the feedforward of depth convolutional network and fully-connected network is propagated, value such as Fig. 2 Shown in lower right-most portion matrix.Then, mesh generator is according to this θ value and output mesh coordinate, the anti-sampling for releasing input picture Grid.Finally, character pixels sampler is according to the sampling grid generated, using the interpolation samplings method such as bilinear interpolation defeated The enterprising row interpolation sampling of deformed characters image entered, ultimately generates the character picture " removing " after correction as shown in the lower left corners Fig. 2.

Claims (9)

1. a kind of deformed characters antidote based on deep space converting network, which is characterized in that pass through spatial alternation network Introducing, realize correction to deformed characters;Wherein the spatial alternation network is character space converting network, is specifically included Character locating network, mesh generator and character pixels sampler.
2. the deformed characters antidote according to claim 1 based on deep space converting network, which is characterized in that described logical The introducing of spatial alternation network is crossed, the correction to deformed characters is realized, specifically includes following steps:
The first step, character image data pretreatment:Affine transformation character sample is generated by character picture training set, including flat Move, rotation, scaling, inclination conditions, and record corresponding affine transformation parameter, and by the deformed characters sample of generation and former character sample This pairing, so as to the training that exercises supervision;
Second step designs character space converting network according to spatial alternation Principles of Network;
Third walks, the training of character space converting network:Using becoming in the paired sample and paired sample handled well in the first step The affine transformation parameter of shape character exercises supervision training to the character space converting network of design;
4th step, affine transformation parameter return:Affine transformation parameter recurrence is carried out to deformed characters using character locating network;
5th step, sampling grid generate:Mesh generator calculates output image lattice coordinate points by affine transformation parameter θ and exists Mesh coordinate position is corresponded in input picture;
6th step, character pixels interpolation sampling:According to the input grid being calculated in the 5th step and the corresponding seat for exporting grid Mark relationship carries out the interpolation sampling of character pixels in the input image, to determine the specific picture for exporting each grid in grid Element value, completes the straightening of character.
3. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step Mnist may be used in character picture training set described in rapid one.
4. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that institute The character locating network stated is designed according to character picture size difficulty or ease situation to be solved, specifically include convolutional layer, pond layer, Nonlinear activation elementary layer, full articulamentum and recurrence layer.
5. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step 3 Described in affine transformation parameter supervised training by calculating the Averaged Square Error of Multivariate of net regression value and true tag value simultaneously It carries out reversed gradient propagation and realizes.
6. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step 3 Described in paired sample supervised training it is square by the average pixel value of character sample after calculating former character sample and correcting Error simultaneously carries out reversed gradient propagation and realizes.
7. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step 3 Described in supervised training in, training loss function consist of two parts, i.e., paired sample loss and transformation parameter loss, mathematics Formula indicates as follows:
Loss=MSE (Iin, Iout)+MSE(θevlgt)
Wherein IinAnd IoutThe correcting image of the deformed characters image and network calculations output of input, θ are indicated respectivelyevlAnd θgtRespectively Indicate that the affine transformation parameter and true transformation parameter that deep space converting network returns, MSE indicate Averaged Square Error of Multivariate; Under the action of the loss function, the parameter value of Optimized model is propagated by reversed gradient so that model reaches comparatively ideal state.
8. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step 4 Described in affine transformation parameter return way of realization be:
Regard the convolutional layer of character locating network, pond layer, non-linear unit layer as one group of feature extraction unit, then deformed characters Image will pass through feature extraction unit for several times successively, affine to generate using being input to recurrence layer after several layers of full articulamentum 6 transformation parameter θ of transformation, can be indicated with following formula:
θ=Floc(Cin)。
9. the deformed characters antidote according to claim 2 based on deep space converting network, which is characterized in that step 6 Described in character pixels interpolation sampling specific method be:
The bilinearity character pixels interpolation sampling of following form is carried out in the input image:
Wherein H and W indicates the height and width of input character picture respectively,Picture of the expression input character picture in coordinate points (n, m) Element value,Indicate the pixel value of i-th of coordinate points of output character image;Particularly, when the input mesh point that required solution obtainsIt has been more than the section of (- 1,1), then has meaned that some coordinate points for exporting grid do not have corresponding input coordinate to click through Row mapping, can take Boundary filling strategy nearby.
CN201810181595.3A 2018-03-06 2018-03-06 A kind of deformed characters antidote based on deep space converting network Pending CN108399408A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810181595.3A CN108399408A (en) 2018-03-06 2018-03-06 A kind of deformed characters antidote based on deep space converting network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810181595.3A CN108399408A (en) 2018-03-06 2018-03-06 A kind of deformed characters antidote based on deep space converting network

Publications (1)

Publication Number Publication Date
CN108399408A true CN108399408A (en) 2018-08-14

Family

ID=63091899

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810181595.3A Pending CN108399408A (en) 2018-03-06 2018-03-06 A kind of deformed characters antidote based on deep space converting network

Country Status (1)

Country Link
CN (1) CN108399408A (en)

Cited By (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109447911A (en) * 2018-10-18 2019-03-08 百度在线网络技术(北京)有限公司 Method, apparatus, storage medium and the terminal device of image restoration
CN109801234A (en) * 2018-12-28 2019-05-24 南京美乐威电子科技有限公司 Geometric image correction method and device
CN109829848A (en) * 2019-01-17 2019-05-31 柳州康云互联科技有限公司 A kind of system and method for Image space transformation neural network based in internet detection
CN109886264A (en) * 2019-01-08 2019-06-14 深圳禾思众成科技有限公司 A kind of character detecting method, equipment and computer readable storage medium
CN110321894A (en) * 2019-04-23 2019-10-11 浙江工业大学 A kind of library book method for rapidly positioning based on deep learning OCR
CN110443782A (en) * 2019-07-03 2019-11-12 杭州深睿博联科技有限公司 Chest x-ray piece model alignment schemes and device, storage medium
CN110929784A (en) * 2019-11-21 2020-03-27 上海智臻智能网络科技股份有限公司 Method for recognizing characters in picture, computer equipment and storage medium
CN110956133A (en) * 2019-11-29 2020-04-03 上海眼控科技股份有限公司 Training method of single character text normalization model, text recognition method and device
CN111046859A (en) * 2018-10-11 2020-04-21 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN111402156A (en) * 2020-03-11 2020-07-10 腾讯科技(深圳)有限公司 Restoration method and device for smear image, storage medium and terminal equipment
CN111539184A (en) * 2020-04-29 2020-08-14 上海眼控科技股份有限公司 Text data manufacturing method and device based on deep learning, terminal and storage medium
CN111583099A (en) * 2020-04-14 2020-08-25 上海联影智能医疗科技有限公司 Image rectification method, computer device, and storage medium
CN111783761A (en) * 2020-06-30 2020-10-16 苏州科达科技股份有限公司 Certificate text detection method and device and electronic equipment
CN113642573A (en) * 2021-07-20 2021-11-12 南京红松信息技术有限公司 Picture separation method based on grids
CN114782961A (en) * 2022-03-23 2022-07-22 华南理工大学 Character image augmentation method based on shape transformation
WO2024045442A1 (en) * 2022-08-30 2024-03-07 青岛云天励飞科技有限公司 Image correction model training method, image correction method, device and storage medium

Cited By (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111046859B (en) * 2018-10-11 2023-09-29 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN111046859A (en) * 2018-10-11 2020-04-21 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN109447911A (en) * 2018-10-18 2019-03-08 百度在线网络技术(北京)有限公司 Method, apparatus, storage medium and the terminal device of image restoration
CN109801234A (en) * 2018-12-28 2019-05-24 南京美乐威电子科技有限公司 Geometric image correction method and device
CN109801234B (en) * 2018-12-28 2023-09-22 南京美乐威电子科技有限公司 Image geometry correction method and device
CN109886264A (en) * 2019-01-08 2019-06-14 深圳禾思众成科技有限公司 A kind of character detecting method, equipment and computer readable storage medium
CN109829848A (en) * 2019-01-17 2019-05-31 柳州康云互联科技有限公司 A kind of system and method for Image space transformation neural network based in internet detection
CN110321894A (en) * 2019-04-23 2019-10-11 浙江工业大学 A kind of library book method for rapidly positioning based on deep learning OCR
CN110443782A (en) * 2019-07-03 2019-11-12 杭州深睿博联科技有限公司 Chest x-ray piece model alignment schemes and device, storage medium
CN110929784A (en) * 2019-11-21 2020-03-27 上海智臻智能网络科技股份有限公司 Method for recognizing characters in picture, computer equipment and storage medium
CN110956133A (en) * 2019-11-29 2020-04-03 上海眼控科技股份有限公司 Training method of single character text normalization model, text recognition method and device
CN111402156A (en) * 2020-03-11 2020-07-10 腾讯科技(深圳)有限公司 Restoration method and device for smear image, storage medium and terminal equipment
CN111583099A (en) * 2020-04-14 2020-08-25 上海联影智能医疗科技有限公司 Image rectification method, computer device, and storage medium
CN111539184A (en) * 2020-04-29 2020-08-14 上海眼控科技股份有限公司 Text data manufacturing method and device based on deep learning, terminal and storage medium
CN111783761A (en) * 2020-06-30 2020-10-16 苏州科达科技股份有限公司 Certificate text detection method and device and electronic equipment
CN113642573A (en) * 2021-07-20 2021-11-12 南京红松信息技术有限公司 Picture separation method based on grids
CN113642573B (en) * 2021-07-20 2023-10-13 南京红松信息技术有限公司 Picture separation method based on grids
CN114782961A (en) * 2022-03-23 2022-07-22 华南理工大学 Character image augmentation method based on shape transformation
CN114782961B (en) * 2022-03-23 2023-04-18 华南理工大学 Character image augmentation method based on shape transformation
WO2024045442A1 (en) * 2022-08-30 2024-03-07 青岛云天励飞科技有限公司 Image correction model training method, image correction method, device and storage medium

Similar Documents

Publication Publication Date Title
CN108399408A (en) A kind of deformed characters antidote based on deep space converting network
US11556797B2 (en) Systems and methods for polygon object annotation and a method of training an object annotation system
Bilinski et al. Dense decoder shortcut connections for single-pass semantic segmentation
CN105740909B (en) Text recognition method under a kind of natural scene based on spatial alternation
CN111259936B (en) Image semantic segmentation method and system based on single pixel annotation
CN114255238A (en) Three-dimensional point cloud scene segmentation method and system fusing image features
CN112330719B (en) Deep learning target tracking method based on feature map segmentation and self-adaptive fusion
CN107590497A (en) Off-line Handwritten Chinese Recognition method based on depth convolutional neural networks
CN113159466B (en) Short-time photovoltaic power generation prediction system and method
CN111862289A (en) Point cloud up-sampling method based on GAN network
CN109300128B (en) Transfer learning image processing method based on convolution neural network hidden structure
US9934553B2 (en) Method for upscaling an image and apparatus for upscaling an image
CN113312973B (en) Gesture recognition key point feature extraction method and system
CN112560865B (en) Semantic segmentation method for point cloud under outdoor large scene
CN110197255A (en) A kind of deformable convolutional network based on deep learning
CN111105439A (en) Synchronous positioning and mapping method using residual attention mechanism network
CN114757904A (en) Surface defect detection method based on AI deep learning algorithm
CN115376024A (en) Semantic segmentation method for power accessory of power transmission line
CN111340011B (en) Self-adaptive time sequence shift neural network time sequence behavior identification method
CN111861886A (en) Image super-resolution reconstruction method based on multi-scale feedback network
CN115410081A (en) Multi-scale aggregated cloud and cloud shadow identification method, system, equipment and storage medium
CN109389607A (en) Ship Target dividing method, system and medium based on full convolutional neural networks
CN115588237A (en) Three-dimensional hand posture estimation method based on monocular RGB image
Wang An augmentation small object detection method based on NAS-FPN
CN116385466A (en) Method and system for dividing targets in image based on boundary box weak annotation

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
WD01 Invention patent application deemed withdrawn after publication
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20180814