CN107679533A - Character recognition method and device - Google Patents

Character recognition method and device Download PDF

Info

Publication number
CN107679533A
CN107679533A CN201710890221.4A CN201710890221A CN107679533A CN 107679533 A CN107679533 A CN 107679533A CN 201710890221 A CN201710890221 A CN 201710890221A CN 107679533 A CN107679533 A CN 107679533A
Authority
CN
China
Prior art keywords
character image
network
distortion
layer
image
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201710890221.4A
Other languages
Chinese (zh)
Inventor
张水发
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Xiaomi Mobile Software Co Ltd
Original Assignee
Beijing Xiaomi Mobile Software Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Xiaomi Mobile Software Co Ltd filed Critical Beijing Xiaomi Mobile Software Co Ltd
Priority to CN201710890221.4A priority Critical patent/CN107679533A/en
Publication of CN107679533A publication Critical patent/CN107679533A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/20Image preprocessing
    • G06V10/24Aligning, centring, orientation detection or correction of the image
    • G06V10/243Aligning, centring, orientation detection or correction of the image by compensating for image skew or non-uniform image deformations
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V30/00Character recognition; Recognising digital ink; Document-oriented image-based pattern recognition
    • G06V30/10Character recognition

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Multimedia (AREA)
  • Character Discrimination (AREA)

Abstract

The disclosure is directed to a kind of character recognition method and device.This method includes:By character image to be identified input character image generation network, the first reparation character image is obtained;Wherein, the character image generation network trains to obtain by not distorting character image and distortion character image;Character image, which is repaired, according to described first carries out Text region.The character recognition method and device of the disclosure, character image to be identified can be repaired, obtain repairing character image, and be identified to repairing the text information in character image, the accuracy rate thus, it is possible to greatly improve Text region.

Description

Character recognition method and device
Technical field
This disclosure relates to image identification technical field, more particularly to a kind of character recognition method and device.
Background technology
In correlation technique, Text region refers to word in image is identified, verified and recorded using computer etc. The technology of reason.People will handle substantial amounts of word, form and text in production and life, and character recognition technology can be significantly Mitigate the work of people.At present, it is relatively low for the Text region accuracy rate in distortion character image, improve in distortion character image Text region accuracy rate it is significant.
The content of the invention
To overcome problem present in correlation technique, the disclosure provides a kind of character recognition method and device.
According to the first aspect of the embodiment of the present disclosure, there is provided a kind of character recognition method, including:
By character image to be identified input character image generation network, the first reparation character image is obtained;Wherein, the text Word image generates network and trains to obtain by not distorting character image and distortion character image;
Character image, which is repaired, according to described first carries out Text region.
In a kind of possible implementation, methods described also includes:
Distortion processing is carried out to the character image that do not distort, obtains the distortion character image;
Character image and the distortion character image are not distorted according to described, and training differentiates network and generation network, described Differentiate that network is used to differentiate the reparation character image and the uniformity for not distorting character image;
Network and the differentiation network are generated described in repetition training, reaches predetermined threshold value or the differentiation net in frequency of training , will when the differentiation result of network shows the consistent sexual satisfaction preparatory condition that the reparation character image does not distort character image with described Current generation network is defined as the character image generation network.
In a kind of possible implementation, character image and the distortion character image, training are not distorted according to described Differentiate network and generation network, including:
The distortion character image is inputted into the generation network, obtains the second reparation character image;
The character image and described second that do not distort is repaired into the character image input differentiation network, obtains being used for table Show that the second reparation character image does not distort the whether consistent differentiation result of character image with described;
According to the differentiation result, the value for differentiating parameter in network or the generation network is adjusted.
In a kind of possible implementation, the generation network includes the multiple coding modules connected by residual error mode With multiple decoder modules, the coding module includes convolutional layer, line rectification function layer and maximum pond layer, the decoder module Including convolutional layer, line rectification function layer and maximum pond layer.
It is described to differentiate that network includes multiple coding modules, the Duo Gequan being sequentially connected in a kind of possible implementation Articulamentum and threshold function table layer, the coding module include convolutional layer, line rectification function layer and maximum pond layer.
According to the second aspect of the embodiment of the present disclosure, there is provided a kind of character recognition device, including:
Repair module, for character image to be identified input character image to be generated into network, obtain the first reparation word graph Picture;Wherein, the character image generation network trains to obtain by not distorting character image and distortion character image;
Identification module, Text region is carried out for repairing character image according to described first.
In a kind of possible implementation, described device also includes:
Processing module, for carrying out distortion processing to the character image that do not distort, obtain the distortion character image;
Training module, for not distorting character image and the distortion character image according to described, training differentiate network with Network is generated, the differentiation network is used to differentiate the reparation character image and the uniformity for not distorting character image;
Determining module, for generating network and the differentiation network described in repetition training, reach default threshold in frequency of training Value or the differentiation result for differentiating network show that the reparation character image and the uniformity for not distorting character image expire During sufficient preparatory condition, current generation network is defined as the character image and generates network.
In a kind of possible implementation, the training module is used for:
The distortion character image is inputted into the generation network, obtains the second reparation character image;
The character image and described second that do not distort is repaired into the character image input differentiation network, obtains being used for table Show that the second reparation character image does not distort the whether consistent differentiation result of character image with described;
According to the differentiation result, the value for differentiating parameter in network or the generation network is adjusted.
In a kind of possible implementation, the generation network includes the multiple coding modules connected by residual error mode With multiple decoder modules, the coding module includes convolutional layer, line rectification function layer and maximum pond layer, the decoder module Including convolutional layer, line rectification function layer and maximum pond layer.
It is described to differentiate that network includes multiple coding modules, the Duo Gequan being sequentially connected in a kind of possible implementation Articulamentum and threshold function table layer, the coding module include convolutional layer, line rectification function layer and maximum pond layer.
According to the third aspect of the embodiment of the present disclosure, there is provided a kind of character recognition device, including:Processor;For storing The memory of processor-executable instruction;Wherein, the processor is configured as performing above-mentioned method.
According to the fourth aspect of the embodiment of the present disclosure, there is provided a kind of non-volatile computer readable storage medium storing program for executing, deposit thereon Computer program instructions are contained, the computer program instructions realize above-mentioned method when being executed by processor.
The technical scheme provided by this disclosed embodiment can include the following benefits:The character recognition method of the disclosure And device, network is generated by the way that character image to be identified is inputted into character image, the first reparation character image is obtained, according to first Repair character image and carry out Text region, wherein, character image generates network by not distorting character image and distortion word graph Obtained as training, thus, it is possible to repair character image to be identified, obtain repairing character image, and to repairing character image In text information be identified, the accuracy rate thus, it is possible to greatly improve Text region.
It should be appreciated that the general description and following detailed description of the above are only exemplary and explanatory, not The disclosure can be limited.
Brief description of the drawings
Accompanying drawing herein is merged in specification and forms the part of this specification, shows the implementation for meeting the disclosure Example, and be used to together with specification to explain the principle of the disclosure.
Fig. 1 is a kind of flow chart of character recognition method according to an exemplary embodiment.
Fig. 2 is a kind of flow chart of character recognition method according to an exemplary embodiment.
Fig. 3 is the schematic block diagram of the generation network according to an exemplary embodiment.
Fig. 4 is the schematic block diagram of the differentiation network according to an exemplary embodiment.
Fig. 5 is a kind of block diagram of character recognition device according to an exemplary embodiment.
Fig. 6 is an a kind of schematical block diagram of character recognition device according to an exemplary embodiment.
Fig. 7 is a kind of block diagram of device 800 for Text region according to an exemplary embodiment.
Embodiment
Here exemplary embodiment will be illustrated in detail, its example is illustrated in the accompanying drawings.Following description is related to During accompanying drawing, unless otherwise indicated, the same numbers in different accompanying drawings represent same or analogous key element.Following exemplary embodiment Described in embodiment do not represent all embodiments consistent with the disclosure.On the contrary, they be only with it is such as appended The example of the consistent apparatus and method of some aspects be described in detail in claims, the disclosure.
Fig. 1 is a kind of flow chart of character recognition method according to an exemplary embodiment.This method is used for word Identification equipment, the disclosure are without limitation.As shown in figure 1, the method comprising the steps of S11 and step S12.
In step s 11, by character image to be identified input character image generation network, the first reparation word graph is obtained Picture;Wherein, character image generation network trains to obtain by not distorting character image and distortion character image.
Wherein, the input of character image generation network can be character image to be identified, and output can be the first reparation text Word image.First repairs character image repairs character image corresponding to character image to be identified.First repair character image with Character image to be identified is of the same size and resolution ratio.Character image to be identified can be the character image that word is twisted Or the character image that word is not twisted or is not almost twisted, the disclosure are without limitation.
In step s 12, repair character image according to first and carry out Text region.
Wherein, the word in the first reparation character image can include Chinese character, English character, numerical character and symbol One or more in character, the first word for repairing in character image can be one in handwritten text and print hand writing Item is multinomial, and the disclosure is without limitation.
It should be noted that it will be appreciated by those skilled in the art that there are various ways to realize basis in correlation technique First, which repairs character image, carries out Text region, such as template matching method, structured analysis method or feature extraction etc., the disclosure pair This is not limited.
The character recognition method of the disclosure, character image to be identified can be repaired, obtain repairing character image, and It is identified to repairing the text information in character image, the accuracy rate thus, it is possible to greatly improve Text region.
Fig. 2 is a kind of flow chart of character recognition method according to an exemplary embodiment.As shown in Fig. 2 the party Method includes step S21 to step S25.
In the step s 21, distortion processing is carried out to not distorting character image, obtains distorting character image.
Wherein, the character image that word is not twisted or is not almost twisted, distortion text can be referred to by not distorting character image Word image can refer to the character image that word is twisted.
In a kind of possible implementation, selection does not distort character image, and what random distortion was chosen does not distort word graph Picture, obtain distorting character image.
In step S22, according to character image and distortion character image is not distorted, training differentiates network and generation network, Differentiate that network is used to differentiate the uniformity repaired character image and do not distort character image.
It should be noted that it will be appreciated by those skilled in the art that character image and distortion text are not distorted in step S21 Word image, which is used to train, differentiates network and generation network.During hands-on, character image and torsion are not distorted for every group Bent character image, alternately training differentiate network and generation network.Furthermore, it is necessary to obtain it is multigroup it is different do not distort character image and Character image is distorted, repetition training differentiates network and generation network, and the steady of network is generated with the character image that enhancing training obtains Qualitative and adaptability.
Wherein, alternately training differentiates that network and generation network can not distort character image and distortion word with pointer to every group Image, in the case where keeping the parameter constant of generation network, training differentiates network, then differentiates the parameter of network not in holding In the case of change, training generation network.Alternately training differentiates network and generation network, until generation network is according to distortion word graph As generation second repair character image, differentiate network can not differentiate do not distort character image and second repair character image whether one Cause, such as differentiate network output 0.5, i.e., the second reparation character image has 50% probability consistent with not distorting character image, has 50% probability is inconsistent with not distorting character image.
In step S23, repetition training generation network and differentiation network, reach predetermined threshold value in frequency of training or differentiate net When the differentiation result of network shows to repair character image with the consistent sexual satisfaction preparatory condition for not distorting character image, by current life It is defined as character image generation network into network.
In a kind of possible implementation, generation network includes the multiple coding modules connected by residual error mode (Encode) and multiple decoder modules (Decode), coding module include convolutional layer, line rectification function (ReLu, Rectified Linear Unit) layer and maximum pond layer (Max Pooling), decoder module includes convolutional layer, line rectification function layer and most Great Chiization layer.
Wherein, coding module is used to encode image.Decoder module is used for encoding what is obtained by coding module Image is decoded.Coding module and decoder module can change image resolution ratio and image channel number, such as increase image Resolution ratio simultaneously reduces image channel number, or reduces image resolution ratio and increase image channel number.Convolutional layer, line rectification function layer It is a basic processing unit in coding module and decoder module with maximum pond layer.
Fig. 3 is the schematic block diagram of the generation network according to an exemplary embodiment.As shown in figure 3, generation network is One ten layers of coding module-decoder module structure, including 5 coding modules connected by residual error mode and 5 decoding moulds Block.Each coding module and each decoder module include 1 convolutional layer, 1 line rectification function layer and 1 maximum pond Layer.Wherein, coding module be respectively Encode1 (n*32*3), Encode2 (n/2*16*64), Encode3 (n/4*8*128), Encode4 (n/8*4*256) and Encode5 (16/n*2*512).Decoder module be respectively Decode1 (16/n*2*512), Decode2 (n/8*4*256), Decode3 (n/4*8*128), Decode4 (n/2*16*64) and Decode5 (n*32*3).Can With understanding, the n*32 in n*32*3 can represent image resolution ratio, and 32 can represent short side resolution ratio in image, and 3 can be with Represent image channel number.
It is 32 by short side resolution adjustment in character image to be identified, long side is pressed as an example of the implementation According to proportional zoom, the first character image is obtained.First character image is inputted into generation network.First character image is by generation net Coding module in network is encoded, then is decoded by the decoder module in generation network, obtains the first reparation word graph Picture.
In a kind of possible implementation, multiple coding modules, multiple full connections that network includes being sequentially connected are differentiated (FC, Fully Connected Layers) layer and threshold function table (Sigmoid) layer, coding module include convolutional layer, linear whole Stream function layer and maximum pond layer.
Wherein, coding module is used to encode image.Full articulamentum is used to represent the distributed nature learnt It is mapped to sample labeling space.Threshold function table layer is used for variable mappings between [0,1].Coding module can change image point Resolution and image channel number, such as reduce image resolution ratio and increase image channel number.Convolutional layer, line rectification function layer and most Great Chiization layer is a basic processing unit in coding module.
Fig. 4 is the schematic block diagram of the differentiation network according to an exemplary embodiment.As shown in figure 4, differentiate network bag Include 5 coding modules being sequentially connected, 2 full articulamentums and 1 threshold function table layer.Each coding module include 1 convolutional layer, 1 line rectification function layer and 1 maximum pond layer.Wherein, coding module is respectively Encode1 ' (n*32*6), Encode2 ' (n/2*16*64), Encode3 ' (n/4*8*128), Encode4 ' (n/8*4*256) and Encode5 ' (16/n*2*512).Can With understanding, the n*32 in n*32*3 can represent image resolution ratio, and 32 can represent the resolution ratio of short side in image, and 3 can To represent image channel number.
In a kind of possible implementation, according to character image and distortion character image is not distorted, training differentiates network It can include with generation network (step S22):Distortion character image input is generated into network, obtains the second reparation character image; To not distort character image and second repair character image input differentiate network, obtain be used for represent second repair character image with The whether consistent differentiation result of character image is not distorted;According to result is differentiated, adjustment differentiates network or generates parameter in network Value.
Wherein, the value for generating parameter in network can refer to the convolution that each coding module and decoder module include in Fig. 3 The value of parameter in layer, line rectification function layer and maximum pond layer.The value of parameter can refer to each in Fig. 4 in differentiation network Convolutional layer, line rectification function layer and the maximum pond layer and each full articulamentum and threshold function table layer that coding module includes The value of middle parameter.
In a kind of possible implementation, the input for differentiating network can be not distort character image and second to repair text The fused images that word image obtains after being merged, output can be for representing that second repairs whether character image is not distort The differentiation result of character image.For example, image channel number corresponding to not distorting character image and the second reparation character image is 3, it will not distort character image and the second reparation character image is merged, and obtain the fused images that image channel number is 6, will melt Image is closed as the input for differentiating network.
In a kind of possible implementation, alternating training differentiates network and generation network, and is used according to differentiation result Back-propagation algorithm adjustment differentiates network and generates the value of parameter in network, until differentiating that network and generation network are all restrained. Wherein, differentiate that network and generation network are all restrained to refer to and differentiate that result is in stable state or frequency of training reaches default threshold Value.
In a kind of possible implementation, generation network G is determined using formula 1;
Wherein, G represents generation network, and D represents to differentiate network,Represent loss result corresponding to generation network, x Expression does not distort character image, and D (x) represents the differentiation result that x obtains as input,Represent to differentiate and damaged corresponding to network Result is lost, z represents distortion character image, and G (z) represents the generation result that z obtains as input, i.e. G (z) represents that second repairs text Word image, the differentiation result that D (G (z)) expression G (z) obtain as input, E [| | x-G (z) | |1Expression does not distort character image With the smooth loss (Smooth L1Loss) of the second difference for repairing character image.
It is understood that generation network G is the network for generating image, it receives random noise z, passes through noise Z generation image G (z).It is the network for differentiation to differentiate network D, and output differentiates result.Differentiate that result represents that input differentiates network Image whether be true picture probability.Differentiate that result is 1 and represents that input differentiates that the image 100% of network is truly to scheme Picture, differentiate that result is 0 and represents that input differentiates that the image of network is unlikely to be true picture.
During hands-on, the target for generating network G is just to try to generate true picture and go to cheat to differentiate network D. And differentiate network D target and be just to try to the image and true picture of the generation of generation network G to be distinguished from.Thus network is generated G and differentiation network D constitute a dynamic gambling process.Under optimal state, the result of last game is made a living networking Network G can generate the image G (z) for being enough to mix the spurious with the genuine, and differentiate that network D is difficult to differentiate that the image of generation network G generation is actually It is not true, therefore D (G (z))=0.5.
It should be noted that those skilled in the art, it should be appreciated that x represents true picture, z represents input generation net Network G noise, and G (z) represents the image of generation network G generation.D (x) represents to differentiate that network D judges whether true picture x is true Real probability.D (G (z)) represents to differentiate that network D judges to generate the whether real probability of image of network G generation.Because x is exactly True picture, so for differentiating network D, D (x) is better closer to 1.And G (z) is to generate the image that network G generates, institute So that for differentiating network D, D (x) is better closer to 0.Generate the purpose of network G:D (G (z)) is to differentiate network D Judge to generate the whether real probability of image that network G generates, generation network G should wish the image of oneself generation Closer to true better.That is, generation network G wish D (G (z)) as far as possible greatly, at this momentValue can diminish.Therefore the foremost of formula 1 Mark beDifferentiate network D purpose:Differentiate that network D ability is stronger, D (x) should be bigger, and D (G (z)) should be got over It is small, at this momentValue can become big.Therefore formula 1 The mark of foremost be
In step s 24, by character image to be identified input character image generation network, the first reparation word graph is obtained Picture.
Description for the step may refer to step S11.
In step s 25, repair character image according to first and carry out Text region.
Description for the step may refer to step S12.
It should be noted that it will be appreciated by those skilled in the art that step S21 to step S23, which is training, obtains word graph As the process of generation network, step S24 to step S25 is the process that actual use character image generates network.Training process is Abnormal process, the use of process is normality process.
In a kind of possible implementation, the character image that encapsulation training obtains in Text region equipment generates net Network, to cause character image generation network handles identification character image may be reused in Text region equipment to be repaired, Obtain repairing character image, and be identified to repairing the character information in character image, known thus, it is possible to greatly improve word Other accuracy rate.
The character recognition method of the disclosure, using production confrontation network generate the training of network so that generation net Network has preferable repair ability to distortion character image, can repair to obtain and not distort that character image is same or analogous to be repaiied Multiple character image, the accuracy rate thus, it is possible to greatly improve Text region.
Fig. 5 is a kind of block diagram of character recognition device according to an exemplary embodiment.Reference picture 5, the device bag Include:Repair module 51, for character image to be identified input character image to be generated into network, obtain the first reparation character image; Wherein, the character image generation network trains to obtain by not distorting character image and distortion character image;Identification module 52, Text region is carried out for repairing character image according to described first.
Fig. 6 is an a kind of schematical block diagram of character recognition device according to an exemplary embodiment.Reference picture 6:
In a kind of possible implementation, described device also includes:Processing module 53, for not distorting word to described Image carries out distortion processing, obtains the distortion character image;Training module 54, for according to it is described do not distort character image with The distortion character image, training differentiate network and generation network, and the differentiation network is used to differentiate the reparation character image With the uniformity for not distorting character image;Determining module 55, for generating network and the differentiation net described in repetition training Network, frequency of training reach predetermined threshold value or it is described differentiate network differentiation result show it is described reparation character image and it is described not When distorting the consistent sexual satisfaction preparatory condition of character image, current generation network is defined as the character image and generates net Network.
In a kind of possible implementation, the training module 54 is used for:By described in the distortion character image input Network is generated, obtains the second reparation character image;The character image and described second that do not distort is repaired into character image input The differentiation network, obtain for representing that the second reparation character image does not distort whether character image is consistent to be sentenced with described Other result;According to the differentiation result, the value for differentiating parameter in network or the generation network is adjusted.
In a kind of possible implementation, the generation network includes the multiple coding modules connected by residual error mode With multiple decoder modules, the coding module includes convolutional layer, line rectification function layer and maximum pond layer, the decoder module Including convolutional layer, line rectification function layer and maximum pond layer.
It is described to differentiate that network includes multiple coding modules, the Duo Gequan being sequentially connected in a kind of possible implementation Articulamentum and threshold function table layer, the coding module include convolutional layer, line rectification function layer and maximum pond layer.
On the device in above-described embodiment, wherein modules perform the concrete mode of operation in relevant this method Embodiment in be described in detail, explanation will be not set forth in detail herein.
The character recognition device of the disclosure, character image to be identified can be repaired, obtain repairing character image, and It is identified to repairing the text information in character image, the accuracy rate thus, it is possible to greatly improve Text region.
Fig. 7 is a kind of block diagram of device 800 for Text region according to an exemplary embodiment.For example, dress It can be mobile phone to put 800, computer, digital broadcast terminal, messaging devices, game console, tablet device, medical treatment The equipment that equipment, body-building equipment, personal digital assistant etc. have character identification function.
Reference picture 7, device 800 can include following one or more assemblies:Processing component 802, memory 804, power supply Component 806, multimedia groupware 808, audio-frequency assembly 810, the interface 812 of input/output (I/O), sensor cluster 814, and Communication component 816.
The integrated operation of the usual control device 800 of processing component 802, such as communicated with display, call, data, phase The operation that machine operates and record operation is associated.Processing component 802 can refer to including one or more processors 820 to perform Order, to complete all or part of step of above-mentioned method.In addition, processing component 802 can include one or more modules, just Interaction between processing component 802 and other assemblies.For example, processing component 802 can include multi-media module, it is more to facilitate Interaction between media component 808 and processing component 802.
Memory 804 is configured as storing various types of data to support the operation in device 800.These data are shown Example includes the instruction of any application program or method for being operated on device 800, contact data, telephone book data, disappears Breath, picture, video etc..Memory 804 can be by any kind of volatibility or non-volatile memory device or their group Close and realize, as static RAM (SRAM), Electrically Erasable Read Only Memory (EEPROM) are erasable to compile Journey read-only storage (EPROM), programmable read only memory (PROM), read-only storage (ROM), magnetic memory, flash Device, disk or CD.
Power supply module 806 provides electric power for the various assemblies of device 800.Power supply module 806 can include power management system System, one or more power supplys, and other components associated with generating, managing and distributing electric power for device 800.
Multimedia groupware 808 is included in the screen of one output interface of offer between described device 800 and user.One In a little embodiments, screen can include liquid crystal display (LCD) and touch panel (TP).If screen includes touch panel, screen Curtain may be implemented as touch-screen, to receive the input signal from user.Touch panel includes one or more touch sensings Device is with the gesture on sensing touch, slip and touch panel.The touch sensor can not only sensing touch or sliding action Border, but also detect and touched or the related duration and pressure of slide with described.In certain embodiments, more matchmakers Body component 808 includes a front camera and/or rear camera.When device 800 is in operator scheme, such as screening-mode or During video mode, front camera and/or rear camera can receive outside multi-medium data.Each front camera and Rear camera can be a fixed optical lens system or have focusing and optical zoom capabilities.
Audio-frequency assembly 810 is configured as output and/or input audio signal.For example, audio-frequency assembly 810 includes a Mike Wind (MIC), when device 800 is in operator scheme, during such as call model, logging mode and speech recognition mode, microphone by with It is set to reception external audio signal.The audio signal received can be further stored in memory 804 or via communication set Part 816 is sent.In certain embodiments, audio-frequency assembly 810 also includes a loudspeaker, for exports audio signal.
I/O interfaces 812 provide interface between processing component 802 and peripheral interface module, and above-mentioned peripheral interface module can To be keyboard, click wheel, button etc..These buttons may include but be not limited to:Home button, volume button, start button and lock Determine button.
Sensor cluster 814 includes one or more sensors, and the state for providing various aspects for device 800 is commented Estimate.For example, sensor cluster 814 can detect opening/closed mode of device 800, and the relative positioning of component, for example, it is described Component is the display and keypad of device 800, and sensor cluster 814 can be with 800 1 components of detection means 800 or device Position change, the existence or non-existence that user contacts with device 800, the orientation of device 800 or acceleration/deceleration and device 800 Temperature change.Sensor cluster 814 can include proximity transducer, be configured to detect in no any physical contact The presence of neighbouring object.Sensor cluster 814 can also include optical sensor, such as CMOS or ccd image sensor, for into As being used in application.In certain embodiments, the sensor cluster 814 can also include acceleration transducer, gyro sensors Device, Magnetic Sensor, pressure sensor or temperature sensor.
Communication component 816 is configured to facilitate the communication of wired or wireless way between device 800 and other equipment.Device 800 can access the wireless network based on communication standard, such as WiFi, 2G or 3G, or combinations thereof.In an exemplary implementation In example, communication component 816 receives broadcast singal or broadcast related information from external broadcasting management system via broadcast channel. In one exemplary embodiment, the communication component 816 also includes near-field communication (NFC) module, to promote junction service.Example Such as, in NFC module radio frequency identification (RFID) technology can be based on, Infrared Data Association (IrDA) technology, ultra wide band (UWB) technology, Bluetooth (BT) technology and other technologies are realized.
In the exemplary embodiment, device 800 can be believed by one or more application specific integrated circuits (ASIC), numeral Number processor (DSP), digital signal processing appts (DSPD), PLD (PLD), field programmable gate array (FPGA), controller, microcontroller, microprocessor or other electronic components are realized, for performing the above method.
In the exemplary embodiment, a kind of non-transitorycomputer readable storage medium including instructing, example are additionally provided Such as include the memory 804 of instruction, above-mentioned instruction can be performed to complete the above method by the processor 820 of device 800.For example, The non-transitorycomputer readable storage medium can be ROM, random access memory (RAM), CD-ROM, tape, floppy disk With optical data storage devices etc..
Those skilled in the art will readily occur to the disclosure its after considering specification and putting into practice invention disclosed herein Its embodiment.The application is intended to any modification, purposes or the adaptations of the disclosure, these modifications, purposes or Person's adaptations follow the general principle of the disclosure and including the undocumented common knowledges in the art of the disclosure Or conventional techniques.Description and embodiments are considered only as exemplary, and the true scope of the disclosure and spirit are by following Claim is pointed out.
It should be appreciated that the precision architecture that the disclosure is not limited to be described above and is shown in the drawings, and And various modifications and changes can be being carried out without departing from the scope.The scope of the present disclosure is only limited by appended claim.

Claims (12)

  1. A kind of 1. character recognition method, it is characterised in that including:
    By character image to be identified input character image generation network, the first reparation character image is obtained;Wherein, the word graph Train to obtain by not distorting character image and distortion character image as generating network;
    Character image, which is repaired, according to described first carries out Text region.
  2. 2. according to the method for claim 1, it is characterised in that methods described also includes:
    Distortion processing is carried out to the character image that do not distort, obtains the distortion character image;
    Character image and the distortion character image are not distorted according to described, and training differentiates network and generation network, the differentiation Network is used to differentiate the reparation character image and the uniformity for not distorting character image;
    Network and the differentiation network are generated described in repetition training, reaches predetermined threshold value in frequency of training or described differentiates network , will be current when differentiating that result shows the consistent sexual satisfaction preparatory condition that the reparation character image does not distort character image with described Generation network be defined as character image generation network.
  3. 3. according to the method for claim 2, it is characterised in that do not distort character image and the distortion word according to described Image, training differentiate network and generation network, including:
    The distortion character image is inputted into the generation network, obtains the second reparation character image;
    The character image and described second that do not distort is repaired into the character image input differentiation network, obtains being used to represent institute State the second reparation character image and do not distort the whether consistent differentiation result of character image with described;
    According to the differentiation result, the value for differentiating parameter in network or the generation network is adjusted.
  4. 4. according to the method for claim 1, it is characterised in that the generation network is more including being connected by residual error mode Individual coding module and multiple decoder modules, the coding module include convolutional layer, line rectification function layer and maximum pond layer, institute Stating decoder module includes convolutional layer, line rectification function layer and maximum pond layer.
  5. 5. according to the method for claim 2, it is characterised in that described to differentiate that network includes the multiple coding moulds being sequentially connected Block, multiple full articulamentums and threshold function table layer, the coding module include convolutional layer, line rectification function layer and maximum pond Layer.
  6. A kind of 6. character recognition device, it is characterised in that including:
    Repair module, for character image to be identified input character image to be generated into network, obtain the first reparation character image;Its In, the character image generation network trains to obtain by not distorting character image and distortion character image;
    Identification module, Text region is carried out for repairing character image according to described first.
  7. 7. device according to claim 6, it is characterised in that described device also includes:
    Processing module, for carrying out distortion processing to the character image that do not distort, obtain the distortion character image;
    Training module, for not distorting character image and the distortion character image according to, training differentiates network and generation Network, the differentiation network are used to differentiate the reparation character image and the uniformity for not distorting character image;
    Determining module, for generating network and the differentiation network described in repetition training, frequency of training reach predetermined threshold value or The differentiation result for differentiating network shows that the reparation character image and the consistent sexual satisfaction for not distorting character image are pre- If during condition, current generation network is defined as the character image and generates network.
  8. 8. device according to claim 7, it is characterised in that the training module is used for:
    The distortion character image is inputted into the generation network, obtains the second reparation character image;
    The character image and described second that do not distort is repaired into the character image input differentiation network, obtains being used to represent institute State the second reparation character image and do not distort the whether consistent differentiation result of character image with described;
    According to the differentiation result, the value for differentiating parameter in network or the generation network is adjusted.
  9. 9. device according to claim 6, it is characterised in that the generation network is more including being connected by residual error mode Individual coding module and multiple decoder modules, the coding module include convolutional layer, line rectification function layer and maximum pond layer, institute Stating decoder module includes convolutional layer, line rectification function layer and maximum pond layer.
  10. 10. device according to claim 7, it is characterised in that described to differentiate that network includes the multiple codings being sequentially connected Module, multiple full articulamentums and threshold function table layer, the coding module include convolutional layer, line rectification function layer and maximum pond Layer.
  11. A kind of 11. character recognition device, it is characterised in that including:
    Processor;
    For storing the memory of processor-executable instruction;
    Wherein, the processor is configured as the method described in any one in perform claim requirement 1 to 5.
  12. 12. a kind of non-volatile computer readable storage medium storing program for executing, is stored thereon with computer program instructions, it is characterised in that institute State and method in claim 1 to 5 described in any one is realized when computer program instructions are executed by processor.
CN201710890221.4A 2017-09-27 2017-09-27 Character recognition method and device Pending CN107679533A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710890221.4A CN107679533A (en) 2017-09-27 2017-09-27 Character recognition method and device

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710890221.4A CN107679533A (en) 2017-09-27 2017-09-27 Character recognition method and device

Publications (1)

Publication Number Publication Date
CN107679533A true CN107679533A (en) 2018-02-09

Family

ID=61137533

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710890221.4A Pending CN107679533A (en) 2017-09-27 2017-09-27 Character recognition method and device

Country Status (1)

Country Link
CN (1) CN107679533A (en)

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109871841A (en) * 2019-02-14 2019-06-11 腾讯科技(深圳)有限公司 Image processing method, device, terminal and storage medium
CN110084172A (en) * 2019-04-23 2019-08-02 北京字节跳动网络技术有限公司 Character recognition method, device and electronic equipment
CN110363189A (en) * 2018-04-09 2019-10-22 珠海金山办公软件有限公司 A kind of document content restorative procedure, device, electronic equipment and readable storage medium storing program for executing
CN110889469A (en) * 2019-09-19 2020-03-17 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN111402156A (en) * 2020-03-11 2020-07-10 腾讯科技(深圳)有限公司 Restoration method and device for smear image, storage medium and terminal equipment
CN111723788A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN111783756A (en) * 2019-04-03 2020-10-16 北京市商汤科技开发有限公司 Text recognition method and device, electronic equipment and storage medium
WO2022047662A1 (en) * 2020-09-02 2022-03-10 Intel Corporation Method and system of neural network object recognition for warpable jerseys with multiple attributes

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101661613A (en) * 2009-08-27 2010-03-03 北京交通大学 Image restoration method based on image segmentation, and system therefor
CN102063621A (en) * 2010-11-30 2011-05-18 汉王科技股份有限公司 Method and device for correcting geometric distortion of character lines
CN103413271A (en) * 2013-07-18 2013-11-27 西安交通大学 Document image rectifying method based on local information
CN104809733A (en) * 2015-05-08 2015-07-29 中北大学 Ancient building wall polluted inscription character image edge extraction method
CN104899571A (en) * 2015-06-12 2015-09-09 成都数联铭品科技有限公司 Random sample generation method for recognition of complex character
CN106648418A (en) * 2016-11-14 2017-05-10 北京爱知之星科技股份有限公司 Handwriting correction method and device
CN107133934A (en) * 2017-05-18 2017-09-05 北京小米移动软件有限公司 Image completion method and device

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101661613A (en) * 2009-08-27 2010-03-03 北京交通大学 Image restoration method based on image segmentation, and system therefor
CN102063621A (en) * 2010-11-30 2011-05-18 汉王科技股份有限公司 Method and device for correcting geometric distortion of character lines
CN103413271A (en) * 2013-07-18 2013-11-27 西安交通大学 Document image rectifying method based on local information
CN104809733A (en) * 2015-05-08 2015-07-29 中北大学 Ancient building wall polluted inscription character image edge extraction method
CN104899571A (en) * 2015-06-12 2015-09-09 成都数联铭品科技有限公司 Random sample generation method for recognition of complex character
CN106648418A (en) * 2016-11-14 2017-05-10 北京爱知之星科技股份有限公司 Handwriting correction method and device
CN107133934A (en) * 2017-05-18 2017-09-05 北京小米移动软件有限公司 Image completion method and device

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
RAYMOND A. YEH: "Semantic Image Inpainting with Deep Generative Models", 《CVPR2017》 *

Cited By (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110363189A (en) * 2018-04-09 2019-10-22 珠海金山办公软件有限公司 A kind of document content restorative procedure, device, electronic equipment and readable storage medium storing program for executing
CN110363189B (en) * 2018-04-09 2021-09-24 珠海金山办公软件有限公司 Document content restoration method and device, electronic equipment and readable storage medium
CN109871841B (en) * 2019-02-14 2023-04-21 腾讯科技(深圳)有限公司 Image processing method, device, terminal and storage medium
CN109871841A (en) * 2019-02-14 2019-06-11 腾讯科技(深圳)有限公司 Image processing method, device, terminal and storage medium
CN111723788A (en) * 2019-03-19 2020-09-29 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN111723788B (en) * 2019-03-19 2023-12-29 杭州海康威视数字技术股份有限公司 Character recognition method and device
CN111783756B (en) * 2019-04-03 2024-04-16 北京市商汤科技开发有限公司 Text recognition method and device, electronic equipment and storage medium
CN111783756A (en) * 2019-04-03 2020-10-16 北京市商汤科技开发有限公司 Text recognition method and device, electronic equipment and storage medium
CN110084172A (en) * 2019-04-23 2019-08-02 北京字节跳动网络技术有限公司 Character recognition method, device and electronic equipment
CN110889469A (en) * 2019-09-19 2020-03-17 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN110889469B (en) * 2019-09-19 2023-07-21 北京市商汤科技开发有限公司 Image processing method and device, electronic equipment and storage medium
CN111402156B (en) * 2020-03-11 2021-08-03 腾讯科技(深圳)有限公司 Restoration method and device for smear image, storage medium and terminal equipment
CN111402156A (en) * 2020-03-11 2020-07-10 腾讯科技(深圳)有限公司 Restoration method and device for smear image, storage medium and terminal equipment
WO2022047662A1 (en) * 2020-09-02 2022-03-10 Intel Corporation Method and system of neural network object recognition for warpable jerseys with multiple attributes

Similar Documents

Publication Publication Date Title
CN107679533A (en) Character recognition method and device
CN107609560A (en) Character recognition method and device
CN107679483A (en) Number plate recognition methods and device
CN107527059B (en) Character recognition method and device and terminal
CN107798327A (en) Character identifying method and device
CN106530227A (en) Image restoration method and device
CN107845062A (en) image generating method and device
CN106651955A (en) Method and device for positioning object in picture
CN107993210A (en) Image repair method, device and computer-readable recording medium
CN106792170A (en) Method for processing video frequency and device
CN107944447A (en) Image classification method and device
CN105426857A (en) Training method and device of face recognition model
CN105335754A (en) Character recognition method and device
CN105335713A (en) Fingerprint identification method and device
CN110096209A (en) Handwriting trace display methods and device
TW201234213A (en) Multimedia input method
CN106682736A (en) Image identification method and apparatus
CN108737891A (en) Video material processing method and processing device
CN106778531A (en) Face detection method and device
CN107230137A (en) Merchandise news acquisition methods and device
CN105354560A (en) Fingerprint identification method and device
CN109819288A (en) Determination method, apparatus, electronic equipment and the storage medium of advertisement dispensing video
CN108345581A (en) A kind of information identifying method, device and terminal device
CN107038428A (en) Vivo identification method and device
CN110717399A (en) Face recognition method and electronic terminal equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination