CN105654127A - End-to-end-based picture character sequence continuous recognition method - Google Patents

End-to-end-based picture character sequence continuous recognition method Download PDF

Info

Publication number
CN105654127A
CN105654127A CN201511018552.6A CN201511018552A CN105654127A CN 105654127 A CN105654127 A CN 105654127A CN 201511018552 A CN201511018552 A CN 201511018552A CN 105654127 A CN105654127 A CN 105654127A
Authority
CN
China
Prior art keywords
neural network
recurrent neural
data
neural networks
moment
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201511018552.6A
Other languages
Chinese (zh)
Inventor
刘世林
何宏靖
陈炳章
吴雨浓
姚佳
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chengdu Business Big Data Technology Co Ltd
Original Assignee
Chengdu Business Big Data Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chengdu Business Big Data Technology Co Ltd filed Critical Chengdu Business Big Data Technology Co Ltd
Priority to CN201511018552.6A priority Critical patent/CN105654127A/en
Publication of CN105654127A publication Critical patent/CN105654127A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/088Non-supervised learning, e.g. competitive learning

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computational Linguistics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Health & Medical Sciences (AREA)
  • Biomedical Technology (AREA)
  • Biophysics (AREA)
  • Evolutionary Biology (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention belongs to the picture character recognition field and relates to an end-to-end-based picture character sequence continuous recognition method. According to the method of the invention, CNN (Convolutional Neural Network) and RNN (Recurrent Neural Network) technologies are adopted; feature extraction is performed on a whole picture containing a plurality of characters through a CNN; identical features are transmitted to an RNN so as to be subjected to repeatedly recursive use; and continuous prediction of the plurality of characters can be realized. With the method adopted, a defect that picture segmentation is required before OCR (optical character recognition) can be eliminated, the early-stage processing process of picture character recognition can be simplified, and the efficiency of character recognition can be significantly improved; and since the RNN recursively uses output data of the last round, the recognition accuracy of character and word sequences can be better improved, and the processing efficiency of character recognition can be further improved.

Description

Based on picture character Sequentially continuous recognition methods end to end
Technical field
The present invention relates to pictograph identification field, particularly to based on picture character Sequentially continuous recognition methods end to end.
Background technology
Development along with society, create a large amount of to digitized demands of paper media such as ancient books, document, bill, business cards, here digitized is not limited only to use scanner or camera to carry out " photo ", the more important thing is to change into these paper documents and store with document readable, editable, realize this process to need the picture scanned is carried out pictograph identification, and traditional pictograph is identified as optical character identification (OCR), optical character identification is identified being scanned on the basis of electronic image by paper document to be identified. It is contemplated that the quality of the scanning quality of effect, paper document itself is (not such as printing quality, font clarity, font specification degree etc.), contents and distribution's (arranging situation of word, ratio plain text and table text and bill) difference, the actual effect of OCR is always not satisfactory. And requirement for the recognition accuracy of different paper documents is variant, the identification of such as bill, requirement to accuracy rate is very high, because if a numeral identifies that mistake this may result in fatal consequence, traditional OCR identifies can not meet so high-precision identification requirement.
Conventional OCR method includes the processing procedures such as the cutting of picture, feature extraction, monocase identification, and wherein the cutting of picture contains substantial amounts of Image semantic classification process, such as Slant Rectify, background denoising, the extraction of monocase; These processing procedures are not only loaded down with trivial details consuming time, and it would furthermore be possible to make picture lose a lot of available informations; And when picture to be identified comprises the character string of multiple word, traditional OCR method needs that former character string is cut into some little pictures comprising single word and identifies respectively, the method major problem is that: the cutting difficulty of a monocase picture, particularly it is mixed with the man of left and right radical, letter, numeral, symbol, or when the distortion of background noise, character, bonding, cutting is more difficult. And once cutting there is a problem, be difficult to obtain recognition result accurately. Identify that need to be badly in need of can image character recognition method rapidly and efficiently in the face of huge.
Summary of the invention
It is an object of the invention to overcome above-mentioned deficiency existing in prior art, it is provided that based on picture character Sequentially continuous recognition methods end to end.Invention applies convolutional neural networks (CNN) and the technology of recurrent neural network (RNN), by CNN, the whole picture comprising multiple character is carried out feature extraction, then same feature feeding RNN is carried out recurrence to reuse, to realize predicting continuously the purpose of multiple characters. The optical character sequence recognition that the inventive method realizes, overcoming before OCR identifies of system first to carry out the drawback of picture cutting, greatly improves the recognition efficiency of pictograph.
In order to realize foregoing invention purpose, the invention provides techniques below scheme:
Based on picture character Sequentially continuous recognition methods end to end, comprise implemented below step:
(1) convolutional neural networks and recurrent neural networks model are built, each moment of wherein said recurrent neural network input signal includes: the sample characteristics data that described convolutional neural networks extracts, the vector data that the words that the output data of a upper moment recurrent neural network and a upper moment recurrent neural network identify changes into;
(2) training sample set is used to train described convolutional neural networks and recurrent neural networks model;
(3) in described convolutional neural networks pictograph sequence inputting to be identified trained and recurrent neural network, the characteristic of picture to be identified is extracted by described convolutional neural networks, it is input in described recurrent neural network, through the iteration successively of described recurrent neural network, export the complete recognition result of pictograph sequence to be identified.
Concrete, the computing formula of the recurrent neural network forward algorithm used in the inventive method is as follows:
a h t = Σ i I w i h x i t + Σ h ′ H w h ′ h b h ′ t - 1
b h t = θ ( a h t )
a k t = Σ h H w h k b h t
y k t = exp ( a k t ) Σ k ′ k exp ( a k ′ t )
Wherein I is the dimension of input vector, and H is the neuron number of hidden layer, and K is the neuron number of output layer, and x is the characteristic that convolutional neural networks extracts,For the input of hidden layer neuron in current time recurrent neural network,Output for current time recurrent neural network hidden layer neuron; wih, wh��h, forCorresponding weight parameter.For the current time neuronic input of recurrent neural network output layer; whkFor the weight that each neuron of output layer is corresponding;For the current time neuronic output of recurrent neural network output layer,It is a probit, represents the current time correspondence neuron output value ratio adding sum relative to all neuron output values of output layer.
Further, in the inventive method, the parameter w used during the transmission of signal forwardih, wh��hAll share across sequential, this avoid the linear increase of model complexity, cause possible over-fitting.
Further, the present invention adopts above-mentioned forward algorithm to transmit operational data in convolutional neural networks and recurrent neural network step by step, identification (prediction) data are got at output layer, when the annotation results with training sample that predicts the outcome has deviation, adjust each weight in neutral net by error backpropagation algorithm classical in neutral net.
Further, in neural network training process, checked the training result of neutral net by development set, adjust the training direction of neutral net in time, prevent the generation of over-fitting situation, in model training process, be only only remained in the training pattern that in development set, recognition accuracy is the highest.
Further, this comprises implemented below step based on the neural network training process of picture character Sequentially continuous recognition methods end to end:
(2-1) training sample manually marked is input in convolutional neural networks;
(2-2) by described convolutional network, input training sample is carried out feature extraction;
(2-3) characteristic extracted by described convolutional neural networks inputs in the first moment recurrent neural network as the first data;
(2-4) the first prediction data is exported through the calculating of the first moment recurrent neural network; Obtain the words recognition result of this moment recurrent neural network according to the first prediction data, this result is defined as: the first recognition result;
(2-5) vector data of correspondence and by the first recognition result is changed into;
(2-6) by the first data, first recognition result of the first prediction data and vectorization is as the input data of the second moment recurrent neural network, calculating through recurrent neural network exports the second prediction data, and obtains second recognition result corresponding by the second prediction data;
(27) using input data as the 3rd moment recurrent neural network of the first data and the second prediction data;
Recurrence successively, until reaching the recurrence number of times set, terminates identifying; Each moment RNN is doped word (or word) is recorded successively and just final is obtained complete string content.
Further, when carrying out model training, comprising and be normalized and artificial annotation process by training sample icon, wherein normalized process includes: sets most long word (or word) number that picture sentence is possible, such as sets the length of sentence as 25.
Further, in being normalized process, in order to avoid transformation of data, the zoom of size uses the mode of equal proportion, with the region background colour polishing of target size disappearance.
Further, normalized picture is manually marked, it is possible to first natural language is carried out word segmentation processing, such as " this thing is very good " is become " this thing is very good "; If the sentence number of words of mark is less than 25, using a special word:<EOS>carries out polishing (length to 25), then randomly select the data of 75% as training set, the data of 25% are as development set.
Compared with prior art, beneficial effects of the present invention: the present invention provides based on picture character Sequentially continuous recognition methods end to end, the present invention adopts convolutional neural networks that word sequence picture to be identified is carried out entirety and levies extraction, and the characteristic extracted is input in the recurrent neural network in each moment as the first data, the pictograph sequence recognition that the inventive method realizes, picture global feature is extracted by convolutional neural networks, it is being made without the basis of single character cutting and noise filtering achieves the identification of whole word sequence, relative to traditional OCR method, present invention, avoiding the inaccurate irreversible identification mistake that may result in of character segmentation, greatly simplify the early stage processing procedure of pictograph identification, significantly improve the efficiency of Text region.
In addition the inventive method realizes the continuous identification of character in word sequence by recurrent neural network, when using recurrent neural network to identify character, the input signal of each moment recurrent neural network also comprises the output data of a moment recurrent neural network, each moment recurrent neural network is when carrying out the Text region of correspondence, namely the picture global feature that convolutional neural networks extracts has been relied on, the output data of a upper moment recurrent neural network are also relied on it, compared to OCR method, recognition accuracy is higher, and simplify the end processing sequences identifying word, recognition efficiency is higher, recognition result is more accurately and reliably.
In a word, the inventive method simplifies the processing procedure of pictograph sequence recognition, significantly improve recognition efficiency and accuracy rate, make developer can focus more on the deposit of the tuning in model and data, improving development efficiency, the inventive method has high using value and is widely applied prospect in pictograph identification field.
Accompanying drawing illustrates:
Fig. 1 be the inventive method realize process schematic.
Fig. 2 is convolutional neural networks structural representation.
Fig. 3 is that the inventive method word sequence identification process signal flows to schematic diagram.
Detailed description of the invention
Below in conjunction with test example and detailed description of the invention, the present invention is described in further detail. But this should not being interpreted as, the scope of the above-mentioned theme of the present invention is only limitted to below example, and all technology realized based on present invention belong to the scope of the present invention.
The present invention provides based on picture character Sequentially continuous recognition methods end to end. Invention applies convolutional neural networks (CNN) and the technology of recurrent neural network (RNN), by CNN, the whole picture comprising multiple character is carried out feature extraction, then same feature feeding RNN is carried out recurrence to reuse, to realize predicting continuously the purpose of multiple characters. The optical character sequence recognition that the inventive method realizes, overcoming before OCR identifies of system first to carry out the drawback of picture cutting, greatly improve the recognition efficiency of pictograph, developer is made to focus more on the deposit of the tuning in model and data, improve development efficiency, again due in the process of model training and application RNN recurrence employ last round of output data, identification accurately higher.
Techniques below scheme is the invention provides: based on picture character Sequentially continuous recognition methods end to end, comprise implemented below step as shown in Figure 1 in order to realize foregoing invention purpose:
(1) convolutional neural networks and recurrent neural networks model are built, each moment of wherein said recurrent neural network input signal includes: the sample characteristics data that described convolutional neural networks extracts, the vector data that the words that the output data of a upper moment recurrent neural network and a upper moment recurrent neural network identify changes into; As shown in Figure 2: described convolutional neural networks is mainly used for the automatic study of picture feature. Wherein, each characteristic pattern (featuremap, shown in vertical setting of types rectangle in figure) generation be all by an own convolution kernel (i.e. little rectangle frame in Fig. 2, it is shared in the characteristic pattern specified) carry out preliminary feature extraction, the feature that convolutional layer is extracted by double sampling layer is sampled and is mainly solved the redundancy of convolutional layer institute extraction feature. In brief, described convolutional neural networks extracts the different characteristic of picture by convolutional layer, by double sampling layer, the feature extracted is sampled, (multiple convolutional layer can be comprised in a convolutional neural networks removing redundancy, double sampling layer and full articulamentum), finally by full articulamentum different characteristic patterns is together in series and constitutes final full picture feature, the inventive method uses a convolutional neural networks, whole pictures is carried out disposable feature extraction, it is entirely avoided the irreversible identification mistake that picture cutting may result in.
(2) training sample set is used to train described convolutional neural networks and recurrent neural networks model.
(3) in described convolutional neural networks pictograph sequence inputting to be identified trained and recurrent neural network, the characteristic of picture to be identified is extracted by described convolutional neural networks, it is input in described recurrent neural network, through the iteration successively of described recurrent neural network, export the complete recognition result of pictograph sequence to be identified.
Concrete, the computing formula of the recurrent neural network forward algorithm used in the inventive method is as follows:
a h t = &Sigma; i I w i h x i t + &Sigma; h &prime; H w h &prime; h b h &prime; t - 1
b h t = &theta; ( a h t )
a k t = &Sigma; h H w h k b h t
y k t = exp ( a k t ) &Sigma; k &prime; k exp ( a k &prime; t )
Wherein I is the dimension of input vector, and H is the neuron number of hidden layer, and K is the neuron number of output layer, and x is the characteristic that convolutional neural networks extracts,For the input of hidden layer neuron in current time recurrent neural network,Output (b for current time recurrent neural network hidden layer neuron0=0), �� () isArriveFunction; wih, wh��h, forCorresponding weight parameter, in a forward algorithm transmittance process, parameter wih, wh��hAll sharing across sequential, so-called sharing across sequential refers to that recurrent neural network is in signal forward transmittance process, each moment wih, wh��hIdentical (the not w of valueih=wh��h), the not w of RNN in the same timeih, wh��hIt is worth identical, reduces the complexity of model parameter, it also avoid the linear increase of model complexity and cause possible over-fitting.For the current time neuronic input of recurrent neural network output layer; whkFor the weight that each neuron of output layer is corresponding;For the current time neuronic output of recurrent neural network output layer,It is a probit, represents the current time correspondence neuron output value ratio adding sum relative to all neuron output values of output layer, generally, will selectThe recognition result that classification is this moment recurrent neural network that the maximum output neuron of value is corresponding.
Can be seen that the input data of hidden layer neuron in the recurrent neural network used the inventive method include the CNN training sample feature extracted, the output data of a upper moment recurrent neural network hidden layer from above-mentioned formula. Therefore the recurrent neural network that the present invention uses is when the word (word) of prediction current time, both the feature of image had been relied on, be also relied on the feature (language model) of upper moment output, signal transduction process as it is shown on figure 3, recognition efficiency and accuracy rate.
Further, the present invention adopts above-mentioned forward algorithm to transmit operational data in convolutional neural networks and recurrent neural network step by step, identification (prediction) data are got at output layer, when the annotation results with training sample that predicts the outcome has deviation, each weight in neutral net is adjusted by error backpropagation algorithm classical in neutral net, error back propagation step by step is shared all neurons of each layer by error back propagation method, obtain the neuronic error signal of each layer, and then revise each neuronic weight. Successively transmitted operational data by forward algorithm, and the process being revised a neuronic weight by backward algorithm gradually is exactly the training process of neutral net; Repeating said process, until the accuracy that predicts the outcome reaches the threshold value set, deconditioning, now it is believed that neural network model is trained completes.
Further, in neural network training process, checked the training result of neutral net by development set, adjust the training direction of neutral net in time, prevent the generation of over-fitting situation, in model training process, be only only remained in the training pattern that in development set, recognition accuracy is the highest.
Further, this comprises implemented below step based on the neural network training process of picture character Sequentially continuous recognition methods end to end:
(2-1) training sample manually marked is input in convolutional neural networks;
(2-2) by described convolutional network, input training sample is carried out feature extraction;
(2-3) characteristic extracted by described convolutional neural networks inputs in the first moment recurrent neural network as the first data;
(2-4) the first prediction data is exported through the calculating of the first moment recurrent neural network; Obtain the words recognition result of this moment recurrent neural network according to the first prediction data, this result is defined as: the first recognition result;
(2-5) vector data of correspondence and by the first recognition result is changed into;
(2-6) by the first data, first recognition result of the first prediction data and vectorization is as the input data of the second moment recurrent neural network, calculating through recurrent neural network exports the second prediction data, and obtains second recognition result corresponding by the second prediction data;
(2-7) using input data as the 3rd moment recurrent neural network of the first data and the second prediction data;
Recurrence successively, the vector of words (recognition result) correspondence that characteristic (the first data), the output data (prediction data) of a upper moment RNN and the upper moment RNN extracted by CNN identifies, as the input data of current time RNN, the prediction through RNN exports a word (or word); Until reaching the recurrence number of times set, terminate identifying; Each moment RNN is doped word (or word) is recorded successively and just final is obtained complete string content.
Further, when carrying out model training, comprise and training sample icon is normalized and artificial annotation process, normalized sample, the basic parameter making sample is homogeneous, reduces data unrelated complexity during model training, is conducive to simplified model training process; Wherein normalized process includes: set most long word (or word) number that picture sentence is possible, such as set the length of sentence as 25, the length of word sequence to be identified is corresponding with the maximum recurrence number of times of recurrent neural network, the longest number of characters that word sequence to be identified is set when being trained sample and preparing can be corresponding the maximum recurrence number of times of default recurrent neural network, increase the stability of model and predictable.
Further, in being normalized process, in order to avoid transformation of data, the zoom of size uses the mode of equal proportion, with the region background colour polishing of target size disappearance.
Further, normalized picture is manually marked, it is possible to first natural language is carried out word segmentation processing, such as " this thing is very good " is become " this thing is very good ". If sentence word (or word) number of mark less than the maximum word (or word) number (less than 25) arranged, uses one special word to carry out polishing (such as use "<EOS>" by the samples pictures polishing less than 25 characters (or word) to the length of 25 characters (or word)).
Further, after above-mentioned normalized and artificial mark, randomly select the data of 75% as training sample set, choose the data of 25% as development sample collection. Neutral net is only saved in the model that in development set, recognition accuracy is the highest, development sample and the uniform format of training sample in the training process, is conducive to improving the training effectiveness of neutral net.

Claims (7)

1. based on picture character Sequentially continuous recognition methods end to end, it is characterised in that comprise implemented below step:
(1) building convolutional neural networks and recurrent neural networks model, input of wherein said recurrent neural network each moment signal includes: the sample characteristics data that described convolutional neural networks extracts, the output data of a upper moment recurrent neural network;
(2) training sample set is used to train described convolutional neural networks and recurrent neural networks model;
(3), in the described convolutional neural networks trained by pictograph sequence inputting to be identified and recurrent neural network, the complete recognition result of pictograph sequence to be identified is exported.
2. the method for claim 1, it is characterised in that: the recurrent neural networks model used in this method adopts following forward algorithm formula:
a h t = &Sigma; i I w i h x i t + &Sigma; h &prime; H w h &prime; h b h &prime; t - 1
b h t = &theta; ( a h t )
a k t = &Sigma; h H w h k b h t
y k t = exp ( a k t ) &Sigma; k &prime; k exp ( a k &prime; t )
Wherein I is the dimension of input vector, and H is the neuron number of hidden layer, and K is the neuron number of output layer, and x is the characteristic that convolutional neural networks extracts,For the input of hidden layer neuron in current time recurrent neural network,Output for current time recurrent neural network hidden layer neuron;For the current time neuronic input of recurrent neural network output layer;For the current time neuronic output of recurrent neural network output layer,It is a probit, represents the current time correspondence neuron output value ratio adding sum relative to all neuron output values of output layer.
3. method as claimed in claim 2, it is characterised in that: described parameter wih, wh��hFor sharing across sequential, the w that each moment uses in a sample training processih, wh��hIt is worth identical.
4. method as claimed in claim 3, it is characterised in that: in neural network training process, checked the training result of neutral net by development set, be only remained in convolutional neural networks and recurrent neural networks model that in development set, recognition accuracy is the highest.
5. the method as described in one of claims 1 to 3, it is characterised in that: comprise implemented below step:
(2-1) training sample manually marked is input in convolutional neural networks;
(2-2) by described convolutional network, input training sample is carried out feature extraction;
(2-3) characteristic extracted by described convolutional neural networks inputs in the first moment recurrent neural network as the first data;
(2-4) the first prediction data is exported through the calculating of the first moment recurrent neural network; Obtain the words recognition result of this moment recurrent neural network according to the first prediction data, this result is defined as: the first recognition result;
(2-5) vector data of correspondence and by the first recognition result is changed into;
(2-6) by the first data, first recognition result of the first prediction data and vectorization is as the input data of the second moment recurrent neural network, calculating through recurrent neural network exports the second prediction data, and obtains second recognition result corresponding by the second prediction data;
(2-7) using input data as the 3rd moment recurrent neural network of the first data and the second prediction data;
Recurrence successively, until when reaching the recurrence number of times set, terminating calculating.
6. method as claimed in claim 5, it is characterised in that: when preparing training sample and development sample, samples pictures being normalized, described normalized includes: the most long word number arranging that picture to be identified allows or word number.
7. method as claimed in claim 6, it is characterised in that: when the sample being normalized manually is marked, when the number of words comprised in samples pictures is less than the most long word number arranged, use the marker character set by the number of words polishing in samples pictures.
CN201511018552.6A 2015-12-30 2015-12-30 End-to-end-based picture character sequence continuous recognition method Pending CN105654127A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201511018552.6A CN105654127A (en) 2015-12-30 2015-12-30 End-to-end-based picture character sequence continuous recognition method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201511018552.6A CN105654127A (en) 2015-12-30 2015-12-30 End-to-end-based picture character sequence continuous recognition method

Publications (1)

Publication Number Publication Date
CN105654127A true CN105654127A (en) 2016-06-08

Family

ID=56477503

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201511018552.6A Pending CN105654127A (en) 2015-12-30 2015-12-30 End-to-end-based picture character sequence continuous recognition method

Country Status (1)

Country Link
CN (1) CN105654127A (en)

Cited By (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446782A (en) * 2016-08-29 2017-02-22 北京小米移动软件有限公司 Image identification method and device
CN106530569A (en) * 2016-10-17 2017-03-22 北京小米移动软件有限公司 Method and device for fire monitoring
CN106570497A (en) * 2016-10-08 2017-04-19 中国科学院深圳先进技术研究院 Text detection method and device for scene image
CN107784303A (en) * 2016-12-15 2018-03-09 平安科技(深圳)有限公司 Licence plate recognition method and device
CN107844794A (en) * 2016-09-21 2018-03-27 北京旷视科技有限公司 Image-recognizing method and device
CN108932533A (en) * 2018-07-12 2018-12-04 北京木瓜移动科技股份有限公司 Identification model construction method and device, character identifying method and device
CN109214386A (en) * 2018-09-14 2019-01-15 北京京东金融科技控股有限公司 Method and apparatus for generating image recognition model
WO2019071660A1 (en) * 2017-10-09 2019-04-18 平安科技(深圳)有限公司 Bill information identification method, electronic device, and readable storage medium
WO2019071662A1 (en) * 2017-10-09 2019-04-18 平安科技(深圳)有限公司 Electronic device, bill information identification method, and computer readable storage medium
CN110147785A (en) * 2018-03-29 2019-08-20 腾讯科技(深圳)有限公司 Image-recognizing method, relevant apparatus and equipment
CN110414519A (en) * 2019-06-27 2019-11-05 众安信息技术服务有限公司 A kind of recognition methods of picture character and its identification device
CN112906696A (en) * 2021-05-06 2021-06-04 北京惠朗时代科技有限公司 English image region identification method and device

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20070025137A (en) * 2005-08-31 2007-03-08 매그나칩 반도체 유한회사 Multi-frequency oscillator
CN104572892A (en) * 2014-12-24 2015-04-29 中国科学院自动化研究所 Text classification method based on cyclic convolution network

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR20070025137A (en) * 2005-08-31 2007-03-08 매그나칩 반도체 유한회사 Multi-frequency oscillator
CN104572892A (en) * 2014-12-24 2015-04-29 中国科学院自动化研究所 Text classification method based on cyclic convolution network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
丛爽: "《智能控制系统及其应用》", 31 August 2013 *
宣森炎等: "基于联合卷积和递归神经网络的交通标志识别", 《传感器与微系统》 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106446782A (en) * 2016-08-29 2017-02-22 北京小米移动软件有限公司 Image identification method and device
CN107844794A (en) * 2016-09-21 2018-03-27 北京旷视科技有限公司 Image-recognizing method and device
CN106570497A (en) * 2016-10-08 2017-04-19 中国科学院深圳先进技术研究院 Text detection method and device for scene image
CN106530569A (en) * 2016-10-17 2017-03-22 北京小米移动软件有限公司 Method and device for fire monitoring
CN107784303A (en) * 2016-12-15 2018-03-09 平安科技(深圳)有限公司 Licence plate recognition method and device
WO2019071660A1 (en) * 2017-10-09 2019-04-18 平安科技(深圳)有限公司 Bill information identification method, electronic device, and readable storage medium
WO2019071662A1 (en) * 2017-10-09 2019-04-18 平安科技(深圳)有限公司 Electronic device, bill information identification method, and computer readable storage medium
CN110147785A (en) * 2018-03-29 2019-08-20 腾讯科技(深圳)有限公司 Image-recognizing method, relevant apparatus and equipment
CN110147785B (en) * 2018-03-29 2023-01-10 腾讯科技(深圳)有限公司 Image recognition method, related device and equipment
CN108932533A (en) * 2018-07-12 2018-12-04 北京木瓜移动科技股份有限公司 Identification model construction method and device, character identifying method and device
CN109214386A (en) * 2018-09-14 2019-01-15 北京京东金融科技控股有限公司 Method and apparatus for generating image recognition model
CN110414519A (en) * 2019-06-27 2019-11-05 众安信息技术服务有限公司 A kind of recognition methods of picture character and its identification device
CN112906696A (en) * 2021-05-06 2021-06-04 北京惠朗时代科技有限公司 English image region identification method and device

Similar Documents

Publication Publication Date Title
CN105654127A (en) End-to-end-based picture character sequence continuous recognition method
CN105654135A (en) Image character sequence recognition system based on recurrent neural network
CN105654129A (en) Optical character sequence recognition method
CN105678292A (en) Complex optical text sequence identification system based on convolution and recurrent neural network
CN105678293A (en) Complex image and text sequence identification method based on CNN-RNN
CN105678300A (en) Complex image and text sequence identification method
Mathew et al. Docvqa: A dataset for vqa on document images
CN105654130A (en) Recurrent neural network-based complex image character sequence recognition system
CN109190722B (en) Font style migration transformation method based on Manchu character picture
CN104966097B (en) A kind of complex script recognition methods based on deep learning
CN108230339A (en) A kind of gastric cancer pathological section based on pseudo label iteration mark marks complementing method
CN111126386A (en) Sequence field adaptation method based on counterstudy in scene text recognition
CN108427953A (en) A kind of character recognition method and device
CN108764317A (en) A kind of residual error convolutional neural networks image classification method based on multichannel characteristic weighing
CN105045900A (en) Data extraction method and apparatus
CN106339753A (en) Method for effectively enhancing robustness of convolutional neural network
CN104008401A (en) Method and device for image character recognition
CN106980817A (en) A kind of terrified video frequency identifying method based on Caffe frameworks
CN105975497A (en) Automatic microblog topic recommendation method and device
CN112861864A (en) Topic entry method, topic entry device, electronic device and computer-readable storage medium
CN109766918A (en) Conspicuousness object detecting method based on the fusion of multi-level contextual information
CN108961270B (en) Bridge crack image segmentation model based on semantic segmentation
CN109508712A (en) A kind of Chinese written language recognition methods based on image
CN111881880A (en) Bill text recognition method based on novel network
CN117011638A (en) End-to-end image mask pre-training method and device

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160608

RJ01 Rejection of invention patent application after publication