CN105609107A - Text processing method and device based on voice identification - Google Patents

Text processing method and device based on voice identification Download PDF

Info

Publication number
CN105609107A
CN105609107A CN201510982716.0A CN201510982716A CN105609107A CN 105609107 A CN105609107 A CN 105609107A CN 201510982716 A CN201510982716 A CN 201510982716A CN 105609107 A CN105609107 A CN 105609107A
Authority
CN
China
Prior art keywords
word
text
lexeme
probability
mark
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201510982716.0A
Other languages
Chinese (zh)
Inventor
曹松军
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Qihoo Technology Co Ltd
Qizhi Software Beijing Co Ltd
Original Assignee
Beijing Qihoo Technology Co Ltd
Qizhi Software Beijing Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Qihoo Technology Co Ltd, Qizhi Software Beijing Co Ltd filed Critical Beijing Qihoo Technology Co Ltd
Priority to CN201510982716.0A priority Critical patent/CN105609107A/en
Publication of CN105609107A publication Critical patent/CN105609107A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/9032Query formulation
    • G06F16/90332Natural language query formulation or dialogue systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/903Querying
    • G06F16/90335Query processing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/211Syntactic parsing, e.g. based on context-free grammar [CFG] or unification grammars
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/205Parsing
    • G06F40/221Parsing markup language streams
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/279Recognition of textual entities

Abstract

The embodiment of the invention provides a text processing method and a device based on voice identification. The method comprises a step of obtaining a first text which is obtained by the voice identification of voice data, a step of punctuating the first text, and obtaining one or more text segments, and a step of adding punctuations to the one or more text segments, and forming a second text through combination. According to the embodiment of the invention, the automatic adding of the punctuations is realized, the manual positioning and punctuation adding of a user are avoided, and the convenience of voice input is improved greatly.

Description

A kind of text handling method and device based on speech recognition
Technical field
The present invention relates to speech recognition technology field, particularly relate to a kind of text place based on speech recognitionReason method and a kind of text processing apparatus based on speech recognition.
Background technology
At present, the fast development of mobile Internet has driven wide as the mobile device such as mobile phone, panel computerGeneral universal, and as one of natural mode of man-machine interaction most convenient on mobile device, phonetic entry just byGradually accepted by users.
Speech recognition is at present generally the extensive Speaker-independent continuous speech recognition of application, its objective is defeatedThe voice that enter are converted into word output, and the recognition result of general output is a series of text.
In the application scenarios such as input method, user needs manually to the text of identification is positioned and addedPunctuation mark, especially for long statement and mobile device, its screen is less, and the position of interface display is alsoLittle, the positioning action of punctuate, the interpolation of punctuation mark operate all very loaded down with trivial details, make the operation of phonetic entryVery loaded down with trivial details.
Summary of the invention
In view of the above problems, the present invention has been proposed to provide one to overcome the problems referred to above or at least part ofA kind of text handling method based on speech recognition of addressing the above problem and corresponding a kind of based on languageThe text processing apparatus of sound identification.
According to one aspect of the present invention, provide a kind of text handling method based on speech recognition, bagDraw together:
Obtain the first text that speech data is carried out to speech recognition acquisition;
Described the first text is made pauses in reading unpunctuated ancient writings, obtain one or more text fragments;
Described one or more text fragments are added to punctuation mark, be combined into the second text.
Alternatively, described described the first text is made pauses in reading unpunctuated ancient writings, obtain the step of one or more text fragmentsSuddenly comprise:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
Alternatively, the step that the lexeme of the described one or more words of described identification is put comprises:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
Alternatively, described according to the order of word, the probability of putting according to each lexeme of previous word, meterThe step that calculates the probability that each lexeme of a rear word puts comprises:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
Alternatively, described sequence labelling model is conditional random field models, based on training text and for instituteState the lexeme of the sign note in training text and put training generation, the punctuation mark of described training text is replaced.
Alternatively, described according to the backward of word, the lexeme according to a rear word based on probability mark is put, markOutpouring the step that the lexeme of previous word puts comprises:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
Alternatively, described lexeme is put and is comprised one or more in prefix, suffix, word, in monosyllabic word;
The step of making pauses in reading unpunctuated ancient writings in the word position of the described appointment at described the first text comprises:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
Alternatively, the described step that described one or more text fragments are added to punctuation mark comprises:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.
According to a further aspect in the invention, provide a kind of text processing apparatus based on speech recognition, bagDraw together:
The first text acquisition module, is suitable for obtaining the first text that speech data is carried out to speech recognition acquisition;
Punctuate module, is suitable for described the first text to make pauses in reading unpunctuated ancient writings, and obtains one or more text fragments;
Punctuation mark adds module, is suitable for described one or more text fragments to add punctuation mark, groupSynthetic the second text.
Alternatively, described punctuate module is also suitable for:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
Alternatively, described punctuate module is also suitable for:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
Alternatively, described punctuate module is also suitable for:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
Alternatively, described sequence labelling model is conditional random field models, based on training text and for instituteState the lexeme of the sign note in training text and put training generation, the punctuation mark of described training text is replaced.
Alternatively, described punctuate module is also suitable for:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
Alternatively, described lexeme is put and is comprised one or more in prefix, suffix, word, in monosyllabic word;
Described punctuate module is also suitable for:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
Alternatively, described punctuation mark interpolation module is also suitable for:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.
The embodiment of the present invention is carried out the result of speech recognition to speech data, the first text, makes pauses in reading unpunctuated ancient writings,Text fragments after punctuate is added to punctuation mark, be combined into the second text, realized punctuation markAutomatically add, avoid user manually to position, add punctuate, the phonetic entry of raising is easy greatlyProperty.
Above-mentioned explanation is only the general introduction of technical solution of the present invention, in order to better understand skill of the present inventionArt means, and can being implemented according to the content of description, and for allow of the present invention above-mentioned and otherObject, feature and advantage can become apparent, below especially exemplified by the specific embodiment of the present invention.
Brief description of the drawings
By reading below detailed description of the preferred embodiment, various other advantage and benefit for thisIt is cheer and bright that field those of ordinary skill will become. Accompanying drawing is only for the object of preferred embodiment is shown,And do not think limitation of the present invention. And in whole accompanying drawing, represent by identical reference symbolIdentical parts. In the accompanying drawings:
Fig. 1 shows a kind of according to an embodiment of the invention text handling method based on speech recognitionThe flow chart of steps of embodiment; And
Fig. 2 shows a kind of according to an embodiment of the invention text processing apparatus based on speech recognitionThe structured flowchart of embodiment.
Detailed description of the invention
Exemplary embodiment of the present disclosure is described below with reference to accompanying drawings in more detail. Although show in accompanying drawingExemplary embodiment of the present disclosure, but should be appreciated that and can realize the disclosure and not with various formsThe embodiment that should be set forth here limits. On the contrary, providing these embodiment is for can be more thoroughlyUnderstand the disclosure, and can be by the those skilled in the art that conveys to complete the scope of the present disclosure.
With reference to Fig. 1, show a kind of according to an embodiment of the invention text place based on speech recognitionThe flow chart of steps of reason embodiment of the method, specifically can comprise the steps:
Step 101, obtains the first text that speech data is carried out to speech recognition acquisition;
In specific implementation, user can be by disposing the electronic equipment input of the sound card equipment such as microphoneLanguage data.
Wherein, this electronic equipment can be mobile device, as mobile phone, panel computer, personal digital assistant,Wearing equipment (as glasses, wrist-watch etc.) etc. can be also permanent plant, as PC, intelligenceTV, Smart Home/household electrical appliances (as air-conditioning, electric cooker) etc., the embodiment of the present invention is not limit thisSystem.
In the time that electronic equipment receives the language data of sound card equipment conversion, can carry out in this locality voice knowledgeNot, add punctuation mark, also can be sent to server carry out speech recognition, add punctuation mark, thisInventive embodiments is not limited this.
In specific implementation, the speech recognition system of carrying out speech recognition can be based on WFST (WeightedFinite-stateTransducer, weighting FST) build network, conventionally by following baseThis module forms:
1, signal is processed and characteristic extracting module; The main task of this module is to extract from voice dataFeature, for acoustic model processing. Meanwhile, it has generally also comprised some signal processing technologies, to the greatest extent canCan reduce the impact that the factors such as ambient noise, channel, speaker cause feature.
2, acoustic model; Adopt of speech recognition system carried out modeling based on single order HMM more.
3, pronunciation dictionary; Pronunciation dictionary comprises the speech recognition system treatable word finder of institute and pronunciation thereof.The actual mapping that acoustic model and language model are provided of pronunciation dictionary.
4, language model; Language model to speech recognition system for language carry out modeling. TheoreticalUpper, comprise regular language, context-free grammar can be served as language mould at interior various language modelsType, but current various system N unit's syntax that generally adopt or based on statistics and variant thereof.
5, decoder; Decoder is one of core of speech recognition system, and its task is the letter to inputNumber, according to acoustics, language model and dictionary, searching can be exported with maximum probability the word string of this signal.Can more clearly understand the relation between above-mentioned module from mathematical angle.
In embodiments of the present invention, decoder can use GMM (GaussianMixtureModel,Gauss hybrid models) and DNN (DeepNeuralNetworks, degree of depth neural network model) trainingThe acoustic model going out.
Due to HMM (HiddenMarkovModel, hidden Markov model), model can be wellThe time variation of description voice and in short-term stationarity, be widely used in large vocabulary continuous speech recognition systemAcoustic Modeling.
In embodiments of the present invention, using syllogic half syllable as basic pronunciation unit, or claim phone, andUsing context-sensitive three-tone as Acoustic Modeling unit.
Each three-tone unit represents with the HMM of 5 states, and the 1st, the 5th stateFor non-emissive state, in the time of training and identification, do not take speech frame; 2nd, 3,4 states are transmitting state,When training and identification, at least take frame voice.
In the starting stage, adopting GMM is state modeling, and adopts based on maximum-likelihood criterionBaum-Welch algorithm iteration is optimized HMM and GMM parameter. When model parameter reaches the condition of convergenceOr reach predefined iterations, deconditioning. And by HMM-GMM system in voiceThe time boundary of HMM state carries out cutting.
Training when DNN, is characterized as input with speech frame, with state corresponding to current speech frame (byThe cutting of HMM-GMM system obtains) be output, maximizing under the object function of cross entropy, adoptGradient descent algorithm based on mini-batch is trained DNN parameter.
Step 102, makes pauses in reading unpunctuated ancient writings to described the first text, obtains one or more text fragments;
In embodiments of the present invention, can, to the result of speech recognition (i.e. the first text), carry out semantemeExcavating, makes pauses in reading unpunctuated ancient writings in the position of interrupting at semanteme.
In the optional embodiment of one of the present invention, step 102 can comprise following sub-step:
Sub-step S11, cuts word processing to described the first text, obtains one or more words;
In embodiments of the present invention, can carry out pretreatment to the first text, be cut into single character.
For example, if the first text of speech recognition is " your good Beijing ", can be cut into " you ", " good "," north ", " capital " is totally four words.
Sub-step S12, identifies the lexeme of described one or more words and puts;
Lexeme is put, and represents the position of this word in certain word, be specifically as follows comprise following a kind of orMultiple:
Prefix (begin, B), represents word first position at a word;
Suffix (end, E), represents word last position at a word;
In word (middle, M), represent that word is in the centre position of a word, non-first position, non-A rear position;
Monosyllabic word (single, S), represents that a word is independent word.
In embodiments of the present invention, can application conditions random field (ConditionalRandomField,CRF) lexeme of the one or more words after mark cutting is put.
CRF is a kind of discriminate probabilistic model, is the one of random field, can be for mark.
Suppose that P (Y|X) is linear chain condition random field, under the condition that is x in stochastic variable X value,Stochastic variable Y value is that the conditional probability of y has following form:
P ( y | x ) = 1 Z ( x ) exp ( Σ i , k λ k t k ( y i - 1 , y i , x , i ) + Σ i , 1 μ l s l ( y i , x , i ) )
Wherein,
Z ( x ) = Σ y exp ( Σ i , k λ k t k ( y i - 1 , y i , x , i ) + Σ i , l μ l s l ( y i , x , i ) )
In above-mentioned formula, tkWith sl be characteristic function, λkAnd μlCorresponding weights, Z (x) be standardization because ofSon is sued for peace on all possible output sequence.
Based on CRF punctuate, can adopt the label sets of 4-tag, i.e. BEMS, the feature templates of designAs follows:
#Unigram
U00:%x[-2,0]
U01:%x[-1,0]
U02:%x[0,0]
U03:%x[1,0]
U04:%x[2,0]
U05:%x[-2,0]/%x[-1,0]/%x[0,0]
U06:%x[-1,0]/%x[0,0]/%x[1,0]
U07:%x[0,0]/%x[1,0]/%x[2,0]
U08:%x[-1,0]/%x[0,0]
U09:%x[0,0]/%x[1,0]
#Bigram
B
Wherein Unigram is corresponding to status flag sl, current word x, characteristic function need to be considered front and back twoIndividual word; Bigram is corresponding to transfer characteristic tk, for convenience of calculation, set transition probability herein for instituteSome list entries x are all the same.
According to above characteristic Design, can use improved iteration method of scales to carry out CRF training, instructionPractise sequence labelling model, sequence labelling model is conditional random field models, based on training text and pinThe lexeme of the sign note in described training text is put to training to be generated.
In order to meet the use scenes on line, corpus can adopt the mark of speech recognition training language materialData, data volume can reach totally 80 ten thousand statements.
In the time of training, first can carry out pretreatment by training text, by the punctuate symbol occurring in training textNumber with space replace, the punctuation mark of training text is replaced.
For example, to training text "! Alright, say again next time. " in punctuation mark replace,Obtain " uh good next time say again ".
Then use BEMS to mark each word in training text, for example, to training textThe mark of " uh good next time say again " is as follows:
Uh S
Good B
E
Lower B
Inferior M
M again
Say E
The training text that region marks, can carry out the training of CRF model.
The model finally obtaining mainly comprises the content of two aspects, is the Chinese that characteristic function F is corresponding on the one handWord sequence is the weight w that characteristic function is corresponding on the other hand.
In the optional embodiment of one of the present invention, sub-step S12 can comprise following sub-step:
Sub-step S121, according to the order of word, the probability of putting according to each lexeme of previous word,Calculate the probability that each lexeme of a rear word is put;
In embodiments of the present invention, the lexeme that can calculate in the word sequence of input by the mode of recursion is putProbability.
If application sequence marking model, can input model characteristic vector F (y, x) and weight vectors w,Word sequence (being the sequence that the word after cutting forms) x=(x1,x2,…,xn)。
In the time of calculating probability, can first initialize:
δ1(j)=w*F1(y0=start,y1=j,x),j=1,2,…,m
Wherein, x is the word sequence of input, and y is the mark sequence that the first text is corresponding, w be weight toAmount, m is the number of mark sequence, such as m corresponding to BEMS is that 4, i is the i in word sequenceIndividual word, j is j mark.
Particularly, for the first character of word sequence, by default sequence labelling model, according to spyLevy function F and corresponding weight w thereof, calculate the probability δ that each lexeme of the 1st word is put1(j)。
Each lexeme of i the word of probability recursion of then, putting by each lexeme of i-1 word is putProbability, record the historical record Ψ of optimal path simultaneouslyi, i is greater than 1 positive integer, i.e. i=2, and 3 ..., n.
δ i ( l ) = m a x 1 ≤ j ≤ m { δ i - 1 ( j ) + w * F i ( y i - 1 = j , y i = l , x ) } , l = 1 , 2 , ... , m
Ψ i ( l ) = argmax 1 ≤ j ≤ m { δ i - 1 ( j ) + w * F i ( y i - 1 = j , y i = l , x ) } , l = 1 , 2 , ... , m
Wherein, x is the word sequence of input, and y is the mark sequence that the first text is corresponding, w be weight toAmount, m is the number of mark sequence, such as m corresponding to BEMS is that 4, i is the i in word sequenceIndividual word, j is j mark.
Particularly, by default sequence labelling model, according to feature F and weight w, based on i-1The probability δ that each lexeme of individual word is puti-1, calculate the probability δ that each lexeme of i word is puti(j)。
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
For example, for " you ", " good ", " north ", " capital " these four words, can first calculate " you " and belong toIn the probability of BEMS.
The probability recursion " good " that belongs to B based on " you " belongs to the probability of BEMS;
The probability recursion " good " that belongs to E based on " you " belongs to the probability of BEMS;
The probability recursion " good " that belongs to M based on " you " belongs to the probability of BEMS;
The probability recursion " good " that belongs to S based on " you " belongs to the probability of BEMS;
For " good ", its probability that has that belongs to B has four, is respectively δB-B、δE-B、δM-B、δS-B,The subscript of probability represents the path of its recursion, and the highest probability of selected value from these four probability, as" good " belongs to the probability of B eventually, and its probability that belongs to EMS is also processed equally.
The probability recursion " north " that belongs to BEMS based on " good " equally, respectively belongs to the general of BEMSRate, the probability recursion " capital " that belongs to BEMS based on " north " belongs to the probability of BEMS, until passPush away.
Sub-step S122, according to the backward of word, the lexeme according to a rear word based on probability mark is put,The lexeme that marks out previous word is put.
When all words of the first text all calculating probability complete after, get the corresponding all marks of the last characterIn optimal value as last optimal value.
m a x y ( w * F ( y , x ) ) = m a x 1 ≤ j ≤ m δ n ( j )
y n * = argmax 1 ≤ j ≤ m δ n ( j )
Wherein, x is the word sequence of input, and y is the mark sequence that the first text is corresponding, w be weight toAmount, m is the number of mark sequence, such as m corresponding to BEMS is that 4, i is the i in word sequenceIndividual word, j is j mark.
Particularly, for the last character, the affiliated lexeme of probability that mark value is the highest is put.
For example, for " you ", " good ", " north ", " capital " these four words, " capital " is the last character,Belong to these four probability of BEMS from " capital ", the probability that selective value is the highest, marks its lexeme and puts,As E.
Recall and obtain final mark path according to last optimal value, complete last mark, try to achieveShortest path y * = ( y 1 * , y 2 * , ... , y n * ) .
y i * = Ψ i + 1 ( y i + 1 * )
Wherein, x is the word sequence of input, and y is the mark sequence that the first text is corresponding, and i is greater than 1Positive integer, i.e. i=n-1, n-2 ..., 1.
Particularly, in the time that i word determines that lexeme is put, query count goes out the general of i words positionThe probability of i-1 word of rate;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
For example, for " you ", " good ", " north ", " capital " these four words, " capital " has marked lexeme and has putFor E.
Search the formerly historical record Ψ in the path of recordi, be δB-E, " capital " belongs to the probability of E,Be the probability recursion that is belonged to B by previous word " north ", can be to capital " mark lexeme is set to B,Also process equally for " good ", " north ", finally complete " your good Beijing " is marked:
You are B
Good E
North B
Capital E
Sub-step S13, makes pauses in reading unpunctuated ancient writings in the word position of the appointment of described the first text, obtain one orMultiple text fragments.
In specific implementation, lexeme is put and is comprised one or more in prefix, suffix, word, in monosyllabic word.
Therefore,, in the time of punctuate, can before the prefix of the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings, alsoCan after the suffix of the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
It should be noted that, when prefix and/or monosyllabic word are the first character in the first text, before itPunctuate is insignificant, and suffix and/or monosyllabic word are when the rear word in the first text, after itPunctuate is insignificant, can ignore.
In addition, if the punctuate carrying out after suffix and/or monosyllabic word with before prefix and/or monosyllabic wordMake pauses in reading unpunctuated ancient writings overlapping, can once make pauses in reading unpunctuated ancient writings.
For example, for the first text " your good Beijing ", can make pauses in reading unpunctuated ancient writings afterwards at suffix " good ", obtainObtain text fragments " hello ", " Beijing ".
Step 103, adds punctuation mark to described one or more text fragments, is combined into the second text.
The application embodiment of the present invention, the mapping that can set up in advance punctuation mark and one or more keywordsRelation.
The example of these mapping relations is as shown in the table:
Punctuation mark Keyword
What, how, where,, why, who
Oh, eh,
Other (i.e. words of other keyword in mapping relations)
Therefore, for each text fragments, can identify keyword, search the punctuate that keyword is correspondingSymbol adds described punctuation mark after text fragments.
For example, for text fragments " hello ", " Beijing ", can add afterwards ", " " hello ",Text fragments and punctuation mark are combined into the second text " hello, Beijing ".
If make pauses in reading unpunctuated ancient writings, add punctuate in electronic equipment this locality, can directly show second after combinationText, shows user.
If make pauses in reading unpunctuated ancient writings, add punctuate at server, can be by the second text return electron after combinationEquipment shows.
The embodiment of the present invention is carried out the result of speech recognition to speech data, the first text, makes pauses in reading unpunctuated ancient writings,Text fragments after punctuate is added to punctuation mark, be combined into the second text, realized punctuation markAutomatically add, avoid user manually to position, add punctuate, the phonetic entry of raising is easy greatlyProperty.
For embodiment of the method, for simple description, thus it is all expressed as to a series of combination of actions,But those skilled in the art should know, the embodiment of the present invention is not subject to the limit of described sequence of movementSystem, because according to the embodiment of the present invention, some step can adopt other orders or carry out simultaneously. ItsInferior, those skilled in the art also should know, the embodiment described in description all belongs to preferred enforcementExample, related action might not be that the embodiment of the present invention is necessary.
With reference to Fig. 2, show a kind of according to an embodiment of the invention text place based on speech recognitionThe structured flowchart of reason device embodiment, specifically can comprise as lower module:
The first text acquisition module 201, is suitable for obtaining speech data is carried out to first of speech recognition acquisitionText;
Punctuate module 202, is suitable for described the first text to make pauses in reading unpunctuated ancient writings, and obtains one or more text sheetsSection;
Punctuation mark adds module 203, is suitable for described one or more text fragments to add punctuation mark,Be combined into the second text.
In the optional embodiment of one of the present invention, described punctuate module 202 can also be suitable for:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
In the optional embodiment of one of the present invention, described punctuate module 202 can also be suitable for:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
In the optional embodiment of one of the present invention, described punctuate module 202 can also be suitable for:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
In specific implementation, described sequence labelling model is conditional random field models, based on training text andPut training for the lexeme of the sign note in described training text and generate, the punctuation mark of described training textBe replaced.
In the optional embodiment of one of the present invention, described punctuate module 202 can also be suitable for:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
In the optional embodiment of one of the present invention, described lexeme put comprise in prefix, suffix, word, singleOne or more in words;
Described punctuate module 202 can also be suitable for:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
In the optional embodiment of one of the present invention, described punctuation mark adds module 203 and can also be suitable for:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.
For device embodiment, because it is substantially similar to embodiment of the method, so the comparison of describingSimply, relevant part is referring to the part explanation of embodiment of the method.
The algorithm providing at this and demonstration are solid with any certain computer, virtual system or miscellaneous equipmentHave relevant. Various general-purpose systems also can with based on using together with this teaching. According to description above,It is apparent constructing the desired structure of this type systematic. In addition, the present invention is not also for any specificProgramming language. It should be understood that and can utilize various programming languages to realize content of the present invention described here,And the description of above language-specific being done is in order to disclose preferred forms of the present invention.
In the description that provided herein, a large amount of details are described. But, can understand, thisInventive embodiment can be put into practice in the situation that there is no these details. In some instances, notBe shown specifically known method, structure and technology, so that not fuzzy understanding of this description.
Similarly, should be appreciated that in order to simplify the disclosure and to help to understand in each inventive aspectOr multiple, in the above in the description of exemplary embodiment of the present invention, each feature of the present invention is sometimesBe grouped together into single embodiment, figure or in its description. But, should be by the disclosureMethod be construed to the following intention of reflection: the present invention for required protection requires than in each claimThe more feature of the middle feature of clearly recording. Or rather, as claims below reflectLike that, inventive aspect is to be less than all features of disclosed single embodiment above. Therefore, followClaims of detailed description of the invention are incorporated to this detailed description of the invention, wherein each right thus clearlyRequirement itself is all as independent embodiment of the present invention.
Those skilled in the art are appreciated that and can carry out certainly the module in the equipment in embodimentChange adaptively and they are arranged in one or more equipment different from this embodiment. CanModule in embodiment or unit or assembly are combined into a module or unit or assembly, and in addition canTo put them into multiple submodules or subelement or sub-component. Except such feature and/or process orAt least some in unit are, outside mutually repelling, can adopt any combination (to comprise companion to this descriptionWith claim, summary and accompanying drawing) in disclosed all features and so disclosed any method orAll processes or the unit of person's equipment combine. Unless clearly statement in addition, this description (comprises companionWith claim, summary and accompanying drawing) in disclosed each feature can be by providing identical, being equal to or phaseAlternative features like object replaces.
In addition, although those skilled in the art will appreciate that embodiment more described herein comprise itIncluded some feature instead of further feature in its embodiment, but the group of the feature of different embodimentClose and mean within scope of the present invention and form different embodiment. For example, power belowIn profit claim, the one of any of embodiment required for protection can make with combination arbitrarilyWith.
All parts embodiment of the present invention can realize with hardware, or with in one or more processingThe software module of moving on device realizes, or realizes with their combination. Those skilled in the art shouldUnderstand, can use in practice microprocessor or digital signal processor (DSP) to realize basisOne of some or all parts in the text-processing equipment based on speech recognition of the embodiment of the present inventionA little or repertoire. The present invention can also be embodied as the part for carrying out method as described hereinOr whole equipment or device program (for example, computer program and computer program). ThisThe realizing program of the present invention and can be stored on computer-readable medium of sample, or can have one orThe form of the multiple signals of person. Such signal can be downloaded and obtain from internet website, or at carrierOn signal, provide, or provide with any other form.
It should be noted above-described embodiment the present invention will be described instead of limit the invention,And those skilled in the art can design replacement in the case of not departing from the scope of claimsEmbodiment. In the claims, any reference symbol between bracket should be configured to rightThe restriction requiring. Word " comprises " not to be got rid of existence and is not listed as element or step in the claims. PositionWord " one " before the element or " one " do not get rid of and have multiple such elements. The present invention canWith by means of including the hardware of some different elements and realizing by means of the computer of suitably programming.In the unit claim of having enumerated some devices, several in these devices can be by sameIndividual hardware branch carrys out imbody. The use of word first, second and C grade does not represent any order.Can be title by these word explanations.
The embodiment of the invention discloses A1, a kind of text handling method based on speech recognition, comprising:
Obtain the first text that speech data is carried out to speech recognition acquisition;
Described the first text is made pauses in reading unpunctuated ancient writings, obtain one or more text fragments;
Described one or more text fragments are added to punctuation mark, be combined into the second text.
A2, method as described in A1, describedly make pauses in reading unpunctuated ancient writings to described the first text, obtains one or manyThe step of individual text fragments comprises:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
A3, method as described in A2, the step bag that the lexeme of the described one or more words of described identification is putDraw together:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
A4, method as described in A3, described according to the order of word, according to each word of previous wordThe probability of position, the step that calculates the probability that each lexeme of a rear word puts comprises:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
A5, method as described in A4, described sequence labelling model is conditional random field models, based on instructionPractice text and put training for the lexeme of the sign note in described training text generating, described training textPunctuation mark is replaced.
A6, method as described in A3 or A4 or A5, described according to the backward of word, according to latter oneThe lexeme of word based on probability mark put, and marks out the step that the lexeme of previous word puts and comprises:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
A7, method as described in A2 or A3 or A4 or A5, described lexeme put comprise prefix, suffix,One or more in word, in monosyllabic word;
The step of making pauses in reading unpunctuated ancient writings in the word position of the described appointment at described the first text comprises:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
A8, method as described in A1 or A2 or A3 or A4 or A5, described to described one or manyThe step that individual text fragments adds punctuation mark comprises:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.
The embodiment of the invention also discloses B9, a kind of text processing apparatus based on speech recognition, comprising:
The first text acquisition module, is suitable for obtaining the first text that speech data is carried out to speech recognition acquisition;
Punctuate module, is suitable for described the first text to make pauses in reading unpunctuated ancient writings, and obtains one or more text fragments;
Punctuation mark adds module, is suitable for described one or more text fragments to add punctuation mark, groupSynthetic the second text.
B10, device as described in B9, described punctuate module is also suitable for:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
B11, device as described in B10, described punctuate module is also suitable for:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
B12, device as described in B11, described punctuate module is also suitable for:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
B13, device as described in B11, described sequence labelling model is conditional random field models, based onTraining text and the lexeme of noting for the sign in described training text are put training and are generated, described training textPunctuation mark be replaced.
B14, device as described in B11 or B12 or B13, described punctuate module is also suitable for:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
B15, device as described in B10 or B11 or B12 or B13, described lexeme put comprise prefix,One or more in suffix, word, in monosyllabic word;
Described punctuate module is also suitable for:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
B16, device as described in B9 or B10 or B11 or B12 or B13, described punctuation mark addsAdding module is also suitable for:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.

Claims (10)

1. the text handling method based on speech recognition, comprising:
Obtain the first text that speech data is carried out to speech recognition acquisition;
Described the first text is made pauses in reading unpunctuated ancient writings, obtain one or more text fragments;
Described one or more text fragments are added to punctuation mark, be combined into the second text.
2. the method for claim 1, is characterized in that, described described the first text is carried outPunctuate, the step that obtains one or more text fragments comprises:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
3. method as claimed in claim 2, is characterized in that, described identification is described one or moreThe step that the lexeme of word is put comprises:
According to the order of word, the probability of putting according to each lexeme of previous word, calculates a rear wordThe probability put of each lexeme;
According to the backward of word, the lexeme according to a rear word based on probability mark is put, and marks out previous wordLexeme put.
4. method as claimed in claim 3, is characterized in that, described according to the order of word, according toThe probability that each lexeme of previous word is put, calculates probability that each lexeme of a rear word putsStep comprises:
By default sequence labelling model, calculate the probability that each lexeme of the 1st word is put;
By default sequence labelling model, the probability that the each lexeme based on i-1 word is put, calculatesGo out the probability that each lexeme of i word is put, i is greater than 1 positive integer;
The probability of putting for each lexeme of i word, to be worth, the highest probability puts as described lexemeProbability.
5. method as claimed in claim 4, is characterized in that, described sequence labelling model is conditionRandom field models, puts training life based on training text and for the lexeme of the sign note in described training textBecome, the punctuation mark of described training text is replaced.
6. the method as described in claim 3 or 4 or 5, is characterized in that, described contrary according to wordOrder, the lexeme according to a rear word based on probability mark is put, and marks out the step that the lexeme of previous word is putComprise:
For the last character, the affiliated lexeme of probability that mark value is the highest is put;
In the time that i word determines that lexeme is put, query count goes out the i-1 of the probability of i words positionThe probability of word, i is greater than 1 positive integer;
For i-1 word, the affiliated lexeme of probability that marks described i-1 word is put.
7. the method as described in claim 2 or 3 or 4 or 5, is characterized in that, described lexeme is putComprise one or more in prefix, suffix, word, in monosyllabic word;
The step of making pauses in reading unpunctuated ancient writings in the word position of the described appointment at described the first text comprises:
Before the prefix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings;
And/or,
After the suffix of described the first text and/or monosyllabic word, make pauses in reading unpunctuated ancient writings.
8. the method as described in claim 1 or 2 or 3 or 4 or 5, is characterized in that, described rightThe step that described one or more text fragments adds punctuation mark comprises:
For each text fragments, identify keyword;
Search the punctuation mark that described keyword is corresponding;
After described text fragments, add described punctuation mark.
9. the text processing apparatus based on speech recognition, comprising:
The first text acquisition module, is suitable for obtaining the first text that speech data is carried out to speech recognition acquisition;
Punctuate module, is suitable for described the first text to make pauses in reading unpunctuated ancient writings, and obtains one or more text fragments;
Punctuation mark adds module, is suitable for described one or more text fragments to add punctuation mark, groupSynthetic the second text.
10. device as claimed in claim 9, is characterized in that, described punctuate module is also suitable for:
Described the first text is cut to word processing, obtain one or more words;
Identifying the lexeme of described one or more words puts;
Making pauses in reading unpunctuated ancient writings in word position in the appointment of described the first text, obtains one or more text fragments.
CN201510982716.0A 2015-12-23 2015-12-23 Text processing method and device based on voice identification Pending CN105609107A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201510982716.0A CN105609107A (en) 2015-12-23 2015-12-23 Text processing method and device based on voice identification

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201510982716.0A CN105609107A (en) 2015-12-23 2015-12-23 Text processing method and device based on voice identification

Publications (1)

Publication Number Publication Date
CN105609107A true CN105609107A (en) 2016-05-25

Family

ID=55988991

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201510982716.0A Pending CN105609107A (en) 2015-12-23 2015-12-23 Text processing method and device based on voice identification

Country Status (1)

Country Link
CN (1) CN105609107A (en)

Cited By (31)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106356054A (en) * 2016-11-23 2017-01-25 广西大学 Method and system for collecting information of agricultural products based on voice recognition
CN106886364A (en) * 2017-02-14 2017-06-23 深圳市金立通信设备有限公司 A kind of text handling method and terminal based on speech recognition
CN107066477A (en) * 2016-12-13 2017-08-18 合网络技术(北京)有限公司 A kind of method and device of intelligent recommendation video
CN107221330A (en) * 2017-05-26 2017-09-29 北京搜狗科技发展有限公司 Punctuate adding method and device, the device added for punctuate
CN107247706A (en) * 2017-06-16 2017-10-13 中国电子技术标准化研究院 Text punctuate method for establishing model, punctuate method, device and computer equipment
CN107247700A (en) * 2017-04-27 2017-10-13 北京捷通华声科技股份有限公司 A kind of method and device for adding text marking
CN107424612A (en) * 2017-07-28 2017-12-01 北京搜狗科技发展有限公司 Processing method, device and machine readable media
CN107679033A (en) * 2017-09-11 2018-02-09 百度在线网络技术(北京)有限公司 Text punctuate location recognition method and device
CN107767870A (en) * 2017-09-29 2018-03-06 百度在线网络技术(北京)有限公司 Adding method, device and the computer equipment of punctuation mark
CN108090038A (en) * 2016-11-11 2018-05-29 科大讯飞股份有限公司 Text punctuate method and system
CN108268429A (en) * 2017-06-15 2018-07-10 广东神马搜索科技有限公司 The determining method and apparatus of online literature chapters and sections
CN108597517A (en) * 2018-03-08 2018-09-28 深圳市声扬科技有限公司 Punctuation mark adding method, device, computer equipment and storage medium
CN108628813A (en) * 2017-03-17 2018-10-09 北京搜狗科技发展有限公司 Treating method and apparatus, the device for processing
CN108694939A (en) * 2018-05-23 2018-10-23 广州视源电子科技股份有限公司 Phonetic search optimization method, device and system
CN109145282A (en) * 2017-06-16 2019-01-04 贵州小爱机器人科技有限公司 Punctuate model training method, punctuate method, apparatus and computer equipment
CN109274845A (en) * 2018-08-31 2019-01-25 平安科技(深圳)有限公司 Intelligent sound pays a return visit method, apparatus, computer equipment and storage medium automatically
CN109558576A (en) * 2018-11-05 2019-04-02 中山大学 A kind of punctuation mark prediction technique based on from attention mechanism
CN109614627A (en) * 2019-01-04 2019-04-12 平安科技(深圳)有限公司 A kind of text punctuate prediction technique, device, computer equipment and storage medium
CN109708256A (en) * 2018-12-06 2019-05-03 珠海格力电器股份有限公司 A kind of voice determines method, apparatus, storage medium and air-conditioning
CN109754808A (en) * 2018-12-13 2019-05-14 平安科技(深圳)有限公司 Method, apparatus, computer equipment and the storage medium of voice conversion text
CN109783648A (en) * 2018-12-28 2019-05-21 北京声智科技有限公司 A method of ASR language model is improved using ASR recognition result
CN109979435A (en) * 2017-12-28 2019-07-05 北京搜狗科技发展有限公司 Data processing method and device, the device for data processing
CN110399878A (en) * 2019-06-14 2019-11-01 南京火眼锐视信息科技有限公司 Table format restoration methods, computer-readable medium and computer
CN110599028A (en) * 2019-09-09 2019-12-20 深圳前海微众银行股份有限公司 Text positioning method, device, equipment and storage medium
CN111326160A (en) * 2020-03-11 2020-06-23 南京奥拓电子科技有限公司 Speech recognition method, system and storage medium for correcting noise text
CN111785259A (en) * 2019-04-04 2020-10-16 北京猎户星空科技有限公司 Information processing method and device and electronic equipment
CN111985208A (en) * 2020-08-18 2020-11-24 沈阳东软智能医疗科技研究院有限公司 Method, device and equipment for realizing punctuation mark filling
CN112036174A (en) * 2019-05-15 2020-12-04 南京大学 Punctuation marking method and device
CN112151019A (en) * 2019-06-26 2020-12-29 阿里巴巴集团控股有限公司 Text processing method and device and computing equipment
CN112634876A (en) * 2021-01-04 2021-04-09 北京有竹居网络技术有限公司 Voice recognition method, voice recognition device, storage medium and electronic equipment
CN112712794A (en) * 2020-12-25 2021-04-27 苏州思必驰信息科技有限公司 Speech recognition marking training combined system and device

Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6067514A (en) * 1998-06-23 2000-05-23 International Business Machines Corporation Method for automatically punctuating a speech utterance in a continuous speech recognition system
CN101876965A (en) * 2009-04-30 2010-11-03 国际商业机器公司 Method and system used for processing text
CN102231278A (en) * 2011-06-10 2011-11-02 安徽科大讯飞信息科技股份有限公司 Method and system for realizing automatic addition of punctuation marks in speech recognition
CN103020034A (en) * 2011-09-26 2013-04-03 北京大学 Chinese words segmentation method and device
CN103077164A (en) * 2012-12-27 2013-05-01 新浪网技术(中国)有限公司 Text analysis method and text analyzer
CN103971684A (en) * 2013-01-29 2014-08-06 腾讯科技(深圳)有限公司 Method and system for adding punctuations and method and device for establishing language model for adding punctuations
CN104142915A (en) * 2013-05-24 2014-11-12 腾讯科技(深圳)有限公司 Punctuation adding method and system

Patent Citations (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6067514A (en) * 1998-06-23 2000-05-23 International Business Machines Corporation Method for automatically punctuating a speech utterance in a continuous speech recognition system
CN101876965A (en) * 2009-04-30 2010-11-03 国际商业机器公司 Method and system used for processing text
CN102231278A (en) * 2011-06-10 2011-11-02 安徽科大讯飞信息科技股份有限公司 Method and system for realizing automatic addition of punctuation marks in speech recognition
CN103020034A (en) * 2011-09-26 2013-04-03 北京大学 Chinese words segmentation method and device
CN103077164A (en) * 2012-12-27 2013-05-01 新浪网技术(中国)有限公司 Text analysis method and text analyzer
CN103971684A (en) * 2013-01-29 2014-08-06 腾讯科技(深圳)有限公司 Method and system for adding punctuations and method and device for establishing language model for adding punctuations
CN104142915A (en) * 2013-05-24 2014-11-12 腾讯科技(深圳)有限公司 Punctuation adding method and system

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
张合 等: ""一种基于层叠CRF的古文断句与句读标记方法"", 《计算机应用研究》 *
李航: "《统计学习方法》", 31 March 2012, 清华大学出版社 *
赵海 等: ""基于有效子串标注的中文分词"", 《中文信息学报》 *

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108090038A (en) * 2016-11-11 2018-05-29 科大讯飞股份有限公司 Text punctuate method and system
CN108090038B (en) * 2016-11-11 2022-01-14 科大讯飞股份有限公司 Text sentence-breaking method and system
CN106356054A (en) * 2016-11-23 2017-01-25 广西大学 Method and system for collecting information of agricultural products based on voice recognition
CN107066477A (en) * 2016-12-13 2017-08-18 合网络技术(北京)有限公司 A kind of method and device of intelligent recommendation video
CN106886364A (en) * 2017-02-14 2017-06-23 深圳市金立通信设备有限公司 A kind of text handling method and terminal based on speech recognition
CN108628813B (en) * 2017-03-17 2022-09-23 北京搜狗科技发展有限公司 Processing method and device for processing
CN108628813A (en) * 2017-03-17 2018-10-09 北京搜狗科技发展有限公司 Treating method and apparatus, the device for processing
CN107247700A (en) * 2017-04-27 2017-10-13 北京捷通华声科技股份有限公司 A kind of method and device for adding text marking
CN107221330A (en) * 2017-05-26 2017-09-29 北京搜狗科技发展有限公司 Punctuate adding method and device, the device added for punctuate
CN108268429A (en) * 2017-06-15 2018-07-10 广东神马搜索科技有限公司 The determining method and apparatus of online literature chapters and sections
CN108268429B (en) * 2017-06-15 2021-08-06 阿里巴巴(中国)有限公司 Method and device for determining network literature chapters
CN107247706A (en) * 2017-06-16 2017-10-13 中国电子技术标准化研究院 Text punctuate method for establishing model, punctuate method, device and computer equipment
CN107247706B (en) * 2017-06-16 2021-06-25 中国电子技术标准化研究院 Text sentence-breaking model establishing method, sentence-breaking method, device and computer equipment
CN109145282B (en) * 2017-06-16 2023-11-07 贵州小爱机器人科技有限公司 Sentence-breaking model training method, sentence-breaking device and computer equipment
CN109145282A (en) * 2017-06-16 2019-01-04 贵州小爱机器人科技有限公司 Punctuate model training method, punctuate method, apparatus and computer equipment
CN107424612A (en) * 2017-07-28 2017-12-01 北京搜狗科技发展有限公司 Processing method, device and machine readable media
US11004448B2 (en) 2017-09-11 2021-05-11 Baidu Online Network Technology (Beijing) Co., Ltd. Method and device for recognizing text segmentation position
CN107679033B (en) * 2017-09-11 2021-12-14 百度在线网络技术(北京)有限公司 Text sentence break position identification method and device
CN107679033A (en) * 2017-09-11 2018-02-09 百度在线网络技术(北京)有限公司 Text punctuate location recognition method and device
CN107767870A (en) * 2017-09-29 2018-03-06 百度在线网络技术(北京)有限公司 Adding method, device and the computer equipment of punctuation mark
CN109979435B (en) * 2017-12-28 2021-10-22 北京搜狗科技发展有限公司 Data processing method and device for data processing
CN109979435A (en) * 2017-12-28 2019-07-05 北京搜狗科技发展有限公司 Data processing method and device, the device for data processing
CN108597517A (en) * 2018-03-08 2018-09-28 深圳市声扬科技有限公司 Punctuation mark adding method, device, computer equipment and storage medium
CN108597517B (en) * 2018-03-08 2020-06-05 深圳市声扬科技有限公司 Punctuation mark adding method and device, computer equipment and storage medium
CN108694939A (en) * 2018-05-23 2018-10-23 广州视源电子科技股份有限公司 Phonetic search optimization method, device and system
CN109274845A (en) * 2018-08-31 2019-01-25 平安科技(深圳)有限公司 Intelligent sound pays a return visit method, apparatus, computer equipment and storage medium automatically
CN109558576B (en) * 2018-11-05 2023-05-23 中山大学 Punctuation mark prediction method based on self-attention mechanism
CN109558576A (en) * 2018-11-05 2019-04-02 中山大学 A kind of punctuation mark prediction technique based on from attention mechanism
CN109708256B (en) * 2018-12-06 2020-07-03 珠海格力电器股份有限公司 Voice determination method and device, storage medium and air conditioner
CN109708256A (en) * 2018-12-06 2019-05-03 珠海格力电器股份有限公司 A kind of voice determines method, apparatus, storage medium and air-conditioning
CN109754808A (en) * 2018-12-13 2019-05-14 平安科技(深圳)有限公司 Method, apparatus, computer equipment and the storage medium of voice conversion text
CN109754808B (en) * 2018-12-13 2024-02-13 平安科技(深圳)有限公司 Method, device, computer equipment and storage medium for converting voice into text
CN109783648A (en) * 2018-12-28 2019-05-21 北京声智科技有限公司 A method of ASR language model is improved using ASR recognition result
CN109783648B (en) * 2018-12-28 2020-12-29 北京声智科技有限公司 Method for improving ASR language model by using ASR recognition result
CN109614627B (en) * 2019-01-04 2023-01-20 平安科技(深圳)有限公司 Text punctuation prediction method and device, computer equipment and storage medium
CN109614627A (en) * 2019-01-04 2019-04-12 平安科技(深圳)有限公司 A kind of text punctuate prediction technique, device, computer equipment and storage medium
CN111785259A (en) * 2019-04-04 2020-10-16 北京猎户星空科技有限公司 Information processing method and device and electronic equipment
CN112036174B (en) * 2019-05-15 2023-11-07 南京大学 Punctuation marking method and device
CN112036174A (en) * 2019-05-15 2020-12-04 南京大学 Punctuation marking method and device
CN110399878A (en) * 2019-06-14 2019-11-01 南京火眼锐视信息科技有限公司 Table format restoration methods, computer-readable medium and computer
CN110399878B (en) * 2019-06-14 2023-05-26 南京火眼锐视信息科技有限公司 Form format recovery method, computer readable medium and computer
CN112151019A (en) * 2019-06-26 2020-12-29 阿里巴巴集团控股有限公司 Text processing method and device and computing equipment
CN110599028B (en) * 2019-09-09 2022-05-17 深圳前海微众银行股份有限公司 Text positioning method, device, equipment and storage medium
CN110599028A (en) * 2019-09-09 2019-12-20 深圳前海微众银行股份有限公司 Text positioning method, device, equipment and storage medium
CN111326160A (en) * 2020-03-11 2020-06-23 南京奥拓电子科技有限公司 Speech recognition method, system and storage medium for correcting noise text
CN111985208A (en) * 2020-08-18 2020-11-24 沈阳东软智能医疗科技研究院有限公司 Method, device and equipment for realizing punctuation mark filling
CN111985208B (en) * 2020-08-18 2024-03-26 沈阳东软智能医疗科技研究院有限公司 Method, device and equipment for realizing punctuation mark filling
CN112712794A (en) * 2020-12-25 2021-04-27 苏州思必驰信息科技有限公司 Speech recognition marking training combined system and device
CN112634876A (en) * 2021-01-04 2021-04-09 北京有竹居网络技术有限公司 Voice recognition method, voice recognition device, storage medium and electronic equipment
CN112634876B (en) * 2021-01-04 2023-11-10 北京有竹居网络技术有限公司 Speech recognition method, device, storage medium and electronic equipment

Similar Documents

Publication Publication Date Title
CN105609107A (en) Text processing method and device based on voice identification
CN108417210B (en) Word embedding language model training method, word recognition method and system
CN110210029B (en) Method, system, device and medium for correcting error of voice text based on vertical field
CN106598939B (en) A kind of text error correction method and device, server, storage medium
Jung et al. An english to korean transliteration model of extended markov window
EP2486470B1 (en) System and method for inputting text into electronic devices
CN110797016B (en) Voice recognition method and device, electronic equipment and storage medium
CN108140019B (en) Language model generation device, language model generation method, and recording medium
WO2020001458A1 (en) Speech recognition method, device, and system
CN103869998B (en) A kind of method and device being ranked up to candidate item caused by input method
CN111739514B (en) Voice recognition method, device, equipment and medium
CN106935239A (en) The construction method and device of a kind of pronunciation dictionary
CN103903619A (en) Method and system for improving accuracy of speech recognition
CN100592385C (en) Method and system for performing speech recognition on multi-language name
Kadyan et al. Refinement of HMM model parameters for punjabi automatic speech recognition (PASR) system
CN114580382A (en) Text error correction method and device
CN102369567B (en) Adaptation for statistical language model
CN102246169A (en) Assigning an indexing weight to a search term
CN110808032A (en) Voice recognition method and device, computer equipment and storage medium
CN1901041B (en) Voice dictionary forming method and voice identifying system and its method
JPWO2012165529A1 (en) Language model construction support apparatus, method and program
CN111435592A (en) Voice recognition method and device and terminal equipment
CN111354343A (en) Voice wake-up model generation method and device and electronic equipment
CN106713111A (en) Processing method for adding friends, terminal and server
CN111428487A (en) Model training method, lyric generation method, device, electronic equipment and medium

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20160525