CN106782560B - Method and device for determining target recognition text - Google Patents
Method and device for determining target recognition text Download PDFInfo
- Publication number
- CN106782560B CN106782560B CN201710127503.9A CN201710127503A CN106782560B CN 106782560 B CN106782560 B CN 106782560B CN 201710127503 A CN201710127503 A CN 201710127503A CN 106782560 B CN106782560 B CN 106782560B
- Authority
- CN
- China
- Prior art keywords
- text
- recognition
- determined
- texts
- candidate
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 38
- 239000013598 vector Substances 0.000 claims description 36
- 238000004364 calculation method Methods 0.000 claims description 3
- 238000012549 training Methods 0.000 description 8
- 230000006870 function Effects 0.000 description 7
- 230000008569 process Effects 0.000 description 5
- 230000003796 beauty Effects 0.000 description 4
- 238000010586 diagram Methods 0.000 description 4
- 238000007781 pre-processing Methods 0.000 description 4
- 238000004458 analytical method Methods 0.000 description 3
- 230000009286 beneficial effect Effects 0.000 description 3
- 230000008878 coupling Effects 0.000 description 3
- 238000010168 coupling process Methods 0.000 description 3
- 238000005859 coupling reaction Methods 0.000 description 3
- 238000005516 engineering process Methods 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 235000021152 breakfast Nutrition 0.000 description 2
- 238000004891 communication Methods 0.000 description 2
- 235000013399 edible fruits Nutrition 0.000 description 2
- 235000013305 food Nutrition 0.000 description 2
- 230000036541 health Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 238000012905 input function Methods 0.000 description 1
- 239000008267 milk Substances 0.000 description 1
- 210000004080 milk Anatomy 0.000 description 1
- 235000013336 milk Nutrition 0.000 description 1
- 239000000203 mixture Substances 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 238000012216 screening Methods 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/30—Information retrieval; Database structures therefor; File system structures therefor of unstructured textual data
- G06F16/35—Clustering; Classification
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F40/00—Handling natural language data
- G06F40/20—Natural language analysis
- G06F40/279—Recognition of textual entities
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Computational Linguistics (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Artificial Intelligence (AREA)
- General Health & Medical Sciences (AREA)
- Databases & Information Systems (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Machine Translation (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The application provides a method and a device for determining a target recognition text, wherein the method comprises the following steps: determining a determined recognition text and a to-be-determined recognition text in at least two candidate recognition texts corresponding to the voice data to be recognized, wherein the determined recognition text is the same part in the at least two candidate recognition texts, and the to-be-determined recognition text is the different part in the at least two candidate recognition texts; calculating the similarity between the texts at the corresponding positions of the recognition text to be determined and the target comparison text, wherein the target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in a preset text library, and the target comparison text comprises the recognition text; configuring candidate recognition texts consisting of the texts to be determined and recognized and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts; the target recognition texts are further screened from the candidate recognition texts, and the accuracy of the target recognition texts is improved.
Description
Technical Field
The present application relates to speech recognition technologies, and in particular, to a method and an apparatus for determining a target recognition text.
Background
With the development of voice control technology, more and more intelligent devices have a voice recognition function, for example, a smart television, a smart refrigerator, a smart air conditioner and the like having a voice control function, and a smart phone and a smart computer having a voice input function.
The current voice recognition mainly comprises the processes of voice preprocessing, acoustic model decoding, pronunciation dictionary analysis, language model decoding and the like, wherein the voice preprocessing is to simply process a received voice signal to obtain a voice feature file and the like; the input of the acoustic model decoding is a feature file of voice, and a phoneme file with the highest probability is obtained through the acoustic model decoding; and then, converting the phoneme information into possible character combinations by inquiring the pronunciation dictionary, and acquiring character combination information with high probability from the character combinations as candidate recognition results through context associated information of the language model. Because the corpus sources in the language model are wide, the accuracy of the recognition result cannot be guaranteed by the candidate recognition result, and therefore, the accurate recognition result needs to be screened by some methods.
However, no suitable selection method exists in the prior art.
Content of application
The application provides a method and a device for determining a target recognition text, which are used for screening out an accurate recognition result from candidate recognition results of voice data to be recognized.
The first aspect of the present application provides a method for determining a target recognition text from at least two candidate recognition texts, comprising:
determining a determined recognition text and a to-be-determined recognition text in at least two candidate recognition texts corresponding to voice data to be recognized, wherein the determined recognition text is the same part in the at least two candidate recognition texts, and the to-be-determined recognition text is the different part in the at least two candidate recognition texts;
calculating the similarity between the text at the corresponding position of the recognition text to be determined and a target comparison text, wherein the target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in a preset text library, and the target comparison text comprises the recognition text to be determined;
and configuring the candidate recognition texts consisting of the texts to be determined and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts.
A second aspect of the present application provides an apparatus for determining a target recognition text from candidate recognition texts, including:
the device comprises a first determining module, a second determining module and a judging module, wherein the first determining module is used for determining a determined recognition text and a to-be-determined recognition text in at least two candidate recognition texts corresponding to voice data to be recognized, the determined recognition text is the same part in the at least two candidate recognition texts, and the to-be-determined recognition text is the different part in the at least two candidate recognition texts;
the calculation module is used for calculating the similarity between the text at the corresponding position of the recognition text to be determined and a target comparison text, wherein the target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in a preset text library, and the target comparison text comprises the determination recognition text;
and the second determining module is used for configuring the candidate recognition texts consisting of the texts to be determined and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts.
The beneficial effect of this application is as follows:
the method for determining the target recognition text comprises the steps of firstly determining a determined recognition text and a text to be determined in at least two candidate recognition texts corresponding to voice data to be recognized, then calculating the similarity between the text to be determined and the text at the corresponding position of a target comparison text aiming at the text to be determined, determining the text to be determined corresponding to the maximum value in the similarity as a correct result corresponding to the voice data to be recognized, further configuring the candidate recognition text consisting of the text to be determined and the determined recognition text as the target recognition text, and determining the text to be determined which is closest to the voice data input by a user according to the target comparison text with the same sentence pattern structure and further according to the similarity between the text to be determined and the text at the corresponding position of the target comparison text when a plurality of candidate recognition texts with close probabilities are obtained, and then the text to be identified and the text to be identified are combined into a target identification text and fed back to the user, namely, different parts in the candidate identification texts with close probabilities are further selected by referring to the target comparison text, so that the accuracy of identifying the voice data to be identified is improved, and the user experience of voice identification is improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the description of the embodiments will be briefly introduced below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art to obtain other drawings based on these drawings without creative efforts.
Fig. 1 is a schematic flowchart of a method for determining a target recognition text from at least two candidate recognition texts according to an embodiment of the present application;
fig. 2 is a schematic flowchart of a method for determining a target recognition text from at least two candidate recognition texts according to another embodiment of the present application;
fig. 3 is a schematic structural diagram of an apparatus for determining a target recognition text from at least two candidate recognition texts according to an embodiment of the present application;
fig. 4 is a schematic structural diagram of an apparatus for determining a target recognition text from at least two candidate recognition texts according to another embodiment of the present application.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention clearer, the present invention will be described in further detail with reference to the accompanying drawings, and it is apparent that the described embodiments are only a part of the embodiments of the present invention, not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
Before explaining the embodiments of the present invention in detail, an application environment of the embodiments of the present invention will be described. The display method for displaying the voice input control instruction provided in the embodiment of the present invention is applied to a terminal, for example, the terminal may be a smart television, a smart phone, a tablet computer, or the like having an Android operating system or an IOS operating system, and the terminal may also be a computer, a PDA (Personal digital assistant), or the like having a Window operating system or an IOS operating system, which is not specifically limited in the embodiment of the present invention.
The application provides a method for determining a target recognition text from at least two candidate recognition texts, and on the basis that a plurality of recognition results are obtained by voice recognition, a final voice recognition text is further analyzed and selected from the plurality of recognition results, so that the accuracy of the voice recognition is improved.
Fig. 1 is a flowchart illustrating a method for determining a target recognition text from at least two candidate recognition texts according to an embodiment of the present application, as shown in fig. 1, the method includes:
s101, determining a determined recognition text and a text to be determined in at least two candidate recognition texts corresponding to the voice data to be recognized.
In a specific implementation process, after a user inputs voice data to be recognized, a plurality of voice recognition texts may be recognized due to reasons such as closeness of pronunciation or recognition accuracy.
For example, a user speaks a sentence "i want to listen to a highly attractive song", and may get a plurality of voice recognition texts "i want to listen to a highly attractive song", and the like.
Candidate recognition texts are determined from the plurality of voice recognition texts, and accurate recognition results are further selected.
The candidate recognition texts are composed of the determined recognition texts and the recognition texts to be determined. The identification text to be determined is the same part in the at least two candidate identification texts. For example, "song" that "i want to listen to" and "song that" i want to listen to "are the identification texts to be determined, and" high win "are the identification texts to be determined.
That is, the same part in the multiple candidate recognition texts can be considered as an accurate result, while the different part is the recognition text to be determined, which needs to be further determined, that is, the recognition text to be determined needs to be further recognized, so as to obtain a more accurate result.
S102, calculating the similarity between the texts at the corresponding positions of the recognition text to be determined and the target contrast text.
The target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in the preset text library, and the target comparison text comprises the determined recognition text.
The preset text library may include a plurality of pre-stored sentences, vocabulary combinations, etc., and the target comparison texts consistent with the candidate recognition text sentence patterns may be matched in the preset text library through word senses, parts of speech (nouns, verbs), etc. For example, "i want to listen to happy songs" may be matched to the target contrast text "i want to listen to zhou jenlen songs" etc. For another example, "please give me a cup of coffee" may match the target contrast text "please give me a cup of milk".
For example, the target contrast text includes the above-mentioned certain identification text, i.e., "i want to listen to a song of zhou jeren" includes the certain identification text "i want to listen to a song".
S103, configuring candidate recognition texts consisting of the texts to be determined and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts.
Optionally, the similarity between the texts at the corresponding positions of the recognition text to be determined and the target contrast text is respectively calculated and determined. For example, the similarity between "grand beauty" and "Zhou Jien" and the similarity between "grand beauty" and "Zhou Jien" are determined, respectively.
If the similarity between the high victory beauty and the Zhoujilun is the largest, the target recognition text is configured as the song that I want to listen to the high victory beauty.
The similarity may refer to semantic similarity, and may also be similarity of belonged types, similarity of parts of speech, and the like, which is not limited herein.
In the embodiment, firstly, a determined recognition text and a text to be determined in at least two candidate recognition texts corresponding to the voice data to be recognized are determined, then, for the text to be determined, the similarity between the text to be determined and the text at the corresponding position of the target comparison text is calculated, the text to be determined and recognized corresponding to the maximum value in the similarity is determined as the correct result corresponding to the voice data to be recognized, and further, the candidate recognition text composed of the text to be determined and the determined recognition text is configured as the target recognition text, so that when a plurality of candidate recognition texts with close probabilities are obtained, the text to be determined and the most similar to the voice data input by the user are determined according to the target comparison text with the same sentence pattern structure as the candidate recognition texts with the close probabilities and further according to the similarity between the text to be determined and the text at the corresponding position in the target comparison text, and then the text to be identified and the text to be identified are combined into a target identification text and fed back to the user, namely, different parts in the candidate identification texts with close probabilities are further selected by referring to the target comparison text, so that the accuracy of identifying the voice data to be identified is improved, and the user experience of voice identification is improved.
Fig. 2 is a flowchart illustrating a method for determining a target recognition text from candidate recognition texts according to another embodiment of the present application. As shown in fig. 2, on the basis of fig. 1, before S101, the method further includes:
s201, acquiring a plurality of voice recognition texts corresponding to the voice data to be recognized.
When a user inputs a segment of speech, the terminal may obtain a plurality of results according to a preset speech recognition decoder, and generally, the preset speech recognition decoder may include one or more models for speech recognition to recognize speech data to be recognized. Because some pronunciation in the voice information is fuzzy or the same pronunciation and similar vocabularies are more, a plurality of voice recognition texts can be recognized.
Specifically, the method comprises the following steps: after the voice data to be recognized is acquired, the voice data to be recognized can be subjected to some preprocessing such as front-end signal processing and endpoint detection processing, voice features are extracted frame by frame, the extracted features are sent to a preset voice recognition decoder, and the preset voice recognition decoder can include: the decoder combines the acoustic model, the language model, and the pronunciation dictionary to obtain a plurality of speech recognition texts.
The acoustic model mainly describes likelihood probability of features under the pronunciation model, and the acoustic model can adopt a Hidden Markov Model (HMM). The Language Model mainly describes continuous occurrence probability among words, the Language Model adopts an n-gram Model, for Chinese, the Model is called a Chinese Language Model (CLM), the Language Model can contain a large amount of linguistic data, the linguistic data can be a large amount of sentences, vocabularies and the like, and the word search result can be constrained according to the statistical probability of co-occurrence between front and back words. The pronunciation dictionary is mainly used for converting words and chords. In specific conversion, the acoustic model decoding is to search the feature file of the sound signal in the acoustic model to generate the optimal phoneme recognition result, wherein phonemes can identify letters. And converting the phoneme recognition result into characters by inquiring the pronunciation dictionary. Finally, the goal of language model decoding is to select the most likely word combination result from the word combinations obtained by querying the pronunciation dictionary as the speech recognition text.
It should be noted that, reference may be made to related technologies for the operation of recognizing the voice data to be recognized to obtain the corresponding voice recognition text, which is not described in detail in the embodiment of the present invention.
For example, the operation of recognizing the speech data to be recognized to obtain the corresponding speech recognition text can be sequentially realized through the following formula.
W1=argmaxP(W|X) (1)
In the formula (1), W represents any character sequence stored in a database, where the character sequence includes words or characters, and the database may be a corpus used for speech recognition; x represents voice data input by a user, W1Represents a character sequence that can be matched with the speech data to be recognized obtained from the stored character sequence, and P (W | X) represents the probability that the speech data to be recognized can become a character. In the above formula (2), W2The matching degree between the voice data to be recognized and the character sequence is shown, P (X | W) shows the probability that the character sequence can pronounce, P (W) shows the probability that the character sequence is a word or a character, and P (X) shows the probability that the voice data to be recognized is audio information.
In the above recognition process, P (W) may be determined by a language model, and P (X | W) may be determined by an acoustic model, so as to complete speech recognition on the speech data to be recognized, and obtain a speech recognition text corresponding to the speech data to be recognized. The language model and the acoustic model will be briefly described below, respectively.
Language model
The language model usually uses the chain rule to break down the probability of a word or character sequence into the product of the probabilities of each word or character, i.e., breaking down W into W1、w2、w3、....wn-1、wnAnd p (w) is determined by the following formula (3).
P(W)=P(w1)P(w2|w1)P(w3|w1,w2)...P(wn|w1,w2,...,wn-1) (3)
In the above formula (3), each term in p (w) is a probability that the current character sequence is a word or a character under the condition that all the character sequences before the character sequence is known are words or characters.
Since when determining p (w) by the above formula (3), if the condition is too long, it is determined that p (w) will be inefficient, thereby affecting subsequent speech recognition. Therefore, to improve the efficiency of determining P (W), P (W) is typically determined by an n-gram language model in the language model. When determining p (w) by the n-gram language model, the probability of the nth word depends only on the (n-1) th word located in front of the word, and p (w) can be determined by the following formula (4).
P(W)=P(w1)P(w2|w1)P(w3|w2)...P(wn|wn-1) (4)
Acoustic model
Since the pronunciation of each word needs to be determined when determining each word, the pronunciation of each word needs to be determined through a dictionary. Where the dictionary is a model juxtaposed to the acoustic model and the language module, and the dictionary can convert a single word into a phoneme string. The acoustic model may determine which sounds should be sequentially pronounced by the words in the user-input speech data through a dictionary, and find the demarcation point of each phoneme through a dynamic rule algorithm such as a Viterbi (Viterbi) algorithm, thereby determining the start-stop time of each phoneme, and thus determining the degree of matching of the user-input speech data with the phoneme string, that is, determining P (X | W).
In general, the distribution of feature vectors of each phoneme can be estimated by a classifier such as a Gaussian mixture model, and in the speech recognition stage, the feature vector x of each frame in the speech data input by the user is determinedtFrom the corresponding phoneme siResulting probability P (x)t|si) The probabilities for each frame are multiplied to obtain P (X | W).
Wherein, the classifier can be obtained by training in advance, and the specific operation is as follows: a large number of feature vectors and phonemes corresponding to each feature vector are extracted from training data through a frequency cepstrum Coefficient (MFCC), and thus a classifier from features to phonemes is trained.
It should be noted that, in practical applications, not only the above-mentioned manner for determining P (X | W) but also other manners, such as directly giving P(s) through a neural network, may be includedi|xt) Can be converted into P (x) by Bayesian formulat|si) And then multiplied to obtain P (X | W), which is only for illustration and does not represent that the embodiment of the present invention is limited thereto.
S202, determining a maximum probability value and a second approximate probability value in probability values corresponding to a plurality of voice recognition texts.
The recognition probability of each speech recognition text can be calculated by adopting a preset algorithm according to the character combination of each speech recognition text.
Alternatively, formulas may be employedCalculating probability value P of each voice recognition textrecWhereinIs the decoding rate of the acoustic model and,is the decoding rate of the pronunciation dictionary,is the decoding rate of the language model.A feature file representing the speech data to be recognized,in order to identify the combination of words,is a sequence of phonemes.
Therefore, the character combination, the phoneme sequence and the feature file of the voice data to be recognized of each voice recognition text are substituted to obtain the corresponding feature file of each voice recognition text And then obtaining the probability value corresponding to each voice recognition text.
Assuming that there are N speech recognition texts in total, the probability value of each speech recognition text is denoted as PnWherein N is 1, 2, … …, N. The maximum probability value P can be further selectedmaxAnd a second rough probability value P2max。
S203, determining whether the difference value between the maximum probability value and the second approximate probability value is larger than a preset probability threshold value.
Further, a difference value between the maximum probability value and the second maximum probability value may be obtained, and if the difference value is greater than or equal to a preset probability threshold value, it indicates that the accuracy of the speech recognition text corresponding to the maximum probability value is higher, and the speech recognition text corresponding to the maximum probability value may be directly determined as the target recognition text.
In specific implementation, the maximum probability value P can be calculated in sequencemaxWith other probability values PnOptionally, using a common equationFormula (II)Calculating the mean absolute value as the acoustic probability value difference EP,EPThe distribution condition of the voice recognition texts is reflected, and the direct difference between the optimal voice recognition text and the rest voice recognition texts is measured. EPWhen the probability value is larger than the preset threshold value, the maximum probability value P can be directly obtainedmaxThe corresponding speech recognition text is determined as the target recognition text without further semantic analysis.
Further, when the difference value between the maximum probability value and the second approximate probability value is smaller than a preset probability threshold, at least two candidate recognized texts from the plurality of voice recognized texts are determined.
Alternatively, the determining of the at least two candidate recognized texts from the plurality of speech recognized texts may be: the method comprises the steps of obtaining a first voice recognition text of which the difference value between the probability value and the maximum probability value is smaller than a preset probability threshold value in a plurality of voice recognition texts, and determining the first voice recognition text and the voice recognition text corresponding to the maximum probability value as at least two candidate recognition texts.
And comparing the maximum probability value with other probability values, and when the difference value is smaller than a preset probability threshold value, taking the voice recognition text corresponding to the compared probability value as a candidate recognition text. If the difference value is larger than or equal to the preset probability threshold value, the probability that the voice recognition text corresponding to the compared probability value becomes the target recognition text is very low, and further analysis is not performed.
Alternatively, the probability values of the plurality of speech recognition texts may be ranked, and a preset number of speech recognition texts with the highest probability values may be selected as candidate recognition texts. Or from top to bottom, selecting candidate recognition texts in sequence according to the difference value of the probability values of two adjacent voice recognition texts, for example, if the difference value of the maximum probability value and the second high probability value is greater than a preset threshold value, directly using the voice recognition text with the highest probability value as a target recognition text, and not continuing to compare; otherwise, the voice recognition text with the highest probability value and the voice recognition text with the second highest probability value are used as candidate recognition texts, the difference value between the second highest probability value and the next probability value is determined in sequence, the candidate recognition texts are determined, and the analogy is repeated until a certain difference value is larger than a preset threshold value, and the comparison is not carried out. Of course, the methods are not limited to these methods, and the candidate recognition texts may be determined flexibly according to the needs, or may be obtained by using a formula or an algorithm.
If only one candidate speech recognition text is determined, this candidate speech recognition text can be directly configured as the target speech recognition text. If there are a plurality of candidate speech recognition texts, a result that best meets the actual situation is further determined as the target speech recognition text.
Optionally, calculating the similarity between the text at the corresponding position of the recognition text to be determined and the target contrast text may include: and determining semantic similarity between the text at the corresponding position of the recognition text to be determined and the target comparison text by adopting a preset word vector model.
The preset word vector model is used for identifying semantic similarity among words through word vector distances.
The preset word vector model can be obtained through word vector training, specifically, the text content can be converted into a real number vector with limited low dimensionality, and the dimensionality is common in 50-dimension and 100-dimension. The distance of the vector can be measured by the most traditional euclidean distance, and can also be measured by the cosine included angle, which is not limited herein. The distance of the vector reflects the distance of the word semantics, i.e. the semantic similarity between words can be expressed by the distance of the vector. The method comprises the following steps that word vector training can be carried out by adopting training tools of word vectors, firstly, training linguistic data capable of comprehensively covering basic words in Chinese are obtained, and corresponding preprocessing is carried out; and then, a training tool of word vectors is called to perform training, and a vector representation form is generated, for example, each word in the corpus has a corresponding 50-dimensional vector representation, which is not limited herein. The larger the vector distance is, the longer the semantic distance between words is, whereas the shorter the semantic distance is.
Specifically, the texts at the corresponding positions of the candidate recognition texts and the target comparison text to be determined appear in the same sentence pattern, and the positions are the same, so that the probability of the same kind of things is very high, and then the similarity is further determined according to the word vector distance.
Table 1 is taken as an example to illustrate:
TABLE 1
Therefore, the word vector distance between the "high-success" and the "zhongjilun" is the closest, then the "song that i want to listen to the high-success" is configured as the target recognition text, and the target recognition text is output and displayed to the user, if the target recognition text is voice information of a control instruction class, related instructions can be executed according to the target recognition text, and details are not repeated herein.
Optionally, the preset word vector model is adopted to determine semantic similarity between the text at the corresponding position of the recognition text to be determined and the target comparison text, and the semantic similarity may be: and when the text to be identified comprises at least two words, respectively determining the semantic similarity between each word in the text to be identified and the word at the corresponding position in the target contrast text by adopting a preset word vector model.
That is, the words at different positions are respectively compared, for example, semantic similarity between the text at the corresponding position of the recognition text "breakfast eating fruit is beneficial to health" and the target comparison text "dinner eating coarse food is beneficial to health" is compared, and semantic similarity between "breakfast" and "dinner" and semantic similarity between "coarse food" and "fruit" can be respectively determined.
Fig. 3 is a schematic structural diagram of an apparatus for determining a target recognition text from at least two candidate recognition texts according to an embodiment of the present application, as shown in fig. 3, the apparatus includes: a first determination module 301, a calculation module 302, and a second determination module 303, wherein:
the first determining module 301 is configured to determine a recognition determining text and a recognition to be determined text in at least two candidate recognition texts corresponding to the voice data to be recognized.
The identification text to be determined is the same part in at least two candidate identification texts, and the identification text to be determined is the different part in at least two candidate identification texts.
A calculating module 302, configured to calculate a similarity between the text at the corresponding position of the to-be-determined recognition text and the target comparison text.
The target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in a preset text library, and the target comparison text comprises the determined recognition text.
A second determining module 303, configured to configure the candidate recognition text composed of the to-be-determined recognition text and the determined recognition text corresponding to the maximum value in the similarity as a target recognition text.
In this embodiment, first, the first determining module 301 determines a determined recognition text and a text to be determined in at least two candidate recognition texts corresponding to the voice data to be recognized, then, the calculating module 302 calculates the similarity between the text to be determined and the text at the corresponding position of the target comparison text for the text to be determined, determines the text to be determined corresponding to the maximum value in the similarity as the correct result corresponding to the voice data to be recognized, and then, the second determining module 302 configures the candidate recognition text composed of the text to be determined and the determined recognition text as the target recognition text, so that when a plurality of candidate recognition texts with close probabilities are obtained, according to the target comparison text with the same sentence pattern structure thereof, further according to the similarity between the text to be determined and the text at the corresponding position in the target comparison text, the method comprises the steps of determining the text to be identified which is closest to voice data input by a user, further forming a target identification text by the text to be identified and the text to be identified, and feeding back the target identification text to the user, namely further selecting different parts in a plurality of candidate identification texts with close probabilities by referring to a target comparison text, so that the accuracy of identifying the voice data to be identified is improved, and the user experience of voice identification is improved.
Fig. 4 is a schematic structural diagram of an apparatus for determining a target recognition text from at least two candidate recognition texts according to another embodiment of the present application, as shown in fig. 4, on the basis of fig. 3, the apparatus further includes: a third determination module 401, wherein:
the third determining module 401 is configured to determine a maximum probability value and a second approximate probability value in the plurality of speech recognition texts corresponding to the speech data to be recognized before the first determining module 301 determines the determined recognition text and the recognized text to be determined in the at least two candidate recognition texts corresponding to the speech data to be recognized.
In this embodiment, the first determining module 301 determines at least two candidate recognized texts from the plurality of speech recognized texts when a difference between the maximum probability value and the second approximate probability value is smaller than a preset probability threshold.
Optionally, the first determining module 301 is specifically configured to acquire a first speech recognition text in the plurality of speech recognition texts, where a difference between the probability value and the maximum probability value is smaller than a preset probability threshold; and determining the first voice recognition text and the voice recognition text corresponding to the maximum probability value as the at least two candidate recognition texts.
Further, the calculating module 302 is specifically configured to determine semantic similarity between the text to be determined and identified and the text at the corresponding position in the target comparison text by using a preset word vector model. The preset word vector model is used for identifying semantic similarity among words through word vector distance.
Optionally, the calculating module 302 is specifically configured to, when the to-be-determined recognition text includes at least two vocabularies, respectively determine semantic similarity between each vocabulary in the to-be-determined recognition text and a vocabulary in a corresponding position in the target comparison text by using the preset word vector model.
It should be noted that: in the apparatus for determining a target recognition text provided in the above embodiment, when determining a target recognition text from at least two candidate recognition texts, only the division of the above function modules is used as an example, in practical applications, the function distribution may be completed by different function modules according to needs, that is, the internal structure of the apparatus is divided into different function modules, so as to complete all or part of the functions described above. In addition, the apparatus for determining a target recognition text and the method for determining a target recognition text provided by the above embodiments belong to the same concept, and specific implementation processes thereof are detailed in the method embodiments and are not described herein again.
In the several embodiments provided in the present application, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described apparatus embodiments are merely illustrative, and for example, the division of the units is only one logical division, and other divisions may be realized in practice, for example, a plurality of units or components may be combined or integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical, mechanical or other form.
In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, or in a form of hardware plus a software functional unit.
The integrated unit implemented in the form of a software functional unit may be stored in a computer readable storage medium. The software functional unit is stored in a storage medium and includes several instructions for enabling a computer device (which may be a personal computer, a server, or a network device) or a processor (processor) to perform some steps of the methods according to the embodiments of the present application. And the aforementioned storage medium includes: a U disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
Finally, it should be noted that: the above embodiments are only used for illustrating the technical solutions of the present application, and not for limiting the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present application.
Claims (10)
1. A method for determining a target recognition text from at least two candidate recognition texts, comprising:
determining a determined recognition text and a to-be-determined recognition text in at least two candidate recognition texts corresponding to voice data to be recognized, wherein the determined recognition text is the same part in the at least two candidate recognition texts, and the to-be-determined recognition text is the different part in the at least two candidate recognition texts;
calculating the similarity between the text at the corresponding position of the recognition text to be determined and a target comparison text, wherein the target comparison text is a text in a preset text library consistent with the sentence pattern structure of the candidate recognition text, the target comparison text comprises the recognition text to be determined, and the positions of the text at the corresponding position of the recognition text to be determined and the text at the corresponding position of the target comparison text in the same sentence pattern structure are the same;
and configuring the candidate recognition texts consisting of the texts to be determined and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts.
2. The method according to claim 1, wherein before determining the determined recognized text and the text to be determined in the at least two candidate recognized texts corresponding to the speech data to be recognized, the method further comprises:
determining a maximum probability value and a second approximate probability value in a plurality of voice recognition texts corresponding to the voice data to be recognized;
determining at least two candidate recognized texts from the plurality of speech recognized texts when a difference between the maximum probability value and the second approximate probability value is smaller than a preset probability threshold.
3. The method of claim 2, wherein determining at least two candidate recognized texts from the plurality of speech recognized texts comprises:
acquiring a first voice recognition text of which the difference value between the probability value and the maximum probability value in the plurality of voice recognition texts is smaller than a preset probability threshold;
and determining the first voice recognition text and the voice recognition text corresponding to the maximum probability value as the at least two candidate recognition texts.
4. The method according to claim 1, wherein the calculating of the similarity between the text at the corresponding position of the recognition text to be determined and the target comparison text specifically comprises:
and determining semantic similarity between the text to be determined and recognized and the text at the corresponding position of the target comparison text by adopting a preset word vector model, wherein the preset word vector model is used for identifying the semantic similarity between words through word vector distance.
5. The method according to claim 4, wherein the determining the semantic similarity between the recognition text to be determined and the text at the corresponding position in the target comparison text by using the preset word vector model specifically comprises:
and when the text to be identified comprises at least two words, respectively determining the semantic similarity between each word in the text to be identified and the word at the corresponding position in the target contrast text by adopting the preset word vector model.
6. An apparatus for determining a target recognized text from at least two candidate recognized texts, comprising:
the device comprises a first determining module, a second determining module and a judging module, wherein the first determining module is used for determining a determined recognition text and a to-be-determined recognition text in at least two candidate recognition texts corresponding to voice data to be recognized, the determined recognition text is the same part in the at least two candidate recognition texts, and the to-be-determined recognition text is the different part in the at least two candidate recognition texts;
the calculation module is used for calculating the similarity between the text at the corresponding position of the recognition text to be determined and the target comparison text, wherein the target comparison text is a text which is consistent with the sentence pattern structure of the candidate recognition text in a preset text library, the target comparison text comprises the recognition text to be determined, and the positions of the text at the corresponding position of the recognition text to be determined and the text at the corresponding position of the target comparison text in the same sentence pattern structure are the same;
and the second determining module is used for configuring the candidate recognition texts consisting of the texts to be determined and the determined recognition texts corresponding to the maximum value in the similarity as target recognition texts.
7. The apparatus of claim 6, further comprising: a third determination module;
the third determining module is configured to determine a maximum probability value and a second approximate probability value in a plurality of voice recognition texts corresponding to the voice data to be recognized before the first determining module determines the determined recognition text and the recognized text to be determined in the at least two candidate recognition texts corresponding to the voice data to be recognized;
the first determining module is specifically configured to determine at least two candidate recognized texts from the plurality of speech recognized texts when a difference between the maximum probability value and the second maximum probability value is smaller than a preset probability threshold.
8. The apparatus according to claim 7, wherein the first determining module is specifically configured to obtain a first speech recognition text in the plurality of speech recognition texts, where a difference between the probability value and the maximum probability value is smaller than a preset probability threshold; and determining the first voice recognition text and the voice recognition text corresponding to the maximum probability value as the at least two candidate recognition texts.
9. The apparatus according to claim 6, wherein the computing module is specifically configured to determine the semantic similarity between the recognized text to be determined and the text at the corresponding position in the target comparison text by using a preset word vector model, where the preset word vector model is used to identify the semantic similarity between words by word vector distance.
10. The apparatus according to claim 9, wherein the computing module is specifically configured to, when the recognition text to be determined includes at least two words, respectively determine semantic similarities between each word in the recognition text to be determined and a word at a corresponding position in the target comparison text by using the preset word vector model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710127503.9A CN106782560B (en) | 2017-03-06 | 2017-03-06 | Method and device for determining target recognition text |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710127503.9A CN106782560B (en) | 2017-03-06 | 2017-03-06 | Method and device for determining target recognition text |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106782560A CN106782560A (en) | 2017-05-31 |
CN106782560B true CN106782560B (en) | 2020-06-16 |
Family
ID=58962349
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710127503.9A Active CN106782560B (en) | 2017-03-06 | 2017-03-06 | Method and device for determining target recognition text |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106782560B (en) |
Families Citing this family (18)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107329843B (en) * | 2017-06-30 | 2021-06-01 | 百度在线网络技术(北京)有限公司 | Application program voice control method, device, equipment and storage medium |
CN107277645A (en) * | 2017-07-27 | 2017-10-20 | 广东小天才科技有限公司 | Error correction method and device for subtitle content |
CN107680585B (en) * | 2017-08-23 | 2020-10-02 | 海信集团有限公司 | Chinese word segmentation method, Chinese word segmentation device and terminal |
CN108197102A (en) | 2017-12-26 | 2018-06-22 | 百度在线网络技术(北京)有限公司 | A kind of text data statistical method, device and server |
CN108417210B (en) * | 2018-01-10 | 2020-06-26 | 苏州思必驰信息科技有限公司 | Word embedding language model training method, word recognition method and system |
CN108364655B (en) * | 2018-01-31 | 2021-03-09 | 网易乐得科技有限公司 | Voice processing method, medium, device and computing equipment |
CN110188338B (en) * | 2018-02-23 | 2023-02-21 | 富士通株式会社 | Text-dependent speaker verification method and apparatus |
CN109829704A (en) * | 2018-12-07 | 2019-05-31 | 创发科技有限责任公司 | Payment channel configuration method, device and computer readable storage medium |
CN111681670B (en) * | 2019-02-25 | 2023-05-12 | 北京嘀嘀无限科技发展有限公司 | Information identification method, device, electronic equipment and storage medium |
CN109918680B (en) * | 2019-03-28 | 2023-04-07 | 腾讯科技(上海)有限公司 | Entity identification method and device and computer equipment |
CN110705274B (en) * | 2019-09-06 | 2023-03-24 | 电子科技大学 | Fusion type word meaning embedding method based on real-time learning |
CN110853635B (en) * | 2019-10-14 | 2022-04-01 | 广东美的白色家电技术创新中心有限公司 | Speech recognition method, audio annotation method, computer equipment and storage device |
CN110706707B (en) | 2019-11-13 | 2020-09-18 | 百度在线网络技术(北京)有限公司 | Method, apparatus, device and computer-readable storage medium for voice interaction |
JP7374756B2 (en) * | 2019-12-20 | 2023-11-07 | キヤノン株式会社 | Information processing device, information processing method, and program |
CN111667821A (en) * | 2020-05-27 | 2020-09-15 | 山西东易园智能家居科技有限公司 | Voice recognition system and recognition method |
CN112614263A (en) * | 2020-12-30 | 2021-04-06 | 浙江大华技术股份有限公司 | Method and device for controlling gate, computer equipment and storage medium |
CN113177114B (en) * | 2021-05-28 | 2022-10-21 | 重庆电子工程职业学院 | Natural language semantic understanding method based on deep learning |
CN113539270B (en) * | 2021-07-22 | 2024-04-02 | 阳光保险集团股份有限公司 | Position identification method and device, electronic equipment and storage medium |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101655837A (en) * | 2009-09-08 | 2010-02-24 | 北京邮电大学 | Method for detecting and correcting error on text after voice recognition |
CN102999483A (en) * | 2011-09-16 | 2013-03-27 | 北京百度网讯科技有限公司 | Method and device for correcting text |
CN105374351A (en) * | 2014-08-12 | 2016-03-02 | 霍尼韦尔国际公司 | Methods and apparatus for interpreting received speech data using speech recognition |
CN105513586A (en) * | 2015-12-18 | 2016-04-20 | 百度在线网络技术(北京)有限公司 | Speech recognition result display method and speech recognition result display device |
CN105869642A (en) * | 2016-03-25 | 2016-08-17 | 海信集团有限公司 | Voice text error correction method and device |
Family Cites Families (7)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
JP2003308094A (en) * | 2002-02-12 | 2003-10-31 | Advanced Telecommunication Research Institute International | Method for correcting recognition error place in speech recognition |
US7587308B2 (en) * | 2005-11-21 | 2009-09-08 | Hewlett-Packard Development Company, L.P. | Word recognition using ontologies |
CN103699530A (en) * | 2012-09-27 | 2014-04-02 | 百度在线网络技术(北京)有限公司 | Method and equipment for inputting texts in target application according to voice input information |
CN104021786B (en) * | 2014-05-15 | 2017-05-24 | 北京中科汇联信息技术有限公司 | Speech recognition method and speech recognition device |
KR102380833B1 (en) * | 2014-12-02 | 2022-03-31 | 삼성전자주식회사 | Voice recognizing method and voice recognizing appratus |
CN106326303B (en) * | 2015-06-30 | 2019-09-13 | 芋头科技(杭州)有限公司 | A kind of spoken semantic analysis system and method |
CN106469554B (en) * | 2015-08-21 | 2019-11-15 | 科大讯飞股份有限公司 | A kind of adaptive recognition methods and system |
-
2017
- 2017-03-06 CN CN201710127503.9A patent/CN106782560B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101655837A (en) * | 2009-09-08 | 2010-02-24 | 北京邮电大学 | Method for detecting and correcting error on text after voice recognition |
CN102999483A (en) * | 2011-09-16 | 2013-03-27 | 北京百度网讯科技有限公司 | Method and device for correcting text |
CN105374351A (en) * | 2014-08-12 | 2016-03-02 | 霍尼韦尔国际公司 | Methods and apparatus for interpreting received speech data using speech recognition |
CN105513586A (en) * | 2015-12-18 | 2016-04-20 | 百度在线网络技术(北京)有限公司 | Speech recognition result display method and speech recognition result display device |
CN105869642A (en) * | 2016-03-25 | 2016-08-17 | 海信集团有限公司 | Voice text error correction method and device |
Also Published As
Publication number | Publication date |
---|---|
CN106782560A (en) | 2017-05-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106782560B (en) | Method and device for determining target recognition text | |
US10283111B1 (en) | Disambiguation in speech recognition | |
US11496582B2 (en) | Generation of automated message responses | |
US11380330B2 (en) | Conversational recovery for voice user interface | |
US10489393B1 (en) | Quasi-semantic question answering | |
US10134388B1 (en) | Word generation for speech recognition | |
US9484021B1 (en) | Disambiguation in speech recognition | |
US20220189458A1 (en) | Speech based user recognition | |
US10210862B1 (en) | Lattice decoding and result confirmation using recurrent neural networks | |
US10332508B1 (en) | Confidence checking for speech processing and query answering | |
US10388274B1 (en) | Confidence checking for speech processing and query answering | |
US9934785B1 (en) | Identification of taste attributes from an audio signal | |
US10121467B1 (en) | Automatic speech recognition incorporating word usage information | |
US11823678B2 (en) | Proactive command framework | |
US10056078B1 (en) | Output of content based on speech-based searching and browsing requests | |
CN106463113B (en) | Predicting pronunciation in speech recognition | |
US7421387B2 (en) | Dynamic N-best algorithm to reduce recognition errors | |
US10963497B1 (en) | Multi-stage query processing | |
US20180137109A1 (en) | Methodology for automatic multilingual speech recognition | |
US9704483B2 (en) | Collaborative language model biasing | |
US11093110B1 (en) | Messaging feedback mechanism | |
US20140195238A1 (en) | Method and apparatus of confidence measure calculation | |
CN109637537B (en) | Method for automatically acquiring annotated data to optimize user-defined awakening model | |
US10366442B1 (en) | Systems and methods to update shopping cart | |
US20050187767A1 (en) | Dynamic N-best algorithm to reduce speech recognition errors |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |