CN109461436B - Method and system for correcting pronunciation errors of voice recognition - Google Patents

Method and system for correcting pronunciation errors of voice recognition Download PDF

Info

Publication number
CN109461436B
CN109461436B CN201811239934.5A CN201811239934A CN109461436B CN 109461436 B CN109461436 B CN 109461436B CN 201811239934 A CN201811239934 A CN 201811239934A CN 109461436 B CN109461436 B CN 109461436B
Authority
CN
China
Prior art keywords
pronunciation
acoustic model
audio
error
prone
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811239934.5A
Other languages
Chinese (zh)
Other versions
CN109461436A (en
Inventor
魏誉荧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Genius Technology Co Ltd
Original Assignee
Guangdong Genius Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Genius Technology Co Ltd filed Critical Guangdong Genius Technology Co Ltd
Priority to CN201811239934.5A priority Critical patent/CN109461436B/en
Publication of CN109461436A publication Critical patent/CN109461436A/en
Application granted granted Critical
Publication of CN109461436B publication Critical patent/CN109461436B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/01Assessment or evaluation of speech recognition systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/10Speech classification or search using distance or distortion measures between unknown speech and reference templates
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
    • G10L25/60Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination for measuring the quality of voice signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • G10L2015/0635Training updating or merging of old and new templates; Mean values; Weighting

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Signal Processing (AREA)
  • Quality & Reliability (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Machine Translation (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

The invention provides a method and a system for correcting pronunciation errors in speech recognition, wherein the method comprises the following steps: establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to pronounce error-prone characters; acquiring user voice information; recognizing the user voice information, and when the pronunciation-prone wrong character is contained in the voice information, extracting an audio segment corresponding to a word containing the pronunciation-prone wrong character in the user voice information; and when the audio frequency segment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss and the pronunciation is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table. The invention prompts and outputs corresponding correct audio when recognizing that the pronunciation of the pronouncing error-prone character of the user is wrong by establishing a mapping table between the standard acoustic model and the error acoustic model.

Description

Method and system for correcting pronunciation errors of voice recognition
Technical Field
The present invention relates to the field of speech recognition technology, and in particular, to a method and system for correcting pronunciation errors in speech recognition.
Background
With the rapid development of the internet, the lives of people become more and more intelligent. Voice interaction is also becoming more popular with users as one of the mainstream communication applications of human-computer interaction in intelligent terminals. The intelligent terminal takes corresponding measures based on the voice input by the user, so the accuracy of the voice input by the user through the terminal seriously influences the feedback made by the intelligent terminal.
The Chinese characters have a large number of polyphonic characters, shape-near characters and the like, and for some users, it is difficult to distinguish rarely used polyphonic characters and shape-near characters, and even more, the common pronunciation of some users of some polyphonic characters and shape-near characters is wrong.
In addition, for pupils, in the learning process, particularly under the condition of small literacy amount, the situation that the words containing polyphone or the characters with similar shapes are wrongly read often occurs, and when the intelligent terminal identifies the pronunciation, the situation can cause the identification error and cannot give out the correct result to be inquired or the corresponding accurate feedback. Therefore, a method and a system for correcting pronunciation errors in speech recognition are needed to solve the above problems.
Disclosure of Invention
The invention aims to provide a method and a system for correcting pronunciation errors of voice recognition, which can prompt and output corresponding correct audio when pronunciation errors of pronunciations of characters easy to be mistaken by a user are recognized by establishing a mapping table between a standard acoustic model and an error acoustic model.
The technical scheme provided by the invention is as follows:
the invention provides a method for correcting pronunciation errors in speech recognition, which is characterized by comprising the following steps:
establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to pronounce error-prone characters;
acquiring user voice information;
recognizing the user voice information, and when the pronunciation-prone wrong character is contained in the voice information, extracting an audio segment corresponding to a word containing the pronunciation-prone wrong character in the user voice information;
and when the audio frequency segment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss and the pronunciation is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
Further, before the establishing of the mapping table between the standard acoustic model and the error acoustic model corresponding to the pronunciation-prone and error-prone word, the method further includes:
acquiring the pronouncing error-prone characters, and generating target words according to the pronouncing error-prone characters;
acquiring the voice audio of the target word, and generating the standard acoustic model according to the voice audio of the target word;
acquiring pronunciation confusion words of the pronunciation easily-wrong words, and replacing the pronunciation easily-wrong words in the target words with the pronunciation confusion words to generate confusion words;
and acquiring the voice audio of the confusing words, and generating the error acoustic model according to the voice audio of the confusing words.
Further, the method also comprises the following steps:
and when the audio frequency segment is matched with the voice audio frequency matching result in the standard acoustic model, prompting the user that the pronunciation of the error-prone character is correct.
Further, the method also comprises the following steps:
when the audio segment does not accord with the voice audio matching result in the acoustic model, converting the audio segment into recognition text, wherein the acoustic model comprises the standard acoustic model and the error acoustic model;
if the target words contain the recognition texts, judging whether the pronunciation of the audio clips is correct, and if so, updating the standard acoustic model according to the audio clips; otherwise, updating the error acoustic model according to the audio fragment, prompting the user that the pronunciation of the pronunciation-prone character is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table;
and if the target word does not contain the identification text, updating the target word according to the identification text, and updating the acoustic model according to the audio fragment.
Further, when the target word does not include the identification text, updating the target word according to the identification text, and updating the acoustic model according to the audio segment specifically includes:
when the target word does not contain the recognition text, updating the target word according to the recognition text;
if the pronunciation of the audio clip is correct, updating the standard acoustic model according to the audio clip, updating the confusion terms according to the updated target terms, and then updating the error acoustic model according to the voice audio of the updated confusion terms;
and if the audio fragment is mispronounced, acquiring correct voice audio of the recognized text, updating the standard acoustic model according to the correct voice audio, updating the confusion terms according to the updated target terms, and updating the wrong acoustic model according to the updated voice audio of the confusion terms.
The present invention also provides a system for correcting pronunciation errors in speech recognition, comprising:
the mapping table establishing module is used for establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronounce error-prone character;
the acquisition module acquires user voice information;
the extraction module is used for identifying the user voice information acquired by the acquisition module, and extracting an audio clip corresponding to a word containing the pronouncing error-prone character in the user voice information when the pronunciation error-prone character is contained in the voice information;
and the processing module prompts a user that the pronunciation is easy to make wrong pronunciation when the audio frequency segment extracted by the extraction module is matched with the voice audio frequency matching result in the wrong acoustic model in the mapping table establishing module, and outputs the corresponding voice audio frequency in the standard acoustic model according to the mapping table established by the mapping table establishing module.
Further, the method also comprises the following steps:
the error-prone character acquisition module is used for acquiring the pronunciation error-prone characters;
the target word generation module is used for generating a target word according to the pronunciation error-prone characters acquired by the error-prone character acquisition module;
the audio acquisition module is used for acquiring the voice audio of the target words generated by the target word generation module;
the acoustic model generating module is used for generating the standard acoustic model according to the voice audio of the target word acquired by the audio acquiring module;
the confusing character acquisition module is used for acquiring the pronunciation confusing characters of the pronunciation confusing characters acquired by the confusing character acquisition module;
a confusion word generation module, which replaces the pronunciation-prone wrong character in the target word generated by the target word generation module with the pronunciation-prone wrong character acquired by the confusion word acquisition module to generate a confusion word;
the audio acquisition module is used for acquiring the voice audio of the confusion word generated by the confusion word generation module;
the acoustic model generating module is used for generating the error acoustic model according to the voice audio of the confusing words acquired by the audio acquiring module.
Further, the method also comprises the following steps:
and the processing module prompts a user that the pronunciation of the error-prone character is correct when the audio frequency segment extracted by the extraction module is matched with the voice audio frequency matching result in the standard acoustic model in the mapping table establishing module.
Further, the method also comprises the following steps:
the processing module is used for converting the audio clip into an identification text when the audio clip extracted by the extraction module is not consistent with a voice audio matching result in an acoustic model in the mapping table establishing module, wherein the acoustic model comprises the standard acoustic model and the error acoustic model;
the control module is used for judging whether the pronunciation of the audio clip is correct or not if the target word contains the recognition text converted by the processing module, and updating the standard acoustic model according to the audio clip if the pronunciation of the audio clip is correct; otherwise, updating the error acoustic model according to the audio fragment, prompting the user that the pronunciation of the pronunciation-prone character is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table;
and if the target word does not contain the identification text converted by the processing module, the control module updates the target word according to the identification text and updates the acoustic model according to the audio segment.
Further, the control module specifically includes:
a target word updating unit, which updates the target word according to the recognition text when the target word does not contain the recognition text converted by the processing module;
the control unit is used for updating the standard acoustic model according to the audio segment if the pronunciation of the audio segment extracted by the extraction module is correct, updating the confusion word according to the target word updated by the target word updating unit, and then updating the wrong acoustic model according to the voice audio of the updated confusion word;
the control unit is used for acquiring correct voice audio of the recognized text if the audio segment extracted by the extraction module is in wrong pronunciation, updating the standard acoustic model according to the correct voice audio, updating the confusion word according to the target word updated by the target word updating unit, and then updating the wrong acoustic model according to the updated voice audio of the confusion word.
The method and the system for correcting the pronunciation error of the voice recognition can bring at least one of the following beneficial effects:
1. in the invention, a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronouncing error-prone character is established so as to be convenient for matching and recognizing the user voice information subsequently and correct the wrong pronouncing.
2. In the invention, when the voice information of the user is matched, only the audio frequency segment corresponding to the word containing the pronouncing easily-wrong character is extracted for matching, so that the difficulty and the workload in the matching process are reduced, the matching speed is accelerated, and the matching accuracy is improved.
Drawings
The above features, technical features, advantages and implementations of a method and system for correcting pronunciation errors in speech recognition will be further described in the following detailed description of preferred embodiments in a clearly understandable manner in conjunction with the accompanying drawings.
FIG. 1 is a flow chart of a first embodiment of a method for correcting pronunciation errors in speech recognition according to the present invention;
FIG. 2 is a flowchart of a second embodiment of a method for correcting pronunciation errors in speech recognition according to the present invention;
FIG. 3 is a flowchart of a third embodiment of a method for correcting pronunciation errors in speech recognition according to the present invention;
FIG. 4 is a schematic structural diagram of a fourth embodiment of a system for correcting pronunciation errors in speech recognition according to the present invention;
FIG. 5 is a schematic structural diagram of a fifth embodiment of a system for correcting pronunciation errors in speech recognition according to the present invention;
fig. 6 is a schematic structural diagram of a sixth embodiment of a system for correcting pronunciation errors in speech recognition according to the present invention.
The reference numbers illustrate:
1000 pronunciation error correction system for speech recognition
1100 mapping table establishing module 1200 obtaining module 1300 extracting module
1400 processing module 1500 error-prone word acquisition module 1600 target word generation module
1700 Audio acquisition Module 1800 Acoustic model Generation Module 1850 confusion word acquisition Module
1900 confusing term generating module
1950 control module 1951 target word update unit 1952 control unit
Detailed Description
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the following description will be made with reference to the accompanying drawings. It is obvious that the drawings in the following description are only some examples of the invention, and that for a person skilled in the art, other drawings and embodiments can be derived from them without inventive effort.
For the sake of simplicity, the drawings only schematically show the parts relevant to the present invention, and they do not represent the actual structure as a product. In addition, in order to make the drawings concise and understandable, components having the same structure or function in some of the drawings are only schematically illustrated or only labeled. In this document, "one" means not only "only one" but also a case of "more than one".
A first embodiment of the present invention, as shown in fig. 1, is a method for correcting pronunciation errors in speech recognition, including:
s100, establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronounceable error-prone character.
Specifically, polyphone characters and Chinese characters with a plurality of similar characters are used as pronouncing error-prone characters, correct pronunciations are used as standard acoustic models, polyphone characters and similar characters are used as error acoustic models, and then mapping tables corresponding to the standard acoustic models and the error acoustic models are established so as to find corresponding correct pronunciations for correcting the error pronunciations.
S200, acquiring user voice information.
Specifically, the user voice information is obtained, which may be input by the user in real time, for example, the user needs to perform voice correction when reading aloud with the intelligent terminal, or a pupil needs to check the learning result of the pupil when learning a new vocabulary through the intelligent terminal. It is also possible to implement recorded audio, for example, to check whether the pronunciation in the audio recorded by the student is accurate. Prompt correction is performed when the system detects a wrong pronunciation, so the user can autonomously select whether the system or a single application starts the function of prompting correction of the wrong pronunciation.
S300, recognizing the user voice information, and when the pronunciation error prone characters are contained in the voice information, extracting audio segments corresponding to the words containing the pronunciation error prone characters in the user voice information.
Specifically, the acquired user voice information is firstly converted into text information, whether the text information contains any one or more pronouncing error-prone characters is identified, if yes, the components of sentences in the text information and the parts of speech of the words are analyzed through a word segmentation technology, the words containing the pronouncing error-prone characters are marked, and finally one or more audio segments corresponding to the marked words are extracted from the user voice information according to the marking result.
S400, when the audio frequency segment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
Specifically, the speech audio of the pronouncing error-prone character corresponding to the standard acoustic model and the wrong acoustic model is found according to the pronouncing error-prone character corresponding to the extracted audio segment, and then the extracted audio segment is matched with the found speech audio one by one.
If the matching result is matched with the voice audio belonging to the wrong acoustic model, the voice audio is indicated as a part of the pronunciation error-prone character in the voice information of the user, and therefore the user is prompted to pronounce the pronunciation error of the pronunciation error-prone character.
Since the same pronunciation-prone character may appear many times in the user voice information, but pronunciation errors do not occur everywhere, when a pronunciation error is prompted, the specific position of the pronunciation-prone character needs to be clearly explained in combination with the context of the pronunciation-prone character in the user voice information.
In addition, the correct voice audio corresponding to the word formed by the pronouncing error-prone character at the position is found through a mapping table between the standard acoustic model and the error acoustic model, and the correct voice audio is output to the user, so that the user can correct pronouncing conveniently.
In this embodiment, after the user voice information is acquired, the pronouncing error-prone character included in the user voice information is firstly recognized, and then an audio segment corresponding to a word including the pronouncing error-prone character in the user voice information is extracted to be matched with a standard acoustic model and a voice audio frequency in an error acoustic model. Only effective audio segments are extracted for matching, so that the voice length required to be matched is reduced, the requirement on the matching capacity of the system is lowered, the matching speed is increased, and the matching accuracy is increased.
A second embodiment of the present invention is an optimized embodiment of the first embodiment, and as shown in fig. 2, the second embodiment includes:
and S010 obtains the pronouncing error-prone character, and generates a target word according to the pronouncing error-prone character.
Specifically, the polyphonic character and the Chinese character with a plurality of shape-similar characters are used as pronunciation-prone characters, a target word is generated according to the pronunciation-prone characters, the target word is a word including pronunciation-prone characters, the pronunciation-prone characters are polyphonic characters or have a plurality of shape-similar characters, and pronunciations of the pronunciation-prone characters in the target word are easy to be confused, such as shen, downloading, pressing, and the like. Or the pronunciation of a part of words does not affect the communication of the daily life of the user, but actually the pronunciation is wrong, such as stubborn, frame and the like, and correction should be given, especially when the elementary school students just start to learn new words, correct pronunciation should be formed from the beginning to avoid later correction.
S020 obtains the voice audio of the target word, and the standard acoustic model is generated according to the voice audio of the target word.
Specifically, the voice audio of each target word is obtained, wherein each pronunciation-prone word may correspond to one or more voice audios, wherein each target word obtains the audio of the standard mandarin pronunciation, and then the audio of the standard mandarin pronunciation is preferentially selected by the system output when the pronunciation of the user needs to be corrected.
However, due to the influence of gender, age, region, dialect accent, cavity tone and the like of the user, the pronunciation of the user is correct but cannot be matched with the pronunciation audio of the standard mandarin, so that besides the audio of the standard mandarin pronunciation, each target word also needs to acquire the audio of the correct pronunciation of people in different ages and regions as much as possible, and the user can independently select whether to prompt and correct the subsequent pronunciation under the condition of correct pronunciation but not standard.
And generating a standard acoustic model according to the voice audio of the target word, classifying the standard acoustic model according to the pronouncing error-prone characters, and classifying again according to the target word under the classification of the pronouncing error-prone characters when a certain pronouncing error-prone character generates more target words or the voice audio corresponding to a certain target word is more.
S030 acquires the pronunciation confusion word of the pronunciation easily-mistaken word, and replaces the pronunciation confusion word in the target word with the pronunciation confusion word to generate the confusion word.
Specifically, the pronunciation confusion word of the pronunciation easy-to-mistake character is obtained, the pronunciation confusion word is the pronunciation easy-to-mistake character of other pronunciations or the shape near character of the pronunciation easy-to-mistake character, the pronunciation easy-to-mistake character in the target word is replaced by the pronunciation confusion word to generate the confusion word, namely, the pronunciation easy-to-mistake character of other pronunciations and the target word are combined or the shape near character of the pronunciation easy-to-mistake character and the target word are combined to generate the confusion word.
The confusing words are words corresponding to voices which may be wrongly read by users when reading the target words, and the confusing words do not conform to the word composition rule in composition, and actually do not have the words, and the confusing words are only used for obtaining matched voice audio subsequently.
S040 obtains the speech audio of the confusing word, and generates the error acoustic model according to the speech audio of the confusing word.
Specifically, the voice audio of each confusing word is obtained, and similarly, due to the influence of the gender, age, region, dialect accent, accent and the like of the user, there may exist a plurality of voice audios corresponding to the same confusing word. And generating an error acoustic model according to the voice audio of the confusing words, classifying the error acoustic model according to the corresponding pronunciation easily-wrong words, and classifying again according to the confusing words under the classification of the pronunciation easily-wrong words when more confusing words correspond to a certain pronunciation easily-wrong word or more voice audio correspond to a certain confusing word.
S100, establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronounceable error-prone character.
S200, acquiring user voice information.
S300, recognizing the user voice information, and when the pronunciation error prone characters are contained in the voice information, extracting audio segments corresponding to the words containing the pronunciation error prone characters in the user voice information.
S400, when the audio frequency segment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
S500, when the audio frequency segment is matched with the voice audio frequency matching result in the standard acoustic model, prompting the user that the pronunciation of the error-prone character is correct.
Specifically, the speech audio of the pronouncing error-prone character corresponding to the standard acoustic model and the wrong acoustic model is found according to the pronouncing error-prone character corresponding to the extracted audio segment, and then the extracted audio segment is matched with the found speech audio one by one.
If the voice audio matched with the matching result is the voice audio belonging to the standard acoustic model, the part of the voice information of the user about the pronounceable wrong character pronounces correctly, but whether the voice audio matched with the matching result is the standard mandarin pronunciation audio needs to be further judged, if the pronunciation of the user is correct and standard, the user is prompted to pronounce the wrong pronounce character correctly but not standard, and the user is asked whether the standard mandarin pronunciation audio needs to be output.
In the embodiment, a standard acoustic model of correct pronunciation of the pronunciation error-prone character and a corresponding incorrect acoustic model of incorrect pronunciation of a combination of the polyphone character and the similar character aiming at the confusable pronunciation of the pronunciation error-prone character are respectively generated, and a mapping table of a corresponding relation between the two acoustic models is established for subsequently identifying the voice information of the user and then carrying out corresponding matching, so that the place of the pronunciation error in the voice information of the user can be accurately identified and corrected.
A third embodiment of the present invention is a preferable embodiment of the first embodiment, and as shown in fig. 3, the third embodiment includes:
s100, establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronounceable error-prone character.
S200, acquiring user voice information.
S300, recognizing the user voice information, and when the pronunciation error prone characters are contained in the voice information, extracting audio segments corresponding to the words containing the pronunciation error prone characters in the user voice information.
S400, when the audio frequency segment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
S600, when the audio segment is not matched with the voice audio matching result in the acoustic model, the audio segment is converted into a recognition text, and the acoustic model comprises the standard acoustic model and the error acoustic model.
Specifically, the acoustic models, namely the speech audio of the pronouncing error-prone characters corresponding to the standard acoustic model and the error acoustic model, are found according to the pronouncing error-prone characters corresponding to the extracted audio segments, and then the extracted audio segments are matched with the found speech audio one by one.
If the matching results of the audio segment and the voice audio in the acoustic model are not consistent, it indicates that the pronunciation corresponding to the audio segment is not included in the acoustic model, and therefore the target word, the confusing word and the acoustic model should be updated according to the audio segment. Identifying the audio clip converts the audio clip into identifying text.
S610, if the target word contains the recognition text, judging whether the pronunciation of the audio clip is correct, and if so, updating the standard acoustic model according to the audio clip; otherwise, updating the error acoustic model according to the audio fragment, prompting the user that the pronunciation of the pronunciation-prone character is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
Specifically, the recognition text is recognized, the recognition text is matched with a target word, if the matching result shows that the recognition text is consistent with a certain target word, whether the pronunciation of the audio clip is correct is judged, if the pronunciation is correct, the audio clip is not recorded in the standard acoustic model possibly due to the fact that accents are less common and the like, therefore, the standard acoustic model is updated according to the audio clip, and then a user is prompted whether the pronunciation of the mispronounced character is correct but not standard, and whether the standard mandarin pronunciation audio needs to be output or not.
If the pronunciation of the audio segment is incorrect, the audio segment is indicated to be a wrong pronunciation of a target word which is not included in the wrong acoustic model, so that the wrong acoustic model is updated according to the audio segment, then the user is prompted to make the pronunciation of the wrong-prone word wrong, and corresponding standard mandarin pronunciation audio is output to correct the user.
S620, if the target word does not contain the identification text, updating the target word according to the identification text, and updating the acoustic model according to the audio fragment.
Specifically, the recognition text is recognized, the recognition text is matched with the target words, if the matching result shows that the recognition text is not matched with all the target words, the recognition text is not included as the target words, therefore, the target words are updated according to the recognition text, and then the acoustic model is updated according to the audio segments.
Step S620, if the target word does not include the identification text, updating the target word according to the identification text, and updating the acoustic model according to the audio piece specifically includes:
s621, when the target word does not contain the recognition text, updating the target word according to the recognition text.
S622, if the pronunciation of the audio segment is correct, updating the standard acoustic model according to the audio segment, updating the confusing word according to the updated target word, and then updating the incorrect acoustic model according to the speech audio of the updated confusing word.
Specifically, after the target word is updated according to the recognition text, the pronunciation of the audio segment is judged, if the pronunciation is correct, the standard acoustic model is updated according to the audio segment, and then other pronunciation audio update standard acoustic models of the recognition text can be obtained.
And then replacing the pronunciation-prone wrong characters in the updated target words with pronunciation-prone wrong characters to correspondingly update the confused words, re-acquiring the voice audio of the updated confused words, and updating the wrong acoustic model according to the re-acquired voice audio.
S623, if the audio segment is mispronounced, acquiring correct voice audio of the recognized text, updating the standard acoustic model according to the correct voice audio, updating the confusion terms according to the updated target terms, and then updating the wrong acoustic model according to the updated voice audio of the confusion terms.
Specifically, after the target word is updated according to the recognized text, the pronunciation of the audio segment is judged, if the pronunciation is incorrect, correct voice audios corresponding to the recognized text, including standard mandarin pronunciation audio and pronunciation audios of other dialects, cavity tones and the like, are obtained, and the standard acoustic model is updated according to the obtained correct voice audios.
And then replacing the pronunciation-prone wrong characters in the updated target words with pronunciation-prone wrong characters to correspondingly update the confused words, re-acquiring the voice audio of the updated confused words, and updating the wrong acoustic model according to the re-acquired voice audio.
In this embodiment, when the extracted audio segment does not match the matching result of the speech audio in the acoustic model, the audio segment is converted into an identification text, the identification text and the audio segment are identified, different situations are classified, the pronunciation of the user is quickly and accurately judged, and then corresponding processing is performed.
A fourth embodiment of the present invention, as shown in fig. 4, is a system 1000 for correcting pronunciation errors in speech recognition, comprising:
the mapping table establishing module 1100 is configured to establish a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronunciation-prone erroneous word.
Specifically, the mapping table establishing module 1100 takes polyphone and chinese characters having a plurality of near-form characters as pronouncing error-prone characters, takes correct pronunciations as a standard acoustic model, takes polyphone and near-form characters as an erroneous acoustic model, and then establishes a mapping table corresponding to each other between the standard acoustic model and the erroneous acoustic model, so as to find corresponding correct pronunciations for correcting the erroneous pronunciations.
The obtaining module 1200 obtains the user voice information.
Specifically, the obtaining module 1200 obtains the user voice information, which may be input by the user in real time, for example, the user needs to perform voice correction when reading aloud with the smart terminal, or the pupil needs to check the learning result of the pupil when learning a new vocabulary through the smart terminal. It is also possible to implement recorded audio, for example, to check whether the pronunciation in the audio recorded by the student is accurate. Prompt correction is performed when the system detects a wrong pronunciation, so the user can autonomously select whether the system or a single application starts the function of prompting correction of the wrong pronunciation.
The extracting module 1300 is configured to identify the user voice information acquired by the acquiring module 1200, and when the voice information includes the pronunciation-prone character, extract an audio segment corresponding to a word including the pronunciation-prone character in the user voice information.
Specifically, the extracting module 1300 converts the acquired user voice information into text information, then identifies whether the text information includes any one or more pronouncing and error-prone characters, analyzes the components of sentences and the parts of speech of the words in the text information by a word segmentation technique if the text information includes any one or more pronouncing and error-prone characters, marks the words including the pronouncing and error-prone characters, and finally extracts one or more audio segments corresponding to the marked words from the user voice information according to the marking result.
The processing module 1400 prompts the user that the pronunciation is prone to wrong words and pronunciation errors when the audio segment extracted by the extraction module 1300 matches the voice audio matching result in the wrong acoustic model in the mapping table establishing module 1100, and outputs the corresponding voice audio in the standard acoustic model according to the mapping table established by the mapping table establishing module 1100.
Specifically, the processing module 1400 finds the speech audio of the pronouncing error-prone character corresponding to the standard acoustic model and the error acoustic model according to the pronouncing error-prone character corresponding to the extracted audio segment, and then matches the extracted audio segment with the found speech audio one by one. If the matching result is matched with the voice audio belonging to the wrong acoustic model, the voice audio is indicated as a part of the pronunciation error-prone character in the voice information of the user, and therefore the user is prompted to pronounce the pronunciation error of the pronunciation error-prone character.
Since the same pronunciation-prone character may appear many times in the user voice information, but pronunciation errors do not occur everywhere, when the processing module 1400 prompts pronunciation errors, the specific position of the pronunciation-prone character needs to be clearly explained in combination with the context of the pronunciation-prone character in the user voice information.
In addition, the correct voice audio corresponding to the word formed by the pronouncing error-prone character at the position is found through a mapping table between the standard acoustic model and the error acoustic model, and the correct voice audio is output to the user, so that the user can correct pronouncing conveniently.
In this embodiment, after the user voice information is acquired, the pronouncing error-prone character included in the user voice information is firstly recognized, and then an audio segment corresponding to a word including the pronouncing error-prone character in the user voice information is extracted to be matched with a standard acoustic model and a voice audio frequency in an error acoustic model. Only effective audio segments are extracted for matching, so that the voice length required to be matched is reduced, the requirement on the matching capacity of the system is lowered, the matching speed is increased, and the matching accuracy is increased.
A fifth embodiment of the present invention is a preferable embodiment of the fourth embodiment, and as shown in fig. 5, the fifth embodiment includes:
the error-prone character acquisition module 1500 acquires the pronunciation error-prone character.
The target word generation module 1600 generates a target word according to the pronunciation error-prone character acquired by the error-prone character acquisition module 1500.
Specifically, the error-prone character acquisition module 1500 uses polyphonic characters and chinese characters with multiple shape-similar characters as pronunciation error-prone characters, and the target term generation module 1600 generates target terms according to the pronunciation error-prone characters, where the target terms are terms including pronunciation error-prone characters, and because the pronunciation error-prone characters are polyphonic characters or have multiple shape-similar characters, and pronunciations of the pronunciation error-prone characters in the target terms are more easily confused, such as shen zi, download, and pressure axis. Or the pronunciation of a part of words does not affect the communication of the daily life of the user, but actually the pronunciation is wrong, such as stubborn, frame and the like, and correction should be given, especially when the elementary school students just start to learn new words, correct pronunciation should be formed from the beginning to avoid later correction.
The audio obtaining module 1700 obtains the voice audio of the target word generated by the target word generating module 1600.
Specifically, the audio obtaining module 1700 obtains a speech audio of each target word, wherein each pronunciation-prone word may correspond to one or more speech audios, wherein each target word obtains an audio of a standard mandarin pronunciation, and subsequently, if the pronunciation of the user needs to be corrected, the audio of the standard mandarin pronunciation is preferentially selected by the system.
However, due to the influence of gender, age, region, dialect accent, cavity tone and the like of the user, the pronunciation of the user is correct but cannot be matched with the pronunciation audio of the standard mandarin, so that besides the audio of the standard mandarin pronunciation, each target word also needs to acquire the audio of the correct pronunciation of people in different ages and regions as much as possible, and the user can independently select whether to prompt and correct the subsequent pronunciation under the condition of correct pronunciation but not standard.
An acoustic model generating module 1800, configured to generate the standard acoustic model according to the speech audio of the target word acquired by the audio acquiring module 1700.
Specifically, the acoustic model generation module 1800 generates a standard acoustic model according to the speech audio of the target word, the standard acoustic model is classified according to the pronouncing error-prone character, and when a certain pronouncing error-prone character generates more target words or a certain target word has more speech audio corresponding to a certain target word, the standard acoustic model can be selected to be classified again according to the target word under the classification of the pronouncing error-prone character.
The confusing word acquiring module 1850 acquires the pronunciation confusing word of the pronunciation confusing word acquired by the error-prone word acquiring module 1500.
The confusing word generating module 1900 replaces the pronunciation-prone character in the target word generated by the target word generating module 1600 with the pronunciation-confusing character acquired by the confusing character acquiring module 1850 to generate the confusing word.
Specifically, the confusing word acquiring module 1850 acquires a confusing pronunciation word of the confusing pronunciation word, where the confusing pronunciation word is a confusing pronunciation word of another pronunciation or a similar word of the confusing pronunciation word, and the confusing word generating module 1900 replaces the confusing pronunciation word in the target word with the confusing pronunciation word to generate the confusing word, that is, combines the confusing pronunciation word of the other pronunciation with the target word or combines the similar word of the confusing pronunciation word with the target word to generate the confusing word.
The confusing words are words corresponding to voices which may be wrongly read by users when reading the target words, and the confusing words do not conform to the word composition rule in composition, and actually do not have the words, and the confusing words are only used for obtaining matched voice audio subsequently.
The audio obtaining module 1700 obtains the speech audio of the confusing word generated by the confusing word generating module 1900.
The acoustic model generation module 1800 generates the error acoustic model according to the speech audio of the confusing word acquired by the audio acquisition module 1700.
Specifically, the audio obtaining module 1700 obtains the voice audio of each confusing word, and similarly, there may exist a plurality of voice audios corresponding to the same confusing word due to the influence of the gender, age, region, dialect accent, and accent of the user. The acoustic model generation module 1800 generates an incorrect acoustic model according to the speech audio of the confusing words, and the incorrect acoustic model is classified according to the corresponding pronunciation-prone character, and when there are more confusing words corresponding to a certain pronunciation-prone character or there are more speech audio corresponding to a certain confusing word, the model can be selected to be classified again according to the confusing words under the classification of pronunciation-prone characters.
The mapping table creating module 1100 creates a mapping table between the standard acoustic model and the error acoustic model corresponding to the pronounceable erroneous word according to the standard acoustic model and the error acoustic model generated by the acoustic model generating module 1800.
The obtaining module 1200 obtains the user voice information.
The extracting module 1300 is configured to identify the user voice information acquired by the acquiring module 1200, and when the voice information includes the pronunciation-prone character, extract an audio segment corresponding to a word including the pronunciation-prone character in the user voice information.
The processing module 1400 prompts the user that the pronunciation is prone to wrong words and pronunciation errors when the audio segment extracted by the extraction module 1300 matches the voice audio matching result in the wrong acoustic model in the mapping table establishing module 1100, and outputs the corresponding voice audio in the standard acoustic model according to the mapping table established by the mapping table establishing module 1100.
The processing module 1400 prompts the user that the pronunciation-prone word pronounces correctly when the audio segment extracted by the extracting module 1300 matches the voice audio matching result in the standard acoustic model in the mapping table establishing module 1100.
Specifically, the processing module 1400 finds the speech audio of the pronouncing error-prone character corresponding to the standard acoustic model and the error acoustic model according to the pronouncing error-prone character corresponding to the extracted audio segment, and then matches the extracted audio segment with the found speech audio one by one.
If the matching result of the processing module 1400 is the speech audio belonging to the standard acoustic model, it indicates that the part of the user speech information related to the mispronounced word is correctly pronounced, but it needs to further determine whether the matching speech audio is the standard mandarin speech audio, if so, it indicates that the user pronunciation is not only correct but also standard, and if not, it prompts the user that the mispronounced word is correctly pronounced but not standard, and asks the user whether the user needs to output the standard mandarin speech audio.
In the embodiment, a standard acoustic model of correct pronunciation of the pronunciation error-prone character and a corresponding incorrect acoustic model of incorrect pronunciation of a combination of the polyphone character and the similar character aiming at the confusable pronunciation of the pronunciation error-prone character are respectively generated, and a mapping table of a corresponding relation between the two acoustic models is established for subsequently identifying the voice information of the user and then carrying out corresponding matching, so that the place of the pronunciation error in the voice information of the user can be accurately identified and corrected.
A sixth embodiment of the present invention is a preferable embodiment of the fourth embodiment, and as shown in fig. 6, the sixth embodiment includes:
the mapping table establishing module 1100 is configured to establish a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronunciation-prone erroneous word.
The obtaining module 1200 obtains the user voice information.
The extracting module 1300 is configured to identify the user voice information acquired by the acquiring module 1200, and when the voice information includes the pronunciation-prone character, extract an audio segment corresponding to a word including the pronunciation-prone character in the user voice information.
The processing module 1400 prompts the user that the pronunciation is prone to wrong words and pronunciation errors when the audio segment extracted by the extraction module 1300 matches the voice audio matching result in the wrong acoustic model in the mapping table establishing module 1100, and outputs the corresponding voice audio in the standard acoustic model according to the mapping table established by the mapping table establishing module 1100.
The processing module 1400, when the audio segment extracted by the extracting module 1300 does not match the voice audio matching result in the acoustic model in the mapping table establishing module 1100, converts the audio segment into a recognition text, where the acoustic model includes the standard acoustic model and the error acoustic model.
Specifically, the processing module 1400 finds the acoustic models, i.e., the standard acoustic model and the speech audio of the pronunciation-prone character corresponding to the error acoustic model, according to the pronunciation-prone character corresponding to the extracted audio segment, and then matches the extracted audio segment with the found speech audio one by one.
If the matching results of the audio segment and the voice audio in the acoustic model are not consistent, it indicates that the pronunciation corresponding to the audio segment is not included in the acoustic model, and therefore the target word, the confusing word and the acoustic model should be updated according to the audio segment. Identifying the audio clip converts the audio clip into identifying text.
A control module 1950, determining whether the pronunciation of the audio clip is correct if the target word includes the recognition text converted by the processing module 1400, and if so, updating the standard acoustic model according to the audio clip; otherwise, updating the error acoustic model according to the audio fragment, prompting the user that the pronunciation of the pronunciation-prone character is wrong, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table.
Specifically, the recognition text is recognized, the recognition text is matched with a target word, if the matching result shows that the recognition text is consistent with a certain target word, whether the pronunciation of the audio clip is correct is judged, if the pronunciation is correct, the audio clip is not recorded in the standard acoustic model possibly due to the fact that accents are less common and the like, therefore, the standard acoustic model is updated according to the audio clip, and then a user is prompted whether the pronunciation of the mispronounced character is correct but not standard, and whether the standard mandarin pronunciation audio needs to be output or not.
If the pronunciation of the audio segment is incorrect, the audio segment is indicated to be a wrong pronunciation of a target word which is not included in the wrong acoustic model, so that the wrong acoustic model is updated according to the audio segment, then the user is prompted to make the pronunciation of the wrong-prone word wrong, and corresponding standard mandarin pronunciation audio is output to correct the user.
The control module 1950, if the target word does not include the recognition text converted by the processing module 1400, updates the target word according to the recognition text and updates the acoustic model according to the audio clip.
Specifically, the recognition text is recognized, the recognition text is matched with the target words, if the matching result shows that the recognition text is not matched with all the target words, the recognition text is not included as the target words, therefore, the target words are updated according to the recognition text, and then the acoustic model is updated according to the audio segments.
The control module 1950 specifically includes:
a target word updating unit 1951, when the target word does not include the recognition text converted by the processing module 1400, updating the target word according to the recognition text.
The control unit 1952, if the audio segment extracted by the extracting module 1300 pronounces correctly, updates the standard acoustic model according to the audio segment, updates the confusing word according to the target word updated by the target word updating unit 1951, and then updates the incorrect acoustic model according to the speech audio of the updated confusing word.
Specifically, after updating the target word according to the recognition text, the control unit 1952 determines the pronunciation of the audio segment, and if the pronunciation is correct, updates the standard acoustic model according to the audio segment, and then may further obtain another pronunciation audio update standard acoustic model of the recognition text.
And then replacing the pronunciation-prone wrong characters in the updated target words with pronunciation-prone wrong characters to correspondingly update the confused words, re-acquiring the voice audio of the updated confused words, and updating the wrong acoustic model according to the re-acquired voice audio.
The control unit 1952, if the audio segment extracted by the extraction module 1300 is mispronounced, obtains a correct speech audio of the recognized text, updates the standard acoustic model according to the correct speech audio, updates the confusing word according to the target word updated by the target word updating unit 1951, and then updates the wrong acoustic model according to the updated speech audio of the confusing word.
Specifically, after updating the target word according to the recognized text, the control unit 1952 determines the pronunciation of the audio segment, and if the pronunciation is incorrect, obtains the correct voice audio corresponding to the recognized text, including the standard mandarin pronunciation audio and pronunciation audio of other dialects, the caval key, etc., and updates the standard acoustic model according to the obtained correct voice audio.
And then replacing the pronunciation-prone wrong characters in the updated target words with pronunciation-prone wrong characters to correspondingly update the confused words, re-acquiring the voice audio of the updated confused words, and updating the wrong acoustic model according to the re-acquired voice audio.
In this embodiment, when the extracted audio segment does not match the matching result of the speech audio in the acoustic model, the audio segment is converted into an identification text, the identification text and the audio segment are identified, different situations are classified, the pronunciation of the user is quickly and accurately judged, and then corresponding processing is performed.
It should be noted that the above embodiments can be freely combined as necessary. The foregoing is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, various modifications and decorations can be made without departing from the principle of the present invention, and these modifications and decorations should also be regarded as the protection scope of the present invention.

Claims (8)

1. A method for correcting pronunciation errors in speech recognition, comprising:
establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to pronounce error-prone characters;
acquiring user voice information;
recognizing the user voice information, and when the pronunciation-prone wrong character is contained in the voice information, extracting an audio segment corresponding to a word containing the pronunciation-prone wrong character in the user voice information;
when the audio frequency fragment is matched with the voice audio frequency matching result in the error acoustic model, prompting a user that the pronunciation is easy to miss, and outputting the corresponding voice audio frequency in the standard acoustic model according to the mapping table;
wherein, before establishing the mapping table between the standard acoustic model and the error acoustic model corresponding to the pronunciation error-prone word, the method further comprises: acquiring the pronouncing error-prone characters, and generating target words according to the pronouncing error-prone characters;
when the audio segment does not accord with the voice audio matching result in the acoustic model, converting the audio segment into recognition text, wherein the acoustic model comprises the standard acoustic model and the error acoustic model;
if the target words contain the recognition texts, judging whether the pronunciation of the audio clips is correct, and if so, updating the standard acoustic model according to the audio clips; otherwise, updating the error acoustic model according to the audio fragment;
and if the target word does not contain the identification text, updating the target word according to the identification text, and updating the acoustic model according to the audio fragment.
2. The method for correcting pronunciation errors in speech recognition according to claim 1, wherein the step of establishing the mapping table between the standard acoustic model and the incorrect acoustic model corresponding to the pronunciation-prone word further comprises:
acquiring the voice audio of the target word, and generating the standard acoustic model according to the voice audio of the target word;
acquiring pronunciation confusion words of the pronunciation easily-wrong words, and replacing the pronunciation easily-wrong words in the target words with the pronunciation confusion words to generate confusion words;
and acquiring the voice audio of the confusing words, and generating the error acoustic model according to the voice audio of the confusing words.
3. The method for correcting pronunciation errors in speech recognition according to claim 1 or 2, further comprising:
and when the audio frequency segment is matched with the voice audio frequency matching result in the standard acoustic model, prompting the user that the pronunciation of the error-prone character is correct.
4. The method according to claim 2, wherein the updating the target word according to the recognized text and the acoustic model according to the audio piece when the target word does not include the recognized text specifically includes:
when the target word does not contain the recognition text, updating the target word according to the recognition text;
if the pronunciation of the audio clip is correct, updating the standard acoustic model according to the audio clip, updating the confusion terms according to the updated target terms, and then updating the error acoustic model according to the voice audio of the updated confusion terms;
and if the audio fragment is mispronounced, acquiring correct voice audio of the recognized text, updating the standard acoustic model according to the correct voice audio, updating the confusion terms according to the updated target terms, and updating the wrong acoustic model according to the updated voice audio of the confusion terms.
5. A system for correcting pronunciation errors in speech recognition, comprising:
the mapping table establishing module is used for establishing a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronounce error-prone character;
the acquisition module acquires user voice information;
the extraction module is used for identifying the user voice information acquired by the acquisition module, and extracting an audio clip corresponding to a word containing the pronouncing error-prone character in the user voice information when the pronunciation error-prone character is contained in the voice information;
the processing module prompts a user that the pronunciation is easy to miss and the pronunciation is wrong when the audio frequency segment extracted by the extraction module is matched with the voice audio frequency matching result in the wrong acoustic model in the mapping table establishing module, and outputs the corresponding voice audio frequency in the standard acoustic model according to the mapping table established by the mapping table establishing module;
the error-prone character acquisition module is used for acquiring the pronunciation error-prone characters before the mapping table establishment module establishes a mapping table between a standard acoustic model and an error acoustic model corresponding to the pronunciation error-prone characters;
the target word generation module is used for generating a target word according to the pronunciation error-prone characters acquired by the error-prone character acquisition module;
the processing module is used for converting the audio clip into an identification text when the audio clip extracted by the extraction module is not consistent with a voice audio matching result in an acoustic model in the mapping table establishing module, wherein the acoustic model comprises the standard acoustic model and the error acoustic model;
the control module is used for judging whether the pronunciation of the audio clip is correct or not if the target word contains the recognition text converted by the processing module, and updating the standard acoustic model according to the audio clip if the pronunciation of the audio clip is correct; otherwise, updating the error acoustic model according to the audio fragment;
and if the target word does not contain the identification text converted by the processing module, the control module updates the target word according to the identification text and updates the acoustic model according to the audio segment.
6. The system for correcting pronunciation errors in speech recognition according to claim 5, further comprising:
the audio acquisition module is used for acquiring the voice audio of the target words generated by the target word generation module;
the acoustic model generating module is used for generating the standard acoustic model according to the voice audio of the target word acquired by the audio acquiring module;
the confusing character acquisition module is used for acquiring the pronunciation confusing characters of the pronunciation confusing characters acquired by the confusing character acquisition module;
a confusion word generation module, which replaces the pronunciation-prone wrong character in the target word generated by the target word generation module with the pronunciation-prone wrong character acquired by the confusion word acquisition module to generate a confusion word;
the audio acquisition module is used for acquiring the voice audio of the confusion word generated by the confusion word generation module;
the acoustic model generating module is used for generating the error acoustic model according to the voice audio of the confusing words acquired by the audio acquiring module.
7. The system for correcting pronunciation errors in speech recognition according to claim 5 or 6, further comprising:
and the processing module prompts a user that the pronunciation of the error-prone character is correct when the audio frequency segment extracted by the extraction module is matched with the voice audio frequency matching result in the standard acoustic model in the mapping table establishing module.
8. The system for correcting pronunciation errors in speech recognition according to claim 6, wherein the control module specifically comprises:
a target word updating unit, which updates the target word according to the recognition text when the target word does not contain the recognition text converted by the processing module;
the control unit is used for updating the standard acoustic model according to the audio segment if the pronunciation of the audio segment extracted by the extraction module is correct, updating the confusion word according to the target word updated by the target word updating unit, and then updating the wrong acoustic model according to the voice audio of the updated confusion word;
the control unit is used for acquiring correct voice audio of the recognized text if the audio segment extracted by the extraction module is in wrong pronunciation, updating the standard acoustic model according to the correct voice audio, updating the confusion word according to the target word updated by the target word updating unit, and then updating the wrong acoustic model according to the updated voice audio of the confusion word.
CN201811239934.5A 2018-10-23 2018-10-23 Method and system for correcting pronunciation errors of voice recognition Active CN109461436B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811239934.5A CN109461436B (en) 2018-10-23 2018-10-23 Method and system for correcting pronunciation errors of voice recognition

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811239934.5A CN109461436B (en) 2018-10-23 2018-10-23 Method and system for correcting pronunciation errors of voice recognition

Publications (2)

Publication Number Publication Date
CN109461436A CN109461436A (en) 2019-03-12
CN109461436B true CN109461436B (en) 2020-12-15

Family

ID=65608234

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811239934.5A Active CN109461436B (en) 2018-10-23 2018-10-23 Method and system for correcting pronunciation errors of voice recognition

Country Status (1)

Country Link
CN (1) CN109461436B (en)

Families Citing this family (10)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110349576A (en) * 2019-05-16 2019-10-18 国网上海市电力公司 Power system operation instruction executing method, apparatus and system based on speech recognition
CN110415679B (en) * 2019-07-25 2021-12-17 北京百度网讯科技有限公司 Voice error correction method, device, equipment and storage medium
CN110600004A (en) * 2019-09-09 2019-12-20 腾讯科技(深圳)有限公司 Voice synthesis playing method and device and storage medium
CN111353066B (en) * 2020-02-20 2023-11-21 联想(北京)有限公司 Information processing method and electronic equipment
CN113920803B (en) * 2020-07-10 2024-05-10 上海流利说信息技术有限公司 Error feedback method, device, equipment and readable storage medium
CN112259092B (en) * 2020-10-15 2023-09-01 深圳市同行者科技有限公司 Voice broadcasting method and device and voice interaction equipment
CN112786052B (en) * 2020-12-30 2024-05-31 科大讯飞股份有限公司 Speech recognition method, electronic equipment and storage device
CN113672144A (en) * 2021-09-06 2021-11-19 北京搜狗科技发展有限公司 Data processing method and device
CN113938708B (en) * 2021-10-14 2024-04-09 咪咕文化科技有限公司 Live audio error correction method, device, computing equipment and storage medium
CN116894442B (en) * 2023-09-11 2023-12-05 临沂大学 Language translation method and system for correcting guide pronunciation

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103853702A (en) * 2012-12-06 2014-06-11 富士通株式会社 Device and method for correcting idiom error in linguistic data
CN105302795A (en) * 2015-11-11 2016-02-03 河海大学 Chinese text verification system and method based on Chinese vague pronunciation and voice recognition
CN105374356A (en) * 2014-08-29 2016-03-02 株式会社理光 Speech recognition method, speech assessment method, speech recognition system, and speech assessment system
CN107741928A (en) * 2017-10-13 2018-02-27 四川长虹电器股份有限公司 A kind of method to text error correction after speech recognition based on field identification
CN107993653A (en) * 2017-11-30 2018-05-04 南京云游智能科技有限公司 The incorrect pronunciations of speech recognition apparatus correct update method and more new system automatically

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8494852B2 (en) * 2010-01-05 2013-07-23 Google Inc. Word-level correction of speech input

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103853702A (en) * 2012-12-06 2014-06-11 富士通株式会社 Device and method for correcting idiom error in linguistic data
CN105374356A (en) * 2014-08-29 2016-03-02 株式会社理光 Speech recognition method, speech assessment method, speech recognition system, and speech assessment system
CN105302795A (en) * 2015-11-11 2016-02-03 河海大学 Chinese text verification system and method based on Chinese vague pronunciation and voice recognition
CN107741928A (en) * 2017-10-13 2018-02-27 四川长虹电器股份有限公司 A kind of method to text error correction after speech recognition based on field identification
CN107993653A (en) * 2017-11-30 2018-05-04 南京云游智能科技有限公司 The incorrect pronunciations of speech recognition apparatus correct update method and more new system automatically

Also Published As

Publication number Publication date
CN109461436A (en) 2019-03-12

Similar Documents

Publication Publication Date Title
CN109461436B (en) Method and system for correcting pronunciation errors of voice recognition
CN108447486B (en) Voice translation method and device
KR102191425B1 (en) Apparatus and method for learning foreign language based on interactive character
US6424935B1 (en) Two-way speech recognition and dialect system
CN110648690B (en) Audio evaluation method and server
US8498857B2 (en) System and method for rapid prototyping of existing speech recognition solutions in different languages
CN109410664B (en) Pronunciation correction method and electronic equipment
US7529678B2 (en) Using a spoken utterance for disambiguation of spelling inputs into a speech recognition system
JP3520022B2 (en) Foreign language learning device, foreign language learning method and medium
CN108431883B (en) Language learning system and language learning program
KR20160122542A (en) Method and apparatus for measuring pronounciation similarity
CN108806719A (en) Interacting language learning system and its method
CN113793593B (en) Training data generation method and device suitable for speech recognition model
CN113362817A (en) Speech recognition error correction device, speech recognition error correction method, and speech recognition error correction program
CN109166569B (en) Detection method and device for phoneme mislabeling
JPH06110494A (en) Pronounciation learning device
CN113299266A (en) Data generating device, data generating method, and recording medium
Kruse et al. Alinha-pb: A phonetic aligner for brazilian portuguese
KR20150103809A (en) Method and apparatus for studying simillar pronounciation
CN111951827B (en) Continuous reading identification correction method, device, equipment and readable storage medium
KR20170056253A (en) Method of and system for scoring pronunciation of learner
KR20090109501A (en) System and Method for Rhythm Training in Language Learning
JP3378547B2 (en) Voice recognition method and apparatus
KR102405547B1 (en) Pronunciation evaluation system based on deep learning
KR20140068292A (en) Speaking Training System for Improving Fluency of Utterence

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant