CN109461459A - Speech assessment method, apparatus, computer equipment and storage medium - Google Patents

Speech assessment method, apparatus, computer equipment and storage medium Download PDF

Info

Publication number
CN109461459A
CN109461459A CN201811493385.4A CN201811493385A CN109461459A CN 109461459 A CN109461459 A CN 109461459A CN 201811493385 A CN201811493385 A CN 201811493385A CN 109461459 A CN109461459 A CN 109461459A
Authority
CN
China
Prior art keywords
pinyin information
information
user
pinyin
matching degree
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201811493385.4A
Other languages
Chinese (zh)
Inventor
贾克尧
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Ping An Technology Shenzhen Co Ltd
Original Assignee
Ping An Technology Shenzhen Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Ping An Technology Shenzhen Co Ltd filed Critical Ping An Technology Shenzhen Co Ltd
Priority to CN201811493385.4A priority Critical patent/CN109461459A/en
Publication of CN109461459A publication Critical patent/CN109461459A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/69Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for evaluating synthetic or decoded voice signals
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification techniques
    • G10L17/04Training, enrolment or model building
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/48Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
    • G10L25/51Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination

Landscapes

  • Engineering & Computer Science (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Computational Linguistics (AREA)
  • Signal Processing (AREA)
  • Machine Translation (AREA)

Abstract

The invention discloses speech assessment method, apparatus, computer equipment and storage mediums.If method includes: to receive matching degree measurement request, from one text sentence of random selection as target text sentence;Typing prompt information is generated according to target text sentence to prompt user's typing voice messaging corresponding with target text sentence;If receiving user according to the voice messaging of typing prompt information institute typing, the voice messaging is converted to by the first Pinyin information by default voice transformation model;Target text sentence is converted into the second Pinyin information by pre-set text transformation model;The matching degree between the first Pinyin information and the second Pinyin information is calculated according to preset matching degree computation model to obtain the speech assessment of the user.The present invention is based on voice processing technologies, and the speech assessment of user is obtained by calculation, and can accurately obtain the matching degree between the voice messaging and mandarin that user is inputted, and meet the actual use demand of user.

Description

Speech assessment method, apparatus, computer equipment and storage medium
Technical field
The present invention relates to field of computer technology more particularly to a kind of speech assessment method, apparatus, computer equipment and deposit Storage media.
Background technique
The range of intelligent sound technical application is more and more extensive, the need pair when passing through intelligent sound technology input voice information The voice messaging inputted is identified, is not suitable for other dialects since speech recognition is only applicable to mandarin, and existing Voice information identification method can not obtain the matching degree between user institute input voice information and mandarin, and inputted in user Voice messaging and mandarin between matching degree it is not high in the case where, significantly reduce to user's institute's input voice information carry out The accuracy rate of identification.Thus, existing voice information identification method presence can not judge to match between voice messaging and mandarin The problem of spending.
Summary of the invention
The embodiment of the invention provides a kind of speech assessment method, apparatus, computer equipment and storage mediums, it is intended to solve The art methods the problem of matching degree between voice messaging and mandarin can not being judged.
In a first aspect, the embodiment of the invention provides a kind of speech assessment methods comprising:
If receiving matching degree measurement request, a text sentence is randomly choosed from pre-set text statement library as target Text sentence;
Typing prompt information is generated according to the target text sentence to prompt user's typing and the target text sentence Corresponding voice messaging;
If receiving the user according to the voice messaging of typing prompt information institute typing, pass through default voice conversion The voice messaging is converted to the first Pinyin information by model;
The target text sentence is converted into the second Pinyin information by pre-set text transformation model;
According to preset matching degree computation model to the matching between first Pinyin information and second Pinyin information Degree is calculated to obtain the speech assessment of the user.
Second aspect, the embodiment of the invention provides a kind of speech assessment devices comprising:
Target text sentence acquiring unit, if for receive matching degree measurement request, from pre-set text statement library with Machine selects a text sentence as target text sentence;
Prompt information generation unit, for generating typing prompt information according to the target text sentence to prompt user to record Enter voice messaging corresponding with the target text sentence;
First Pinyin information acquiring unit, if for receiving the user according to typing prompt information institute typing The voice messaging is converted to the first Pinyin information by default voice transformation model by voice messaging;
Second Pinyin information acquiring unit, for being converted to the target text sentence by pre-set text transformation model Second Pinyin information;
Speech assessment acquiring unit, for according to preset matching degree computation model to first Pinyin information and described the Matching degree between two Pinyin informations is calculated to obtain the speech assessment of the user.
The third aspect, the embodiment of the present invention provide a kind of computer equipment again comprising memory, processor and storage On the memory and the computer program that can run on the processor, the processor execute the computer program Speech assessment method described in the above-mentioned first aspect of Shi Shixian.
Fourth aspect, the embodiment of the invention also provides a kind of computer readable storage mediums, wherein the computer can It reads storage medium and is stored with computer program, it is above-mentioned that the computer program when being executed by a processor executes the processor Speech assessment method described in first aspect.
The embodiment of the invention provides a kind of speech assessment method, apparatus, computer equipment and storage mediums.By random Selection obtains target text sentence, converts to the voice messaging of user institute typing and spelling that target text sentence is converted to Message breath, which carries out matching degree calculating, can obtain the inputted voice of user by speech assessment to obtain the speech assessment of user Matching degree between information and mandarin can accurately calculate the matching degree between voice messaging and mandarin, meet The actual use demand of user.
Detailed description of the invention
Technical solution in order to illustrate the embodiments of the present invention more clearly, below will be to needed in embodiment description Attached drawing is briefly described, it should be apparent that, drawings in the following description are some embodiments of the invention, general for this field For logical technical staff, without creative efforts, it is also possible to obtain other drawings based on these drawings.
Fig. 1 is the flow diagram of speech assessment method provided in an embodiment of the present invention;
Fig. 2 is the sub-process schematic diagram of speech assessment method provided in an embodiment of the present invention;
Fig. 3 is another sub-process schematic diagram of speech assessment method provided in an embodiment of the present invention;
Fig. 4 is another sub-process schematic diagram of speech assessment method provided in an embodiment of the present invention;
Fig. 5 is another flow diagram of speech assessment method provided in an embodiment of the present invention;
Fig. 6 is the schematic block diagram of speech assessment device provided in an embodiment of the present invention;
Fig. 7 is the subelement schematic block diagram of speech assessment device provided in an embodiment of the present invention;
Fig. 8 is another subelement schematic block diagram of speech assessment device provided in an embodiment of the present invention;
Fig. 9 is another subelement schematic block diagram of speech assessment device provided in an embodiment of the present invention;
Figure 10 is another schematic block diagram of speech assessment device provided in an embodiment of the present invention;
Figure 11 is the schematic block diagram of computer equipment provided in an embodiment of the present invention.
Specific embodiment
Following will be combined with the drawings in the embodiments of the present invention, and technical solution in the embodiment of the present invention carries out clear, complete Site preparation description, it is clear that described embodiments are some of the embodiments of the present invention, instead of all the embodiments.Based on this hair Embodiment in bright, every other implementation obtained by those of ordinary skill in the art without making creative efforts Example, shall fall within the protection scope of the present invention.
It should be appreciated that ought use in this specification and in the appended claims, term " includes " and "comprising" instruction Described feature, entirety, step, operation, the presence of element and/or component, but one or more of the other feature, whole is not precluded Body, step, operation, the presence or addition of element, component and/or its set.
It is also understood that mesh of the term used in this description of the invention merely for the sake of description specific embodiment And be not intended to limit the present invention.As description of the invention and it is used in the attached claims, unless on Other situations are hereafter clearly indicated, otherwise " one " of singular, "one" and "the" are intended to include plural form.
It will be further appreciated that the term "and/or" used in description of the invention and the appended claims is Refer to any combination and all possible combinations of one or more of associated item listed, and including these combinations.
Referring to Fig. 1, Fig. 1 is the flow diagram of speech assessment method provided in an embodiment of the present invention.The speech assessment Method is applied in user terminal, and user terminal is to be scored with realizing voice messaging for executing speech assessment method Terminal device, such as mobile phone, tablet computer, desktop computer, laptop etc..
As shown in Figure 1, the method comprising the steps of S110~S150.
If S110, receiving matching degree measurement request, a text sentence is randomly choosed from pre-set text statement library and is made For target text sentence.
If receiving the matching degree measurement request that user is inputted, randomly choosed from the pre-set text statement library of user terminal One text sentence is as target text sentence.Matching degree measurement request is user (user of user terminal) at user's end The solicited message measured to the speech assessment of the user inputted in end, pre-set text statement library are in user terminal It is default for storing the statement library of text sentence, it is for selection comprising a plurality of text sentence in pre-set text sentence.To increase The accuracy that the speech assessment of user is measured by force, it is pre- from comprising a plurality of text sentence by randomly selected mode If the occasional case in measurement process can be reduced as target text sentence by selecting one in text sentence library, to improve measurement Accuracy.
S120, typing prompt information is generated to prompt user's typing and the target text according to the target text sentence The corresponding voice messaging of sentence.
Typing prompt information is generated according to the target text sentence to prompt user's typing and the target text sentence Corresponding voice messaging, specifically, generating the prompt information of typing voice messaging and in user in the target text sentence It is shown in terminal, to prompt the user corresponding voice messaging of typing, user in the target text sentence to watch the prompt Information simultaneously carries out typing to voice messaging according to the content of prompt information.
For example, generating corresponding prompt information is " please to be read aloud with mandarin following if target text sentence is " XXXX " Content: XXXX ".
If S130, receiving the user according to the voice messaging of typing prompt information institute typing, by presetting language The voice messaging is converted to the first Pinyin information by sound transformation model.
If receiving the user according to the voice messaging of typing prompt information institute typing, pass through default voice conversion The voice messaging is converted to the first Pinyin information by model.For the voice messaging to the typing of user institute whether with corresponding text language Sentence, which matches, to be judged, need to be by the way that the voice messaging of user institute typing is converted to the first Pinyin information, voice transformation model It is the concrete model for voice messaging to be converted to the first Pinyin information, specifically includes acoustic model in voice transformation model With phonetic feature dictionary.
In one embodiment, as shown in Fig. 2, step S130 includes sub-step S131 and S132.
S131, cutting is carried out to voice messaging to obtain institute in voice messaging according to the acoustic model in voice transformation model The multiple phonemes for including.
Cutting is carried out to obtain included in voice messaging to voice messaging according to the acoustic model in voice transformation model Multiple phonemes.Specifically, the voice messaging of user institute typing is made of, the sound of a character the phoneme of multiple character sounds Element includes the frequency and tone color of the character sound.It include the phoneme of all character sounds in acoustic model, by by audio-frequency information It is matched with phoneme all in acoustic model, cutting can be carried out to the phoneme of character single in audio-frequency information, by cutting Divide and finally obtains multiple phonemes included in the audio-frequency information.
S132, obtained phoneme is matched with by all sounds according to the phonetic feature dictionary in Text region model Element carries out being converted to the first Pinyin information.
According to the phonetic feature dictionary in Text region model to obtained phoneme matched with by all phonemes into Row is converted to the first Pinyin information.It include the corresponding phoneme information of all character phonetics in phonetic feature dictionary, by by institute Obtained phoneme phoneme information corresponding with character phonetic is matched, and can be phonetic feature by the phoneme conversion of single character The character phonetic to match in dictionary with the phoneme, obtaining the corresponding character phonetic of all characters in the voice messaging can be real It is now the first Pinyin information by all phoneme conversions included in voice messaging.
S140, the target text sentence is converted to by the second Pinyin information by pre-set text transformation model.
The target text sentence is converted into the second Pinyin information by pre-set text transformation model.Pre-set text conversion Model is the model for target text sentence to be converted to the second Pinyin information, the Chinese pronunciation of each Chinese character or number Phonetic can be led to be indicated, include the Pinyin information of all characters in text conversion model, by by target text sentence Included in Chinese character or number matched with the character in text conversion model, target text sentence can be converted to the Two Pinyin informations.
In one embodiment, as shown in figure 3, step S140 includes sub-step S141 and S142.
S141, the Pinyin information that each character in target text sentence is obtained by pre-set text transformation model.
The Pinyin information of each character in target text sentence is obtained by pre-set text transformation model.Specifically, passing through Each Chinese character in target text sentence or number are matched with the character in text conversion model, the corresponding Chinese can be obtained The Pinyin information of word or number in text conversion model, that is to say the phonetic for acquiring each character in target text sentence Information.
For example, the Pinyin information for acquiring " fruit " is " gu ǒ ", the Pinyin information of "and" is " h é/h ú/hu ó/hu ò ".
S142, obtained Pinyin information is combined by pre-set text transformation model to obtain the second phonetic letter Breath.
Obtained character is spelled in position of each character in target text sentence by pre-set text transformation model Sound is combined to obtain the second Pinyin information.Specifically, if some character is not polyphone, directly by the reading of the character Sound is added to the character the location of in target text sentence;If some character is polyphone, need through default text This transformation model is by carrying out part of speech analysis to the character and front and back character, to select and the word from a variety of pronunciations of the character It accords with a kind of pronunciation that current part of speech matches and is added to the character the location of in target text sentence.
For example, if in target text sentence a certain section be " apple and grape ", due to "and" be polyphone, then to its front and back Character carries out part of speech analysis, and apple and grape are noun, then the part of speech of "and" at this time is determined by pre-set text transformation model For structural auxiliary word, the Pinyin information when "and" is as structural auxiliary word is " h é ".
S150, according to preset matching degree computation model between first Pinyin information and second Pinyin information Matching degree is calculated to obtain the speech assessment of the user.
The matching degree between the first Pinyin information and the second Pinyin information is counted according to preset matching degree computation model It calculates to obtain the matching degree between user institute input voice information and mandarin, that is to say the speech assessment of the user.Specifically, Matching degree computation model is to spell to the model that matching degree is calculated between the first Pinyin information and the second Pinyin information, first Message breath is the Pinyin information being converted to according to user institute input voice information, and the second Pinyin information is respective objects text language The Pinyin information of sentence standard Chinese pronunciation, by calculating the matching degree between the first Pinyin information and the second Pinyin information, The matching degree between user's institute's input voice information and the standard Chinese pronunciation of respective objects text sentence can be obtained, that is to say Obtain the speech assessment of the user.
In one embodiment, as shown in figure 4, step S150 includes sub-step S151, S152 and S153.
S151, it is whether equal to the quantity of character phonetic included in first Pinyin information and the second Pinyin information into Row judgement.
It, need to be to included in the first Pinyin information and the second Pinyin information before the speech assessment to user calculates Whether the quantity of character is equal to be judged, is spelled specifically, obtaining character in the first Pinyin information and the second Pinyin information respectively The quantity of sound simultaneously carries out judging that judging result can be obtained.
If judging result be character included in the first Pinyin information and the second Pinyin information quantity it is unequal, to The prompt information of family sending voice messaging typing mistake.It can show that the voice messaging typing is wrong in the display equipment of user terminal Prompt information accidentally can be according to the content of prompt information typing voice messaging again after user receives prompt information.
If the quantity of character phonetic included in S152, the first Pinyin information and the second Pinyin information is equal, first is spelled Each character phonetic is matched with respective symbols phonetic in the second Pinyin information to obtain the matching of character phonetic in message breath Number P.
If the quantity of character phonetic included in the first Pinyin information and the second Pinyin information is equal, by the first Pinyin information In each character phonetic matched with respective symbols phonetic in the second Pinyin information, character phonetic coupling number P can be obtained. The quantity of the character phonetic as included in the first Pinyin information and the second Pinyin information is equal, every in the first Pinyin information An one character phonetic character phonetic corresponding with the second Pinyin information is corresponding, specifically, obtaining the first phonetic letter First character phonetic is matched with first character phonetic in the second Pinyin information in breath, if in the first Pinyin information One character phonetic matches with respective symbols phonetic in the second Pinyin information, then coupling number adds one;If not matching that, obtain Character late phonetic is matched with phase character phonetic in the second Pinyin information in first Pinyin information, until the first phonetic Last character phonetic completes matching in information, and obtaining coupling number at this time is final character phonetic coupling number P.
S153, according to the matching degree calculation formula in matching degree computation model: matching degree S=P/P0The use is calculated The speech assessment at family, wherein P0For the quantity of character phonetic included in the second Pinyin information.
According to the matching degree calculation formula in matching degree computation model: matching degree S=P/P0The user institute is calculated Matching degree between input voice information and mandarin that is to say the speech assessment of the user, wherein P0For the second Pinyin information Included in character phonetic quantity, P be character phonetic coupling number.
Based on the above speech assessment method, the voice messaging inputted to user is scored after obtaining speech assessment, It can be further compared by the way that matching degree threshold value is arranged in the user terminal with resulting speech assessment, if obtained voice Scoring is less than matching degree threshold value, then does not carry out speech recognition to the voice messaging to be identified that user subsequently inputs;If obtained Speech assessment is not less than matching degree threshold value, and the voice messaging to be identified that can be subsequently input to user carries out speech recognition, namely It is that the voice messaging to be identified inputted again to the user after the speech assessment of user reaches certain and requires carries out voice knowledge Not, the accuracy rate that voice messaging to be identified is identified is inputted to user to greatly improve.
In addition, the above speech assessment method also can be directly used for being trained the mandarin level of user itself, such as User that is to say by obtaining speech assessment and obtain its matching degree between input voice information and mandarin, can understand Understand the voice messaging that oneself is inputted whether standard, so as to be instructed by mandarin level of the above method to its own Practice.
In one embodiment, as shown in figure 5, further including step S160 after step S150.
S160, the character phonetic not matched that in first Pinyin information with second Pinyin information is obtained to generate Non-matching prompt information.
It obtains the character phonetic not matched that in the first Pinyin information with the second Pinyin information and generates non-matching prompt letter Breath shown in the user terminal, with the standard Chinese pronunciation of the voice messaging that prompts user to be inputted and target text sentence it Between the character phonetic that does not match that, user watch the non-matching prompt information can obtain when subsequent progress voice input it is required into The improved part of row.
Target text sentence is obtained by random selection, convert simultaneously target text to the voice messaging of user institute typing The Pinyin information that sentence is converted to, which carries out matching degree calculating, can be obtained with obtaining the speech assessment of user by speech assessment Take the matching degree between family institute input voice information and mandarin, can to the matching degree between voice messaging and mandarin into Row is accurate to be calculated, and the actual use demand of user is met.
The embodiment of the present invention also provides a kind of speech assessment device, and the speech assessment device is for executing aforementioned voice scoring Any embodiment of method.Specifically, referring to Fig. 6, Fig. 6 is the schematic of speech assessment device provided in an embodiment of the present invention Block diagram.The speech assessment device can be configured in the terminal devices such as desktop computer, laptop, tablet computer or mobile phone.
As shown in fig. 6, speech assessment device 100 includes target text sentence acquiring unit 110, prompt information generation unit 120, the first Pinyin information acquiring unit 130, the second Pinyin information acquiring unit 140 and speech assessment acquiring unit 150.
Target text sentence acquiring unit 110, if for receiving matching degree measurement request, from pre-set text statement library A text sentence is randomly choosed as target text sentence.
If receiving the matching degree measurement request that user is inputted, randomly choosed from the pre-set text statement library of user terminal One text sentence is as target text sentence.Matching degree measurement request is user (user of user terminal) at user's end The solicited message measured to the speech assessment of the user inputted in end, pre-set text statement library are in user terminal It is default for storing the statement library of text sentence, it is for selection comprising a plurality of text sentence in pre-set text sentence.To increase The accuracy that the speech assessment of user is measured by force, it is pre- from comprising a plurality of text sentence by randomly selected mode If the occasional case in measurement process can be reduced as target text sentence by selecting one in text sentence library, to improve measurement Accuracy.
Prompt information generation unit 120, for generating typing prompt information according to the target text sentence to prompt to use Family typing voice messaging corresponding with the target text sentence.
Typing prompt information is generated according to the target text sentence to prompt user's typing and the target text sentence Corresponding voice messaging, specifically, generating the prompt information of typing voice messaging and in user in the target text sentence It is shown in terminal, to prompt the user corresponding voice messaging of typing, user in the target text sentence to watch the prompt Information simultaneously carries out typing to voice messaging according to the content of prompt information.
First Pinyin information acquiring unit 130, if being recorded for receiving the user according to the typing prompt information The voice messaging is converted to the first Pinyin information by default voice transformation model by the voice messaging entered.
If receiving the user according to the voice messaging of typing prompt information institute typing, pass through default voice conversion The voice messaging is converted to the first Pinyin information by model.For the voice messaging to the typing of user institute whether with corresponding text language Sentence, which matches, to be judged, need to be by the way that the voice messaging of user institute typing is converted to the first Pinyin information, voice transformation model It is the concrete model for voice messaging to be converted to the first Pinyin information, specifically includes acoustic model in voice transformation model With phonetic feature dictionary.
In other inventive embodiments, as shown in fig. 7, the first Pinyin information acquiring unit 130 includes subelement: phoneme Acquiring unit 131 and phoneme matching unit 132.
Phoneme acquiring unit 131, for according to the acoustic model in voice transformation model to voice messaging carry out cutting with Obtain multiple phonemes included in voice messaging.
Cutting is carried out to obtain included in voice messaging to voice messaging according to the acoustic model in voice transformation model Multiple phonemes.Specifically, the voice messaging of user institute typing is made of, the sound of a character the phoneme of multiple character sounds Element includes the frequency and tone color of the character sound.It include the phoneme of all character sounds in acoustic model, by by audio-frequency information It is matched with phoneme all in acoustic model, cutting can be carried out to the phoneme of character single in audio-frequency information, by cutting Divide and finally obtains multiple phonemes included in the audio-frequency information.
Phoneme matching unit 132, for according to the phonetic feature dictionary in Text region model to obtained phoneme into Row matching is to carry out all phonemes to be converted to the first Pinyin information.
According to the phonetic feature dictionary in Text region model to obtained phoneme matched with by all phonemes into Row is converted to the first Pinyin information.It include the corresponding phoneme information of all character phonetics in phonetic feature dictionary, by by institute Obtained phoneme phoneme information corresponding with character phonetic is matched, and can be phonetic feature by the phoneme conversion of single character The character phonetic to match in dictionary with the phoneme, obtaining the corresponding character phonetic of all characters in the voice messaging can be real It is now the first Pinyin information by all phoneme conversions included in voice messaging.
Second Pinyin information acquiring unit 140, for being turned the target text sentence by pre-set text transformation model It is changed to the second Pinyin information.
The target text sentence is converted into the second Pinyin information by pre-set text transformation model.Pre-set text conversion Model is the model for target text sentence to be converted to the second Pinyin information, the Chinese pronunciation of each Chinese character or number Phonetic can be led to be indicated, include the Pinyin information of all characters in text conversion model, by by target text sentence Included in Chinese character or number matched with the character in text conversion model, target text sentence can be converted to the Two Pinyin informations.
In other inventive embodiments, as shown in figure 8, the second Pinyin information acquiring unit 140 includes subelement: character Pinyin information acquiring unit 141 and Pinyin information assembled unit 142.
Character Pinyin information acquiring unit 141, it is every in target text sentence for being obtained by pre-set text transformation model The Pinyin information of one character.
The Pinyin information of each character in target text sentence is obtained by pre-set text transformation model.Specifically, passing through Each Chinese character in target text sentence or number are matched with the character in text conversion model, the corresponding Chinese can be obtained The Pinyin information of word or number in text conversion model, that is to say the phonetic for acquiring each character in target text sentence Information.
Pinyin information assembled unit 142, for carrying out group to obtained Pinyin information by pre-set text transformation model It closes to obtain the second Pinyin information.
Obtained character is spelled in position of each character in target text sentence by pre-set text transformation model Sound is combined to obtain the second Pinyin information.Specifically, if some character is not polyphone, directly by the reading of the character Sound is added to the character the location of in target text sentence;If some character is polyphone, need through default text This transformation model is by carrying out part of speech analysis to the character and front and back character, to select and the word from a variety of pronunciations of the character It accords with a kind of pronunciation that current part of speech matches and is added to the character the location of in target text sentence.
Speech assessment acquiring unit 150 is used for according to preset matching degree computation model to first Pinyin information and institute The matching degree between the second Pinyin information is stated to be calculated to obtain the speech assessment of the user.
The matching degree between the first Pinyin information and the second Pinyin information is counted according to preset matching degree computation model It calculates to obtain the matching degree between user institute input voice information and mandarin, that is to say the speech assessment of the user.Specifically, Matching degree computation model is to spell to the model that matching degree is calculated between the first Pinyin information and the second Pinyin information, first Message breath is the Pinyin information being converted to according to user institute input voice information, and the second Pinyin information is respective objects text language The Pinyin information of sentence standard Chinese pronunciation, by calculating the matching degree between the first Pinyin information and the second Pinyin information, The matching degree between user's institute's input voice information and the standard Chinese pronunciation of respective objects text sentence can be obtained, that is to say Obtain the speech assessment of the user.
In other inventive embodiments, as shown in figure 9, the speech assessment acquiring unit 150 includes subelement: judging unit 151, character phonetic coupling number acquiring unit 152 and scoring computing unit 153.
Judging unit 151, for the number to character phonetic included in first Pinyin information and the second Pinyin information Whether amount is equal to be judged.
It, need to be to included in the first Pinyin information and the second Pinyin information before the speech assessment to user calculates Whether the quantity of character is equal to be judged, is spelled specifically, obtaining character in the first Pinyin information and the second Pinyin information respectively The quantity of sound simultaneously carries out judging that judging result can be obtained.
If judging result be character included in the first Pinyin information and the second Pinyin information quantity it is unequal, to The prompt information of family sending voice messaging typing mistake.It can show that the voice messaging typing is wrong in the display equipment of user terminal Prompt information accidentally can be according to the content of prompt information typing voice messaging again after user receives prompt information.
Character phonetic coupling number acquiring unit 152, if for word included in the first Pinyin information and the second Pinyin information Accord with phonetic quantity it is equal, by respective symbols phonetic in each character phonetic in the first Pinyin information and the second Pinyin information into Row matching is to obtain character phonetic coupling number P.
If the quantity of character phonetic included in the first Pinyin information and the second Pinyin information is equal, by the first Pinyin information In each character phonetic matched with respective symbols phonetic in the second Pinyin information, character phonetic coupling number P can be obtained. The quantity of the character phonetic as included in the first Pinyin information and the second Pinyin information is equal, every in the first Pinyin information An one character phonetic character phonetic corresponding with the second Pinyin information is corresponding, specifically, obtaining the first phonetic letter First character phonetic is matched with first character phonetic in the second Pinyin information in breath, if in the first Pinyin information One character phonetic matches with respective symbols phonetic in the second Pinyin information, then coupling number adds one;If not matching that, obtain Character late phonetic is matched with phase character phonetic in the second Pinyin information in first Pinyin information, until the first phonetic Last character phonetic completes matching in information, and obtaining coupling number at this time is final character phonetic coupling number P.
Score computing unit 153, for according to the matching degree calculation formula in matching degree computation model: matching degree S=P/ P0The speech assessment of the user is calculated, wherein P0For the quantity of character phonetic included in the second Pinyin information.
According to the matching degree calculation formula in matching degree computation model: matching degree S=P/P0The user institute is calculated Matching degree between input voice information and mandarin that is to say the speech assessment of the user, wherein P0For the second Pinyin information Included in character phonetic quantity, P be character phonetic coupling number.
By being arranged matching degree threshold value in the user terminal, and the voice messaging that user is inputted by the above method into Row scoring obtain speech assessment, if obtained speech assessment be less than matching degree threshold value, not to user subsequently input wait know Other voice messaging carries out speech recognition;If obtained speech assessment is not less than matching degree threshold value, user can be subsequently input Voice messaging to be identified carry out speech recognition, that is to say after the speech assessment of user reaches certain and requires again to the user The voice messaging to be identified that is inputted carries out speech recognition, inputs voice messaging to be identified to greatly improve to user and knows Other accuracy rate.
In addition, user that is to say by obtaining speech assessment and obtain its between input voice information and mandarin With degree, i.e., it will be clearly understood that the voice messaging oneself inputted whether standard, therefore user can be by the above method to its own Mandarin level be trained.Teacher can issue the language reading aloud operation and being read aloud using above-mentioned speech assessment method student Message breath scores, by the matching degree read aloud between operation and standard Chinese pronunciation by being completed in a manner of objective to student It is calculated.
In other inventive embodiments, as shown in Figure 10, the speech assessment device 100 further includes subelement: non-matching to mention Show information generating unit 160.
Non-matching prompt information generation unit 160 is believed for obtaining in first Pinyin information with second phonetic The character phonetic not matched that is ceased to generate non-matching prompt information.
It obtains the character phonetic not matched that in the first Pinyin information with the second Pinyin information and generates non-matching prompt letter Breath shown in the user terminal, with the standard Chinese pronunciation of the voice messaging that prompts user to be inputted and target text sentence it Between the character phonetic that does not match that, user watch the non-matching prompt information can obtain when subsequent progress voice input it is required into The improved part of row.
Target text sentence is obtained by random selection, convert simultaneously target text to the voice messaging of user institute typing The Pinyin information that sentence is converted to, which carries out matching degree calculating, can be obtained with obtaining the speech assessment of user by speech assessment Take the matching degree between family institute input voice information and mandarin, can to the matching degree between voice messaging and mandarin into Row is accurate to be calculated, and the actual use demand of user is met.
Above-mentioned speech assessment device can be implemented as the form of computer program, which can be in such as Figure 11 institute It is run in the computer equipment shown.
Figure 11 is please referred to, Figure 11 is the schematic block diagram of computer equipment provided in an embodiment of the present invention.
Refering to fig. 11, which includes processor 502, memory and the net connected by system bus 501 Network interface 505, wherein memory may include non-volatile memory medium 503 and built-in storage 504.
The non-volatile memory medium 503 can storage program area 5031 and computer program 5032.The computer program 5032 are performed, and processor 502 may make to execute speech assessment method.
The processor 502 supports the operation of entire computer equipment 500 for providing calculating and control ability.
The built-in storage 504 provides environment for the operation of the computer program 5032 in non-volatile memory medium 503, should When computer program 5032 is executed by processor 502, processor 502 may make to execute speech assessment method.
The network interface 505 is for carrying out network communication, such as the transmission of offer data information.Those skilled in the art can To understand, structure shown in Figure 11, only the block diagram of part-structure relevant to the present invention program, is not constituted to this hair The restriction for the computer equipment 500 that bright scheme is applied thereon, specific computer equipment 500 may include than as shown in the figure More or fewer components perhaps combine certain components or with different component layouts.
Wherein, the processor 502 is for running computer program 5032 stored in memory, to realize following function Can: if receiving matching degree measurement request, a text sentence is randomly choosed from pre-set text statement library as target text Sentence;Typing prompt information is generated according to the target text sentence to prompt user's typing opposite with the target text sentence The voice messaging answered;If receiving the user according to the voice messaging of typing prompt information institute typing, by presetting language The voice messaging is converted to the first Pinyin information by sound transformation model;By pre-set text transformation model by the target text Sentence is converted to the second Pinyin information;According to preset matching degree computation model to first Pinyin information and second phonetic Matching degree between information is calculated to obtain the speech assessment of the user.
In one embodiment, it is recorded if processor 502 receives the user in execution according to the typing prompt information The voice messaging entered is held when the voice messaging being converted to the step of the first Pinyin information by default voice transformation model The following operation of row: cutting is carried out to voice messaging to obtain being wrapped in voice messaging according to the acoustic model in voice transformation model The multiple phonemes contained;Obtained phoneme is matched with by all sounds according to the phonetic feature dictionary in Text region model Element carries out being converted to the first Pinyin information.
In one embodiment, processor 502 is turned the target text sentence by pre-set text transformation model in execution It when being changed to the step of the second Pinyin information, performs the following operations: being obtained in target text sentence by pre-set text transformation model The Pinyin information of each character;Obtained Pinyin information is combined by pre-set text transformation model to obtain the second spelling Message breath.
In one embodiment, processor 502 is being executed according to preset matching degree computation model to first Pinyin information When matching degree between second Pinyin information is calculated to obtain the step of the speech assessment of the user, execute such as Lower operation: whether equal to the quantity of character phonetic included in first Pinyin information and the second Pinyin information to sentence It is disconnected;It, will be every in the first Pinyin information if the quantity of character phonetic included in the first Pinyin information and the second Pinyin information is equal One character phonetic is matched with respective symbols phonetic in the second Pinyin information to obtain character phonetic coupling number P;According to With the matching degree calculation formula in degree computation model: matching degree S=P/P0The speech assessment of the user is calculated, wherein P0 For the quantity of character phonetic included in the second Pinyin information.
In one embodiment, processor 502 is being executed according to preset matching degree computation model to first Pinyin information After the step of matching degree between second Pinyin information is calculated to obtain the speech assessment of the user, also hold The following operation of row: it is non-to generate to obtain the character phonetic not matched that in first Pinyin information with second Pinyin information Match prompt information.
It will be understood by those skilled in the art that the embodiment of computer equipment shown in Figure 11 is not constituted to computer The restriction of equipment specific composition, in other embodiments, computer equipment may include components more more or fewer than diagram, or Person combines certain components or different component layouts.For example, in some embodiments, computer equipment can only include depositing Reservoir and processor, in such embodiments, the structure and function of memory and processor are consistent with embodiment illustrated in fig. 11, Details are not described herein.
It should be appreciated that in embodiments of the present invention, processor 502 can be central processing unit (Central ProcessingUnit, CPU), which can also be other general processors, digital signal processor (Digital Signal Processor, DSP), specific integrated circuit (Application Specific Integrated Circuit, ASIC), ready-made programmable gate array (Field-Programmable GateArray, FPGA) or other programmable logic devices Part, discrete gate or transistor logic, discrete hardware components etc..Wherein, general processor can be microprocessor or The processor is also possible to any conventional processor etc..
Computer readable storage medium is provided in another embodiment of the invention.The computer readable storage medium can be with For non-volatile computer readable storage medium.The computer-readable recording medium storage has computer program, wherein calculating If machine program performed the steps of when being executed by processor receive matching degree measurement request, from pre-set text statement library with Machine selects a text sentence as target text sentence;Typing prompt information is generated according to the target text sentence to prompt User's typing voice messaging corresponding with the target text sentence;It prompts to believe according to the typing if receiving the user The voice messaging is converted to the first Pinyin information by default voice transformation model by the voice messaging for ceasing institute's typing;Pass through The target text sentence is converted to the second Pinyin information by pre-set text transformation model;According to preset matching degree computation model pair Matching degree between first Pinyin information and second Pinyin information is calculated to obtain the voice of the user and comment Point.
In one embodiment, if described receive the user according to the voice of typing prompt information institute typing letter Breath, by presetting the step of voice messaging is converted to the first Pinyin information by voice transformation model, comprising: turned according to voice Acoustic model in mold changing type carries out cutting to voice messaging to obtain multiple phonemes included in voice messaging;According to text Phonetic feature dictionary in identification model matches obtained phoneme to carry out all phonemes to be converted to the first spelling Message breath.
In one embodiment, described that the target text sentence is converted to by the second phonetic by pre-set text transformation model The step of information, comprising: the Pinyin information of each character in target text sentence is obtained by pre-set text transformation model;Pass through Pre-set text transformation model is combined obtained Pinyin information to obtain the second Pinyin information.
In one embodiment, described that first Pinyin information and described second are spelled according to preset matching degree computation model The step of matching degree between message breath is calculated to obtain the speech assessment of the user, comprising: to first phonetic Whether the quantity of character phonetic included in information and the second Pinyin information is equal to be judged;If the first Pinyin information and second The quantity of character phonetic included in Pinyin information is equal, each character phonetic in the first Pinyin information and the second phonetic are believed Respective symbols phonetic is matched to obtain character phonetic coupling number P in breath;According to the matching degree meter in matching degree computation model Calculate formula: matching degree S=P/P0The speech assessment of the user is calculated, wherein P0For word included in the second Pinyin information Accord with the quantity of phonetic.
In one embodiment, described that first Pinyin information and described second are spelled according to preset matching degree computation model After the step of matching degree between message breath is calculated to obtain the speech assessment of the user, further includes: described in acquisition The character phonetic not matched that in first Pinyin information with second Pinyin information is to generate non-matching prompt information.
It is apparent to those skilled in the art that for convenience of description and succinctly, foregoing description is set The specific work process of standby, device and unit, can refer to corresponding processes in the foregoing method embodiment, and details are not described herein. Those of ordinary skill in the art may be aware that unit described in conjunction with the examples disclosed in the embodiments of the present disclosure and algorithm Step can be realized with electronic hardware, computer software, or a combination of the two, in order to clearly demonstrate hardware and software Interchangeability generally describes each exemplary composition and step according to function in the above description.These functions are studied carefully Unexpectedly the specific application and design constraint depending on technical solution are implemented in hardware or software.Professional technician Each specific application can be used different methods to achieve the described function, but this realization is it is not considered that exceed The scope of the present invention.
In several embodiments provided by the present invention, it should be understood that disclosed unit and method, it can be with It realizes by another way.For example, the apparatus embodiments described above are merely exemplary, for example, the unit It divides, only logical function partition, there may be another division manner in actual implementation, can also will be with the same function Unit set is at a unit, such as multiple units or components can be combined or can be integrated into another system or some Feature can be ignored, or not execute.In addition, shown or discussed mutual coupling, direct-coupling or communication connection can Be through some interfaces, the indirect coupling or communication connection of device or unit, be also possible to electricity, mechanical or other shapes Formula connection.
The unit as illustrated by the separation member may or may not be physically separated, aobvious as unit The component shown may or may not be physical unit, it can and it is in one place, or may be distributed over multiple In network unit.Some or all of unit therein can be selected to realize the embodiment of the present invention according to the actual needs Purpose.
It, can also be in addition, the functional units in various embodiments of the present invention may be integrated into one processing unit It is that each unit physically exists alone, is also possible to two or more units and is integrated in one unit.It is above-mentioned integrated Unit both can take the form of hardware realization, can also realize in the form of software functional units.
If the integrated unit is realized in the form of SFU software functional unit and sells or use as independent product When, it can store in a computer readable storage medium.Based on this understanding, technical solution of the present invention substantially or Person says that all or part of the part that contributes to existing technology or the technical solution can body in the form of software products Reveal and, which is stored in a computer readable storage medium, including some instructions are used so that one Platform computer equipment (can be personal computer, server or the network equipment etc.) executes described in each embodiment of the present invention The all or part of the steps of method.And computer readable storage medium above-mentioned includes: USB flash disk, mobile hard disk, read-only memory The various media that can store program code such as (ROM, Read-Only Memory), magnetic or disk.
The above description is merely a specific embodiment, but scope of protection of the present invention is not limited thereto, any Those familiar with the art in the technical scope disclosed by the present invention, can readily occur in various equivalent modifications or replace It changes, these modifications or substitutions should be covered by the protection scope of the present invention.Therefore, protection scope of the present invention should be with right It is required that protection scope subject to.

Claims (10)

1. a kind of speech assessment method characterized by comprising
If receiving matching degree measurement request, a text sentence is randomly choosed from pre-set text statement library as target text Sentence;
Typing prompt information is generated according to the target text sentence to prompt user's typing opposite with the target text sentence The voice messaging answered;
If receiving the user according to the voice messaging of typing prompt information institute typing, by presetting voice transformation model The voice messaging is converted into the first Pinyin information;
The target text sentence is converted into the second Pinyin information by pre-set text transformation model;
According to preset matching degree computation model to the matching degree between first Pinyin information and second Pinyin information into Row is calculated to obtain the speech assessment of the user.
2. speech assessment method according to claim 1, which is characterized in that described by presetting voice transformation model for language Message breath is converted to the first Pinyin information, comprising:
It is more included in voice messaging to obtain that cutting is carried out to voice messaging according to the acoustic model in voice transformation model A phoneme;
Obtained phoneme is matched to turn all phonemes according to the phonetic feature dictionary in Text region model It gets in return to the first Pinyin information.
3. speech assessment method according to claim 1, which is characterized in that it is described by pre-set text transformation model by institute It states target text sentence and is converted to the second Pinyin information, comprising:
The Pinyin information of each character in target text sentence is obtained by pre-set text transformation model;
Obtained Pinyin information is combined by pre-set text transformation model to obtain the second Pinyin information.
4. speech assessment method according to claim 1, which is characterized in that described according to preset matching degree computation model pair Matching degree between first Pinyin information and second Pinyin information is calculated to obtain the voice of the user and comment Point, comprising:
It is whether equal to the quantity of character phonetic included in first Pinyin information and the second Pinyin information to judge;
It, will be every in the first Pinyin information if the quantity of character phonetic included in the first Pinyin information and the second Pinyin information is equal One character phonetic is matched with respective symbols phonetic in the second Pinyin information to obtain character phonetic coupling number P;
According to the matching degree calculation formula in matching degree computation model: matching degree S=P/P0The voice that the user is calculated is commented Point, wherein P0For the quantity of character phonetic included in the second Pinyin information.
5. speech assessment method according to claim 1, which is characterized in that described according to preset matching degree computation model pair Matching degree between first Pinyin information and second Pinyin information is calculated to obtain the voice of the user and comment / after, further includes:
The character phonetic not matched that in first Pinyin information with second Pinyin information is obtained to generate non-matching mention Show information.
6. a kind of speech assessment device characterized by comprising
Target text sentence acquiring unit, if being selected at random from pre-set text statement library for receiving matching degree measurement request A text sentence is selected as target text sentence;
Prompt information generation unit, for according to the target text sentence generate typing prompt information with prompt user's typing with The corresponding voice messaging of the target text sentence;
First Pinyin information acquiring unit, if for receiving the user according to the voice of typing prompt information institute typing The voice messaging is converted to the first Pinyin information by default voice transformation model by information;
Second Pinyin information acquiring unit, for the target text sentence to be converted to second by pre-set text transformation model Pinyin information;
Speech assessment acquiring unit, for being spelled according to preset matching degree computation model to first Pinyin information and described second Matching degree between message breath is calculated to obtain the speech assessment of the user.
7. speech assessment device according to claim 6, which is characterized in that the first Pinyin information acquiring unit, packet It includes:
Phoneme acquiring unit, for carrying out cutting to voice messaging according to the acoustic model in voice transformation model to obtain voice Multiple phonemes included in information;
Phoneme matching unit, for according to the phonetic feature dictionary in Text region model to obtained phoneme matched with All phonemes are carried out to be converted to the first Pinyin information.
8. speech assessment device according to claim 6, which is characterized in that the second Pinyin information acquiring unit, packet It includes:
Character Pinyin information acquiring unit, for obtaining each character in target text sentence by pre-set text transformation model Pinyin information;
Pinyin information assembled unit, for being combined obtained Pinyin information to obtain by pre-set text transformation model Second Pinyin information.
9. a kind of computer equipment, including memory, processor and it is stored on the memory and can be on the processor The computer program of operation, which is characterized in that the processor realizes such as claim 1 to 5 when executing the computer program Any one of described in speech assessment method.
10. a kind of computer readable storage medium, which is characterized in that the computer-readable recording medium storage has computer journey Sequence, the computer program make the processor execute such as language described in any one of claim 1 to 5 when being executed by a processor Sound methods of marking.
CN201811493385.4A 2018-12-07 2018-12-07 Speech assessment method, apparatus, computer equipment and storage medium Pending CN109461459A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811493385.4A CN109461459A (en) 2018-12-07 2018-12-07 Speech assessment method, apparatus, computer equipment and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811493385.4A CN109461459A (en) 2018-12-07 2018-12-07 Speech assessment method, apparatus, computer equipment and storage medium

Publications (1)

Publication Number Publication Date
CN109461459A true CN109461459A (en) 2019-03-12

Family

ID=65612699

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811493385.4A Pending CN109461459A (en) 2018-12-07 2018-12-07 Speech assessment method, apparatus, computer equipment and storage medium

Country Status (1)

Country Link
CN (1) CN109461459A (en)

Cited By (7)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110085210A (en) * 2019-03-15 2019-08-02 平安科技(深圳)有限公司 Interactive information test method, device, computer equipment and storage medium
CN110136721A (en) * 2019-04-09 2019-08-16 北京大米科技有限公司 A kind of scoring generation method, device, storage medium and electronic equipment
CN111027794A (en) * 2019-03-29 2020-04-17 广东小天才科技有限公司 Dictation operation correcting method and learning equipment
CN111179937A (en) * 2019-12-24 2020-05-19 上海眼控科技股份有限公司 Method, apparatus and computer-readable storage medium for text processing
CN112291281A (en) * 2019-07-09 2021-01-29 钉钉控股(开曼)有限公司 Voice broadcast and voice broadcast content setting method and device
CN112634858A (en) * 2020-12-16 2021-04-09 平安科技(深圳)有限公司 Speech synthesis method, speech synthesis device, computer equipment and storage medium
CN113793611A (en) * 2021-08-27 2021-12-14 上海浦东发展银行股份有限公司 Scoring method, scoring device, computer equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1455386A (en) * 2002-11-01 2003-11-12 中国科学院声学研究所 Imbedded voice synthesis method and system
CN103559880A (en) * 2013-11-08 2014-02-05 百度在线网络技术(北京)有限公司 Voice input system and voice input method
CN107204184A (en) * 2017-05-10 2017-09-26 平安科技(深圳)有限公司 Audio recognition method and system
CN107659847A (en) * 2016-09-22 2018-02-02 腾讯科技(北京)有限公司 Voice interface method and apparatus
CN107731228A (en) * 2017-09-20 2018-02-23 百度在线网络技术(北京)有限公司 The text conversion method and device of English voice messaging
CN108766437A (en) * 2018-05-31 2018-11-06 平安科技(深圳)有限公司 Audio recognition method, device, computer equipment and storage medium

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN1455386A (en) * 2002-11-01 2003-11-12 中国科学院声学研究所 Imbedded voice synthesis method and system
CN103559880A (en) * 2013-11-08 2014-02-05 百度在线网络技术(北京)有限公司 Voice input system and voice input method
CN107659847A (en) * 2016-09-22 2018-02-02 腾讯科技(北京)有限公司 Voice interface method and apparatus
CN107204184A (en) * 2017-05-10 2017-09-26 平安科技(深圳)有限公司 Audio recognition method and system
CN107731228A (en) * 2017-09-20 2018-02-23 百度在线网络技术(北京)有限公司 The text conversion method and device of English voice messaging
CN108766437A (en) * 2018-05-31 2018-11-06 平安科技(深圳)有限公司 Audio recognition method, device, computer equipment and storage medium

Cited By (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110085210A (en) * 2019-03-15 2019-08-02 平安科技(深圳)有限公司 Interactive information test method, device, computer equipment and storage medium
CN110085210B (en) * 2019-03-15 2023-10-13 平安科技(深圳)有限公司 Interactive information testing method and device, computer equipment and storage medium
CN111027794A (en) * 2019-03-29 2020-04-17 广东小天才科技有限公司 Dictation operation correcting method and learning equipment
CN111027794B (en) * 2019-03-29 2023-09-26 广东小天才科技有限公司 Correction method and learning equipment for dictation operation
CN110136721A (en) * 2019-04-09 2019-08-16 北京大米科技有限公司 A kind of scoring generation method, device, storage medium and electronic equipment
CN112291281A (en) * 2019-07-09 2021-01-29 钉钉控股(开曼)有限公司 Voice broadcast and voice broadcast content setting method and device
CN112291281B (en) * 2019-07-09 2023-11-03 钉钉控股(开曼)有限公司 Voice broadcasting and voice broadcasting content setting method and device
CN111179937A (en) * 2019-12-24 2020-05-19 上海眼控科技股份有限公司 Method, apparatus and computer-readable storage medium for text processing
CN112634858A (en) * 2020-12-16 2021-04-09 平安科技(深圳)有限公司 Speech synthesis method, speech synthesis device, computer equipment and storage medium
CN112634858B (en) * 2020-12-16 2024-01-23 平安科技(深圳)有限公司 Speech synthesis method, device, computer equipment and storage medium
CN113793611A (en) * 2021-08-27 2021-12-14 上海浦东发展银行股份有限公司 Scoring method, scoring device, computer equipment and storage medium

Similar Documents

Publication Publication Date Title
CN109461459A (en) Speech assessment method, apparatus, computer equipment and storage medium
CN103714048B (en) Method and system for correcting text
EP2958105B1 (en) Method and apparatus for speech synthesis based on large corpus
EP3504709B1 (en) Determining phonetic relationships
CN109817201B (en) Language learning method and device, electronic equipment and readable storage medium
CN109637521A (en) A kind of lip reading recognition methods and device based on deep learning
CN111402862B (en) Speech recognition method, device, storage medium and equipment
Klatt et al. On the automatic recognition of continuous speech: Implications from a spectrogram-reading experiment
CN106935239A (en) The construction method and device of a kind of pronunciation dictionary
JP2005321817A (en) Method and apparatus for obtaining combining information from speech signals for adaptive interaction in teaching and testing
US11410642B2 (en) Method and system using phoneme embedding
CN108877782A (en) Audio recognition method and device
CN109817244B (en) Spoken language evaluation method, device, equipment and storage medium
CN111833853A (en) Voice processing method and device, electronic equipment and computer readable storage medium
CN109166569B (en) Detection method and device for phoneme mislabeling
CN103050115A (en) Recognizing device, recognizing method, generating device, and generating method
CN104639742B (en) The spoken method and device of mobile terminal CAL
CN110335608A (en) Voice print verification method, apparatus, equipment and storage medium
CN109658931A (en) Voice interactive method, device, computer equipment and storage medium
CN110148413B (en) Voice evaluation method and related device
CN109582775B (en) Information input method, device, computer equipment and storage medium
CN110503956A (en) Audio recognition method, device, medium and electronic equipment
JP2004053652A (en) Pronunciation judging system, server for managing system and program therefor
JP2006208644A (en) Server system and method for measuring linguistic speaking ability
CN109697975B (en) Voice evaluation method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination