CN108735220A - A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method - Google Patents

A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method Download PDF

Info

Publication number
CN108735220A
CN108735220A CN201810319502.9A CN201810319502A CN108735220A CN 108735220 A CN108735220 A CN 108735220A CN 201810319502 A CN201810319502 A CN 201810319502A CN 108735220 A CN108735220 A CN 108735220A
Authority
CN
China
Prior art keywords
voice
document
text
module
error
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810319502.9A
Other languages
Chinese (zh)
Inventor
祝梦影
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Sichuan Feixun Information Technology Co Ltd
Original Assignee
Sichuan Feixun Information Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Sichuan Feixun Information Technology Co Ltd filed Critical Sichuan Feixun Information Technology Co Ltd
Priority to CN201810319502.9A priority Critical patent/CN108735220A/en
Publication of CN108735220A publication Critical patent/CN108735220A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F40/00Handling natural language data
    • G06F40/20Natural language analysis
    • G06F40/232Orthographic correction, e.g. spell checking or vowelisation
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04MTELEPHONIC COMMUNICATION
    • H04M1/00Substation equipment, e.g. for use by subscribers
    • H04M1/72Mobile telephones; Cordless telephones, i.e. devices for establishing wireless links to base stations without route selection
    • H04M1/724User interfaces specially adapted for cordless or mobile telephones
    • H04M1/72403User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality
    • H04M1/7243User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages
    • H04M1/72433User interfaces specially adapted for cordless or mobile telephones with means for local support of applications that increase the functionality with interactive means for internal management of messages for voice messaging, e.g. dictaphones

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • Multimedia (AREA)
  • Acoustics & Sound (AREA)
  • Artificial Intelligence (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Business, Economics & Management (AREA)
  • General Business, Economics & Management (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Signal Processing (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

The present invention provides a kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction methods.In the present invention, the former voice document of learner is obtained first, and carrying out text file conversion to the original voice document using speech recognition technology obtains original text this document;Wrong point list is obtained to original text this document error correction later;Then feedback error point list, and according to wrong point list, each erroneous point in original text this document is replaced with into correct machine language text to be formed and corrects version text file;Version voice document is corrected followed by phonetic synthesis correction version text file acquisition and plays to learner.By error correction, feedback and replacement, when learner hears that the content that earphone plays is the sound of oneself, illustrate this part spoken language inerrancy, when hearing machine language, illustrates wrong herein, need to be corrected with reference to correct machine language, add to practice as needed.

Description

A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method
Technical field
The present invention relates to technical field of intelligent interaction more particularly to a kind of language learning intelligent earphone, intelligent interaction systems System and man-machine interaction method.
Background technology
Wireless Bluetooth headsets are developed to from wired earphone, from function earphone is listened to music merely to can be with the intelligence of human-computer interaction Earphone, the type and function of earphone are more and more abundant.
In addition, as what is exchanged between various countries increases, people are also more and more for the demand of foreign language language.Existing language Learn in product, can only realize the function of listening and read, such as some language learning cell phone applications, it can not be to learner It practises situation and provides specific feedback, this causes learner that can not look into scarce mending-leakage, in place of specific aim is covered the shortage.
Invention content
The goal of the invention of the present invention is to provide a kind of language learning intelligent earphone, intelligent interactive system and human-computer interaction side Method can designate that the mistake place in learner's spoken language and feed back to learner, for learner's repetition learning to reinforce practicing.
To achieve the above object, the present invention provides a kind of language learning intelligent earphone, including:Recording module, voice are known Other module, correction module, feedback module, text processing module, voice synthetic module and playing module;
Wherein, the recording module is used for the study phonetic acquisition original voice document of typing learner;
The sound identification module is used to carry out text file conversion to the former voice document using speech recognition technology, Obtain original text this document;
The correction module is used to obtain wrong point list to original text this document error correction using artificial intelligence technology;
The feedback module is for feeding back the wrong point list;
The wrong point list of the text processing module based on feedback, by each erroneous point in described original text this document Correct machine language text is replaced with, is formed and corrects version text file;
The voice synthetic module is used to carry out voice document to the correction version text file using speech synthesis technique Conversion obtains and corrects version voice document;
The playing module plays the correction version voice document by earphone.
Optionally, the correction module uses artificial intelligence technology to original text this document error correction.
Optionally, the wrong point list of the text processing module based on feedback, also forms revised text file;Institute State the type of error text file and corresponding correct machine language text file that revised text file is each erroneous point;It is described Voice synthetic module also carries out voice document conversion using speech synthesis technique to the revised text file, obtains each erroneous point Type of error voice document and corresponding machine talk file;The playing module also plays the type of error by earphone Voice document and machine talk file.
Optionally, the revised text file is interspersed in the correction version text file by the text processing module, Or it is put in before or after the correction version text file.
Optionally, further include voice storage module, version voice is corrected for storing.
Optionally, further include voice storage module, type of error voice for storing each erroneous point and corresponding correct Machine language voice.
Optionally, the intelligent earphone is translation earphone, the sound identification module, voice synthetic module and broadcasting mould Block is served as by sound identification module, voice synthetic module and the playing module of the translation earphone.
The present invention also provides a kind of language learning intelligent interactive systems, including:
Above-mentioned intelligent earphone and mobile terminal;
Wherein, the mobile terminal and the intelligent earphone interactive communication, the APP on the mobile terminal, which is synchronized, obtains institute State former voice document and the correction version voice document.
The present invention further provides a kind of man-machine interaction methods of intelligent earphone, including:
The study phonetic acquisition original voice document of typing learner;
Text file conversion is carried out to the former voice document using speech recognition technology, obtains original text this document;
To original text this document error correction, wrong point list is obtained;
It feeds back the wrong point list and each erroneous point in described original text this document is replaced with into correct machine language Text forms and corrects version text file;
Voice document conversion is carried out to the correction version text file using speech synthesis technique, obtains and corrects version voice text Part;
The correction version voice document is played by earphone.
Optionally, using artificial intelligence technology to original text this document error correction.
Optionally, the wrong point list is fed back, revised text file is also formed;The revised text file is each mistake The type of error text file and corresponding correct machine language text file of point;It is also repaiied to described using speech synthesis technique It orders text file and carries out voice document conversion, obtain the type of error voice document and corresponding machine talk text of each erroneous point Part;The type of error voice document and machine talk file are played by earphone.
Optionally, when broadcasting, the type of error voice document and machine talk file are interspersed in the correction version voice It is interior, or be put in before or after the correction version voice.
Optionally, the correction version voice is also stored, to call study repeatedly.
Optionally, the type of error voice document and machine talk file are also stored, to call study repeatedly.
Optionally, the different type of error text files of each erroneous point are distinguished using different code.
Optionally, the type of error of each erroneous point includes:Word or byte pronunciation mistake, tone or stress mistake, With pause or not familiar mistake.
Optionally, further include the APP being synchronized to former voice document with the correction version voice document on mobile phone.
Optionally, further include by former voice document, the type of error language for correcting version voice document and each erroneous point APP on sound file and corresponding machine talk file synchronization to mobile phone.
Compared with prior art, the beneficial effects of the present invention are:
1) in the present invention, the former voice document of learner is obtained first, using speech recognition technology to the original voice document It carries out text file conversion and obtains original text this document;Wrong point list is obtained to original text this document error correction later;Then it feeds back Mistake point list, and according to wrong point list, each erroneous point in original text this document is replaced with into correct machine language text Version text file is corrected to be formed;Version voice document is corrected followed by phonetic synthesis correction version text file acquisition and is played to Learner.As can be seen that by error correction, feedback and replacement, when learner hears that the content that earphone plays is the sound of oneself When, illustrate this part spoken language inerrancy, when hearing machine language, illustrate it is wrong herein, need to be with reference to correct machine language It corrects, adds to practice as needed.
2) in alternative, in error correction step, a) the representative normative text number in the data packet that can be prestored with earphone According to being compared, big data error correction of the artificial intelligence technology by internet can also b) be used.
3) in alternative, version voice document is corrected in addition to playing, can also be stored in earphone, be adjusted repeatedly for learner It takes hard of hearing.
Still optionally further, the correction version voice document can be stored, can also store and correct version text file, benefit exists In:Reduce the memory space occupied, or in the case where memory space is certain, stores several sections more.
Still optionally further, memory block can be opened in earphone, can also be opened in the shifting that LAN can be formed with the earphone In dynamic terminal.
4) in alternative, learner can not only hear complete correction version voice, can also hear the mistake of each erroneous point The voice of type and corresponding correct machine language.Above-mentioned purpose by feedback error point list simultaneously, also formed revision text This document;Revised text file is the type of error text file of each erroneous point and corresponding correct machine language text text Part.Later revised text file with correct version text file respectively respectively or together, played after phonetic synthesis.About each mistake The type of error and corresponding correct machine language voice of point can be interspersed in and correct in version voice, or be put in when playing Before or after the correction version voice.
Still optionally further, in addition to playing, it is also sent to the APP of mobile terminal, APP is opened whenever and wherever possible for learner It practises.Above-mentioned transmission opportunity can be:A) learner is actively in the APP of mobile terminal when synchronous study result;Or b) learner When opening the APP of mobile terminal, earphone is actively sent;Or c) mobile terminal is once in earphone under same LAN, earphone Actively send.
5) in alternative, the type of error of each erroneous point includes:Word or byte pronunciation mistake, tone or stress are wrong Mistake and pause or not familiar mistake.Each type of error is refined, the emphasis that learner can be prompted to correct targetedly corrects mistake.
Description of the drawings
Fig. 1 is the module map of the language learning intelligent earphone in one embodiment of the invention;
Fig. 2 is the flow chart of the man-machine interaction method of the intelligent earphone in Fig. 1:
Fig. 3 is the flow chart of the man-machine interaction method of the intelligent earphone in another embodiment of the present invention.
Specific implementation mode
To make the above purposes, features and advantages of the invention more obvious and understandable, below in conjunction with the accompanying drawings to the present invention Specific embodiment be described in detail.
Fig. 1 is the module map of the language learning intelligent earphone in one embodiment of the invention.
Shown in referring to Fig.1, which includes:Recording module 11, sound identification module 12, error correction Module 13, feedback module 14, text processing module 15, voice synthetic module 16 and playing module 17.
The function of each module is introduced individually below.
Recording module 11 is used for the study phonetic acquisition original voice document of typing learner.
The study voice of learner can be one or a few words, one or several phrases, even one or several words.It learns The content of the study voice of habit person can be based at that time, local, mood is extemporaneously played instantly, can also be with reference on intelligent terminal The learning Content presented in APP application programs.
The former voice document of typing can be stored in voice storage module.
Sound identification module 12 is used to carry out text file conversion to former voice document using speech recognition technology, obtains former Text file.
Speech recognition technology can be existing speech recognition technology, such as using at present using more depth nerve net Network (Deep Neural Network, DNN), recurrent neural network (Recurrent Neural Network, RNN).It is converted into First text file occupies memory space small, is second conducive to subsequently compare error correction in big data.
Correction module 13 is used to, to original text this document error correction, obtain wrong point list.
Error correction can be there are two types of approach.First, the study voice for being directed to learner is originated from and is answered with reference to the APP on intelligent terminal The case where with the learning Content presented in program, can be compared with the representative normative text data in pre stored data. Second, for the study voice of learner content be based at that time, it is local, mood is extemporaneously played instantly the case where, using artificial Intellectual technology error correction.The use of artificial intelligence technology error correction refer to capturing to compare in big data on the internet.Earphone can lead to It crosses bluetooth to connect with mobile terminal, big data is obtained using mobile terminal as hot spot.Earphone can also with mobile terminal and Router forms LAN, and big data is obtained using the communication of router.
Mistake point list includes the correspondence of the position and wrong content of each mistake.
Feedback module 14 is for feeding back the mistake point list.
Feedback module 14 forms a kind of feedback mechanism so that learner has practiced spoken language not only through saying, and knows mistake Accidentally place, targetedly to repeat.
Wrong point list of the text processing module 15 based on feedback, each erroneous point in original text this document is replaced with correctly Machine language text, formed correct version text file.
About correct machine language text when replacing, acquisition modes can have:A) correction module 13 not only obtains The position of erroneous point and erroneous point content can also compare according to data packet or big data and obtain correct machine language text;It should Correct machine language text can be together listed in position, wrong content in wrong point list;Or b) text processing module 15 The machine language for calling earphone to prestore.
Voice synthetic module 16 is used to carry out voice document conversion to correcting version text file using speech synthesis technique, obtains Version voice document must be corrected.Speech synthesis technique is referred to existing speech synthesis technique, such as uses WaveNet softwares.
Playing module 17 plays the correction version voice document by earphone.
Version voice document is corrected in addition to playing, can also be stored in voice storage module, be called repeatedly for learner.
Learner can be turned over the former voice document of calling up and down by the button on earphone 1 and/or correct version voice document, also It can be called by the APP of intelligent terminal.
When learner hears that the content that earphone 1 plays is oneself sound, illustrate this part spoken language inerrancy, when hearing When machine language, illustrate wrong herein, need to be corrected with reference to correct machine language, can also add to practice as needed.
Intelligent earphone 1 in the present embodiment can be translation earphone.In other words, which is based on existing translation Earphone carries out functions expanding, in this way:Sound identification module 12, voice synthetic module 16 and playing module 17 can be by translation ears Sound identification module, voice synthetic module and the playing module of machine serve as.
One embodiment of the invention also provides a kind of language learning intelligent interactive system, including:
Above-mentioned intelligent earphone 1 and mobile terminal;
Wherein, mobile terminal and 1 interactive communication of intelligent earphone, the APP on mobile terminal synchronize obtain former voice document with Correct version voice document.
Intelligent earphone 1 can pass through bluetooth and mobile terminal interactive communication.Intelligent earphone 1 can also with mobile terminal and Router forms LAN, the interactive communication in LAN.
Fig. 2 is the flow chart of the man-machine interaction method of the intelligent earphone in Fig. 1.
Each step is introduced individually below.
With reference to shown in Fig. 2, step S1, the study phonetic acquisition original voice document of typing learner is first carried out.
The study voice of learner can be one or a few words, one or several phrases, even one or several words.It learns The content of the study voice of habit person can be based at that time, local, mood is extemporaneously played instantly, can also be with reference on intelligent terminal The learning Content presented in APP application programs.
The former voice document of typing can be stored in voice storage module.
Then step S2 is executed, text file conversion is carried out to the original voice document using speech recognition technology, is obtained former Text file.
Speech recognition technology can be existing speech recognition technology.
Being converted into text file, first to occupy memory space small, is second conducive to subsequently compare error correction in big data.
Followed by step S3 is executed, to original text this document error correction, wrong point list is obtained.
Error correction can be there are two types of method.First, the study voice for being directed to learner is originated from and is answered with reference to the APP on intelligent terminal The case where with the learning Content presented in program, can be compared with the representative normative text data in pre stored data. Second, for the study voice of learner content be based at that time, it is local, mood is extemporaneously played instantly the case where, using artificial Intellectual technology error correction.The use of artificial intelligence technology error correction refer to capturing to compare in big data on the internet.Earphone can lead to It crosses bluetooth to connect with mobile terminal, big data is obtained using mobile terminal as hot spot.Earphone can also with mobile terminal and Router forms LAN, and big data is obtained using the communication of router.
Mistake point list includes the correspondence of the position and wrong content of each mistake.
Step S4 is executed later, and feedback error point list simultaneously replaces with each erroneous point in the original text this document correctly Machine language text forms and corrects version text file.
This feedback step forms a kind of feedback mechanism so that learner has practiced spoken language not only through saying, and knows mistake Accidentally place, targetedly to repeat.
About correct machine language text when replacing, acquisition modes can have:A) error correction step S3 is not only obtained The position of erroneous point and erroneous point content can also compare according to data packet or big data and obtain correct machine language text;It should Correct machine language text can be together listed in position, wrong content in wrong point list;Or b) earphone is called to prestore Machine language.
Step S5 is executed, voice document conversion is carried out to correcting version text file using speech synthesis technique, is corrected Version voice document.
Speech synthesis technique is referred to existing speech synthesis technique.
Then step S6 is executed, which is played by earphone.
Play-back technology is referred to existing play-back technology.
Intelligent earphone in the present embodiment can be translation earphone.In other words, which is based on existing translation ear Machine carries out functions expanding, in this way:Speech recognition steps S2, phonetic synthesis step S5 and broadcasting step S6 can be by translation earphones Sound identification module, voice synthetic module and playing module complete.
Another embodiment of the present invention also provides a kind of Foreigh-language oral-speech practice intelligent earphone.Compared with previous embodiment, this Intelligent earphone in embodiment difference lies in:1) wrong point list of the text processing module 15 based on feedback also forms revision text This document;The revised text file is the type of error text file of each erroneous point and corresponding correct machine language text text Part;2) voice synthetic module 16 also carries out voice document conversion using speech synthesis technique to the revised text file, obtains The type of error voice document of each erroneous point and corresponding machine talk file;3) playing module 17 passes through earphone also playback error Type voice file and machine talk file.
Revised text file can be interspersed in and correct in version text file by text processing module 15 as needed, or put Before or after correcting version text file.
The type of error of each erroneous point may include:Word or byte pronunciation mistake, tone or stress mistake and pause or Not familiar mistake.In text file, different type mistake can be distinguished by different code, can also be distinguished by different storage addresses. Word or byte pronunciation mistake, tone or stress mistake can be placed on not familiar mistake before or after correction version text file.For Pause mistake can be interspersed in and correct in version text file, so that follow-up earphone plays " should not have pause herein " to prompt to use Family.The certain value of full section, such as 80% or more can be accounted for mistake of statistics for the criterion of not familiar mistake, plays " this section life Dredge " to prompt learner.
Fig. 3 is the flow chart of the man-machine interaction method of the intelligent earphone in another embodiment of the present invention.It, can be with reference to Fig. 3 Find out, it is roughly the same with the flow in Fig. 2, difference lies in:
1) in step S4 ', feedback error point list also forms revised text file;Revised text file is each erroneous point Type of error text file and corresponding correct machine language text file;
2) in step S5 ', voice document conversion also is carried out to revised text file using speech synthesis technique, obtains each mistake Overdue type of error voice document and corresponding machine talk file;
3) in step S6 ', pass through earphone also playback error type voice file and machine talk file.
In revised text file, the type of error of each erroneous point may include:Word or byte pronunciation mistake, tone or again Sound mistake and pause or not familiar mistake.In text file, different type mistake can be distinguished by different code, can also be by not It is distinguished with storage address.Word or byte pronunciation mistake, can be placed on not familiar mistake and correct version text tone or stress mistake Before or after file.For pause mistake, it can be interspersed in and correct in version text file, so that the broadcasting of follow-up earphone " is not answered herein Have pause " to prompt user.Can account for the certain value of full section with mistake of statistics for the criterion of not familiar mistake, for example, 80% with On, " this section is not familiar " is played to prompt learner.
The type of error voice document of each erroneous point and corresponding machine talk file are played, it can also be by each erroneous point Type of error voice document and corresponding machine talk file preserve, so that learner calls.Learner can be by earphone Button up and down turn over the type of error voice document for calling each erroneous point and corresponding machine talk file, intelligence can also be passed through The APP of terminal is called.
Intelligent earphone in the present embodiment can also be translation earphone.New phonetic synthesis step S5 ' and broadcasting step S6 ' can be completed by the voice synthetic module and playing module of translation earphone.
Although present disclosure is as above, present invention is not limited to this.Any those skilled in the art are not departing from this It in the spirit and scope of invention, can make various changes or modifications, therefore protection scope of the present invention should be with claim institute Subject to the range of restriction.

Claims (18)

1. a kind of language learning intelligent earphone, which is characterized in that including:Recording module, sound identification module, correction module, Feedback module, text processing module, voice synthetic module and playing module;
Wherein, the recording module is used for the study phonetic acquisition original voice document of typing learner;
The sound identification module is used to carry out text file conversion to the former voice document using speech recognition technology, obtains Original text this document;
The correction module is used to, to original text this document error correction, obtain wrong point list;
The feedback module is for feeding back the wrong point list;
The wrong point list of the text processing module based on feedback, each erroneous point in described original text this document is replaced For correct machine language text, is formed and correct version text file;
The voice synthetic module is used to carry out voice document conversion to the correction version text file using speech synthesis technique, It obtains and corrects version voice document;
The playing module plays the correction version voice document by earphone.
2. language learning intelligent earphone according to claim 1, which is characterized in that the correction module uses artificial intelligence Energy technology is to original text this document error correction.
3. language learning intelligent earphone according to claim 1, which is characterized in that the text processing module is based on anti- The wrong point list of feedback, also forms revised text file;The revised text file is the type of error text of each erroneous point This document and corresponding correct machine language text file;The voice synthetic module is using speech synthesis technique also to described Revised text file carries out voice document conversion, obtains the type of error voice document and corresponding machine talk text of each erroneous point Part;The playing module also plays the type of error voice document and machine talk file by earphone.
4. language learning intelligent earphone according to claim 3, which is characterized in that the text processing module will be described Revised text file is interspersed in the correction version text file, or is put in before or after the correction version text file.
5. language learning intelligent earphone according to claim 1, which is characterized in that further include voice storage module, use Version voice is corrected in storage.
6. language learning intelligent earphone according to claim 3, which is characterized in that further include voice storage module, use In the type of error voice and corresponding correct machine language voice that store each erroneous point.
7. language learning intelligent earphone according to any one of claims 1 to 6, which is characterized in that the intelligent earphone To translate earphone, the sound identification module, voice synthetic module and playing module are by the speech recognition for translating earphone Module, voice synthetic module and playing module serve as.
8. a kind of language learning intelligent interactive system, which is characterized in that including:
Intelligent earphone described in any one of claim 1 to 7 and mobile terminal;
Wherein, the mobile terminal and the intelligent earphone interactive communication, the APP on the mobile terminal, which is synchronized, obtains the original Voice document and the correction version voice document.
9. a kind of man-machine interaction method of intelligent earphone, which is characterized in that including:
The study phonetic acquisition original voice document of typing learner;
Text file conversion is carried out to the former voice document using speech recognition technology, obtains original text this document;
To original text this document error correction, wrong point list is obtained;
It feeds back the wrong point list and each erroneous point in described original text this document is replaced with into correct machine language text, It is formed and corrects version text file;
Voice document conversion is carried out to the correction version text file using speech synthesis technique, obtains and corrects version voice document;
The correction version voice document is played by earphone.
10. man-machine interaction method according to claim 9, which is characterized in that using artificial intelligence technology to the original text This document error correction.
11. man-machine interaction method according to claim 9, which is characterized in that the feedback wrong point list is also formed and repaiied Order text file;The revised text file is the type of error text file of each erroneous point and corresponding correct machine language Text file;Voice document conversion also is carried out to the revised text file using speech synthesis technique, obtains each erroneous point Type of error voice document and corresponding machine talk file;The type of error voice document and machine language are played by earphone Sound file.
12. man-machine interaction method according to claim 11, which is characterized in that when broadcasting, the type of error voice text Part and machine talk file are interspersed in the correction version voice, or are put in before or after the correction version voice.
13. man-machine interaction method according to claim 9, which is characterized in that the correction version voice is also stored, with Study is called repeatedly.
14. man-machine interaction method according to claim 11, which is characterized in that the type of error voice document and machine Voice document is also stored, to call study repeatedly.
15. man-machine interaction method according to claim 11, which is characterized in that the different type of error texts of each erroneous point File is distinguished using different code.
16. the man-machine interaction method according to claim 11 or 15, which is characterized in that the type of error of each erroneous point Including:Word or byte pronunciation mistake, tone or stress mistake and pause or not familiar mistake.
17. man-machine interaction method according to claim 9, which is characterized in that further include entangling former voice document with described Legal voice document is synchronized to the APP on mobile phone.
18. man-machine interaction method according to claim 11, which is characterized in that further include by former voice document, described entangle On the type of error voice document and corresponding machine talk file synchronization to mobile phone of legal voice document and each erroneous point APP。
CN201810319502.9A 2018-04-11 2018-04-11 A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method Pending CN108735220A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810319502.9A CN108735220A (en) 2018-04-11 2018-04-11 A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810319502.9A CN108735220A (en) 2018-04-11 2018-04-11 A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method

Publications (1)

Publication Number Publication Date
CN108735220A true CN108735220A (en) 2018-11-02

Family

ID=63940747

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810319502.9A Pending CN108735220A (en) 2018-04-11 2018-04-11 A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method

Country Status (1)

Country Link
CN (1) CN108735220A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109462790A (en) * 2018-11-14 2019-03-12 王美金 A kind of artificial intelligence wears mill ear financial payment translation earphone cloud system and method
CN114974221A (en) * 2022-04-29 2022-08-30 中移互联网有限公司 Speech recognition model training method and device and computer readable storage medium

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101739870A (en) * 2009-12-03 2010-06-16 深圳先进技术研究院 Interactive language learning system and method
CN103000052A (en) * 2011-09-16 2013-03-27 上海先先信息科技有限公司 Man-machine interactive spoken dialogue system and realizing method thereof
CN203761458U (en) * 2014-02-24 2014-08-06 汉诚信息技术(上海)有限公司 Interactive type automatic voice response system
CN104123858A (en) * 2014-07-30 2014-10-29 广东小天才科技有限公司 Method and device for error detection and correction during back-reading lesson text
CN105025178A (en) * 2014-04-18 2015-11-04 北京艾沃信通讯技术有限公司 Method and system for converting interactive voice response into interactive text response
CN105304080A (en) * 2015-09-22 2016-02-03 科大讯飞股份有限公司 Speech synthesis device and speech synthesis method
CN106528616A (en) * 2016-09-30 2017-03-22 厦门快商通科技股份有限公司 Language error correcting method and system for use in human-computer interaction process
CN106847288A (en) * 2017-02-17 2017-06-13 上海创米科技有限公司 The error correction method and device of speech recognition text
CN107133222A (en) * 2017-04-17 2017-09-05 中译语通科技(北京)有限公司 A kind of real-time language conversion equipment and conversion method based on heterogeneous framework
CN107293284A (en) * 2017-07-27 2017-10-24 上海传英信息技术有限公司 A kind of phoneme synthesizing method and speech synthesis system based on intelligent terminal
CN107622054A (en) * 2017-09-26 2018-01-23 科大讯飞股份有限公司 The error correction method and device of text data
CN107644637A (en) * 2017-03-13 2018-01-30 平安科技(深圳)有限公司 Phoneme synthesizing method and device
CN107798931A (en) * 2017-12-05 2018-03-13 上海电机学院 A kind of intelligent children education learning system and method

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101739870A (en) * 2009-12-03 2010-06-16 深圳先进技术研究院 Interactive language learning system and method
CN103000052A (en) * 2011-09-16 2013-03-27 上海先先信息科技有限公司 Man-machine interactive spoken dialogue system and realizing method thereof
CN203761458U (en) * 2014-02-24 2014-08-06 汉诚信息技术(上海)有限公司 Interactive type automatic voice response system
CN105025178A (en) * 2014-04-18 2015-11-04 北京艾沃信通讯技术有限公司 Method and system for converting interactive voice response into interactive text response
CN104123858A (en) * 2014-07-30 2014-10-29 广东小天才科技有限公司 Method and device for error detection and correction during back-reading lesson text
CN105304080A (en) * 2015-09-22 2016-02-03 科大讯飞股份有限公司 Speech synthesis device and speech synthesis method
CN106528616A (en) * 2016-09-30 2017-03-22 厦门快商通科技股份有限公司 Language error correcting method and system for use in human-computer interaction process
CN106847288A (en) * 2017-02-17 2017-06-13 上海创米科技有限公司 The error correction method and device of speech recognition text
CN107644637A (en) * 2017-03-13 2018-01-30 平安科技(深圳)有限公司 Phoneme synthesizing method and device
CN107133222A (en) * 2017-04-17 2017-09-05 中译语通科技(北京)有限公司 A kind of real-time language conversion equipment and conversion method based on heterogeneous framework
CN107293284A (en) * 2017-07-27 2017-10-24 上海传英信息技术有限公司 A kind of phoneme synthesizing method and speech synthesis system based on intelligent terminal
CN107622054A (en) * 2017-09-26 2018-01-23 科大讯飞股份有限公司 The error correction method and device of text data
CN107798931A (en) * 2017-12-05 2018-03-13 上海电机学院 A kind of intelligent children education learning system and method

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN109462790A (en) * 2018-11-14 2019-03-12 王美金 A kind of artificial intelligence wears mill ear financial payment translation earphone cloud system and method
CN114974221A (en) * 2022-04-29 2022-08-30 中移互联网有限公司 Speech recognition model training method and device and computer readable storage medium
CN114974221B (en) * 2022-04-29 2024-01-19 中移互联网有限公司 Speech recognition model training method and device and computer readable storage medium

Similar Documents

Publication Publication Date Title
US11627221B2 (en) Semiautomated relay method and apparatus
US10255918B2 (en) Command and control of devices and applications by voice using a communication base system
CN105869626B (en) A kind of method and terminal of word speed automatic adjustment
US9547642B2 (en) Voice to text to voice processing
US10217466B2 (en) Voice data compensation with machine learning
US9924012B2 (en) Apparatus for answering a phone call when a recipient of the phone call decides that it is inappropriate to talk, and related method
US20110270601A1 (en) Universal translator
US20190130176A1 (en) Performing artificial intelligence sign language translation services in a video relay service environment
US11528568B1 (en) Assisted hearing aid with synthetic substitution
KR102044689B1 (en) System and method for creating broadcast subtitle
CN108093653B (en) Voice prompt method, recording medium and voice prompt system
KR101819459B1 (en) Voice recognition system and apparatus supporting voice recognition error correction
CN110166729A (en) Cloud video-meeting method, device, system, medium and calculating equipment
US20150287332A1 (en) Distance Education Method and Server Device for Providing Distance Education
CN106981289A (en) A kind of identification model training method and system and intelligent terminal
CN108735220A (en) A kind of language learning intelligent earphone, intelligent interactive system and man-machine interaction method
Liao et al. Formosa speech recognition challenge 2020 and Taiwanese across Taiwan corpus
CN107766333A (en) A kind of intelligent translation apparatus, system and method
CN103646645A (en) Method based on voice translation text output
CN110915239A (en) On-line automatic audio transcription for hearing aid users
US20200075000A1 (en) System and method for broadcasting from a group of speakers to a group of listeners
CN102104657A (en) Alarm clock reminding method, device and mobile terminal
CN114154636A (en) Data processing method, electronic device and computer program product
CN115171645A (en) Dubbing method and device, electronic equipment and storage medium
JP2019159099A (en) Music reproduction system

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20181102

WD01 Invention patent application deemed withdrawn after publication