CN109493658A - Situated human-computer dialogue formula spoken language interactive learning method - Google Patents
Situated human-computer dialogue formula spoken language interactive learning method Download PDFInfo
- Publication number
- CN109493658A CN109493658A CN201910014790.1A CN201910014790A CN109493658A CN 109493658 A CN109493658 A CN 109493658A CN 201910014790 A CN201910014790 A CN 201910014790A CN 109493658 A CN109493658 A CN 109493658A
- Authority
- CN
- China
- Prior art keywords
- content
- spoken
- human
- sentence
- oral
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 20
- 230000002452 interceptive effect Effects 0.000 title claims abstract description 15
- 230000000694 effects Effects 0.000 abstract description 5
- 230000003993 interaction Effects 0.000 abstract description 5
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004891 communication Methods 0.000 description 1
- 238000010586 diagram Methods 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/065—Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Multimedia (AREA)
- General Physics & Mathematics (AREA)
- Educational Administration (AREA)
- Educational Technology (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
The invention discloses a kind of situated human-computer dialogue formula spoken language interactive learning methods, belong to language learning field.Include the following steps, S1, the conversation content being placed under one or more contextual models generate trainable spoken content;S2, Oral Training is opened, into human-computer conversational mode;S3, the preceding sentence that spoken content is given expression to by voice mode, expressed content of the user according to preceding sentence, the content of sentence after being gone out by oral expression;S4, the rear sentence content gone out to user's oral expression are acquired, analyze, score and feed back;S5, step S3, S4 is repeated until situational dialogues terminates, completion spoken dialog is trained.Spoken by human-computer interaction dialogue exercise, correction promotes pronunciation standard, improves Oral Activities and students' ability to use language, and specific aim Oral Training is carried out in certain scenarios, and user is made to be familiar with and grasp the oral expression of some scene.
Description
Technical field
The present invention relates to a kind of situated human-computer dialogue formula spoken language interactive learning methods, belong to language learning field.
Background technique
Language is one of national important feature and the most important the vehicle of communication of the mankind, is the master that people are linked up
Expression way is wanted, the achievement of human civilization is saved and transmitted by language.The study of existing language, mostly being imparted knowledge to students based on books is
Main, such learning method one side efficiency is lower, and oracy cannot effectively improve, be easy to appear it is readable but not
The case where saying;It on the other hand is in specific situation, learner not can be carried out specific aim study, the language table in certain scenarios
It cannot be improved up to mode and ability to express.
Summary of the invention
Technical problem to be solved by the present invention lies in: a kind of situated human-computer dialogue formula spoken language interactive learning side is provided
Method, spoken by human-computer interaction dialogue exercise, correction promotes pronunciation standard, improves Oral Activities and students' ability to use language,
It is familiar with and grasps the oral expression of some scene.
The technical problems to be solved by the invention take following technical scheme to realize:
Situated human-computer dialogue formula spoken language interactive learning method, includes the following steps,
S1, the conversation content being placed under one or more contextual models generate trainable spoken content;
S2, Oral Training is opened, into human-computer conversational mode;
S3, the preceding sentence that spoken content is given expression to by voice mode, user pass through spoken table according to the expressed content of preceding sentence
The content of sentence up to after going out;
S4, the rear sentence content gone out to user's oral expression are acquired, analyze, score and feed back;
S5, step S3, S4 is repeated until situational dialogues terminates, completion spoken dialog is trained.
As preferred embodiment, in step S4, passes through recording form and acquire effective audio content.
As preferred embodiment, in step S4, analytic process specifically: judge that volume, ambient sound interference and acquisition have
The English content phrase of effect, oral expression content.
Scoring element as preferred embodiment, in step S4, in the link that scores are as follows: syllable judgement, fluency, integrity degree,
Accuracy, word speed.
As preferred embodiment, in step S4, feedback element are as follows: according to appraisal result, provide word in oral expression content,
Syllable it is excellent, good, in, difference is as a result, according to the matching degree of oral expression content and preceding sentence content and integrity degree and accuracy
Feedback is the dialogue training for repeatedly practicing also being to continue under next contextual model.
As preferred embodiment, in step s3, for the content that embodies of preceding sentence, it can also provide corresponding keyword and mention
Show and/or whole sentence prompts.
As preferred embodiment, keyword prompt and/or the prompt of whole sentence are instantly prompting.
As preferred embodiment, keyword prompt and/or the prompt of whole sentence have two kinds of Chinese and English.
The beneficial effects of the present invention are: spoken by human-computer interaction dialogue exercise, correction promotes pronunciation standard, improves mouth
Language ability to express and students' ability to use language carry out specific aim Oral Training in certain scenarios, and user is made to be familiar with and grasp some
The oral expression of scene.
Detailed description of the invention
Fig. 1 is flow chart of the invention;
Fig. 2 is a kind of effect picture of specific embodiment in the present invention.
Specific embodiment
In order to be easy to understand to technical means, creative features, achievable purpose and effectiveness of the invention, below with reference to tool
Body diagram, the present invention is further explained.
As shown in Figure 1, situated human-computer dialogue formula spoken language interactive learning method, includes the following steps,
S1, the conversation content being placed under one or more contextual models generate trainable spoken content;
Specific spoken language content is set according to contextual model, can be a variety of situational contexts such as to strike up a conversation with, greet, using question and answer mode into
Row human-computer interaction.
S2, Oral Training is opened, into human-computer conversational mode;
S3, load bearing equipment give expression to the preceding sentence of spoken content by voice mode, and user is led to according to the expressed content of preceding sentence
Cross the content of sentence after oral expression goes out;
S4, the rear sentence content gone out to user's oral expression are acquired, analyze, score and feed back;
S5, step S3, S4 is repeated until situational dialogues terminates, completion spoken dialog is trained.
Load bearing equipment of the invention can be the plurality of devices such as mobile phone, tablet computer, computer, be planted according to the needs of use
Enter.
More specifically, in step S4, effective audio content is acquired by recording form.
More specifically, in step S4, analytic process specifically: judge that volume, ambient sound interference and acquisition are effective
English content phrase, oral expression content.
More specifically, in step S4, score link in scoring element are as follows: syllable judgement, fluency, integrity degree, accurately
Degree, word speed.
More specifically, in step S4, feedback element are as follows: according to appraisal result, provide word, syllable in oral expression content
It is excellent, good, in, difference is as a result, according to the matching degree of oral expression content and preceding sentence content and integrity degree and accuracy feedback
It is the dialogue training for repeatedly practicing also being to continue under next contextual model.
Further, in step s3, for the content that embodies of preceding sentence, corresponding keyword prompt can also be provided
And/or whole sentence prompt.
Further, keyword prompt and/or the prompt of whole sentence are instantly prompting.
Further, keyword prompt and/or the prompt of whole sentence have two kinds of Chinese and English.
In primary training, it can only be trained, can also set just for the spoken dialog in some certain scenarios
The spoken dialog entered in a variety of scenes is repeatedly trained.
Signified "and/or" in the present invention, the scheme represented are one such with while using A and B using A or B
Three kinds of schemes.
As shown in Fig. 2, load bearing equipment is that mobile phone passes through voice when training in a kind of specific embodiment wherein
Sentence content before playing out, and reply rear sentence the Chinese and English prompt for carrying out positive sentence, score after user answers and provide anti-
Feedback, if failing the oral expression content for clearly collecting user, is expressed prompt again.
Spoken by human-computer interaction dialogue exercise in the present invention, correction promotes pronunciation standard, improves Oral Activities
And students' ability to use language, specific aim Oral Training is carried out in certain scenarios, and user is made to be familiar with and grasp the spoken language of some scene
Expression.
The above shows and describes the basic principles and main features of the present invention and the advantages of the present invention.The technology of the industry
Personnel are it should be appreciated that the present invention is not limited to the above embodiments, without departing from the spirit and scope of the present invention, this hair
Bright to will also have various changes and improvements, these changes and improvements are both fallen in the scope of protection of present invention.The present invention claims
Protection scope is defined by the appending claims and its equivalent thereof.
Claims (8)
1. situated human-computer dialogue formula spoken language interactive learning method, it is characterised in that: include the following steps,
S1, the conversation content being placed under one or more contextual models generate trainable spoken content;
S2, Oral Training is opened, into human-computer conversational mode;
S3, the preceding sentence that spoken content is given expression to by voice mode, user pass through spoken table according to the expressed content of preceding sentence
The content of sentence up to after going out;
S4, the rear sentence content gone out to user's oral expression are acquired, analyze, score and feed back;
S5, step S3, S4 is repeated until situational dialogues terminates, completion spoken dialog is trained.
2. situated human-computer dialogue formula spoken language interactive learning method according to claim 1, it is characterised in that: step S4
In, effective audio content is acquired by recording form.
3. situated human-computer dialogue formula spoken language interactive learning method according to claim 1, it is characterised in that: step S4
In, analytic process specifically: judge volume, ambient sound interference and acquire effective English content phrase, in oral expression
Hold.
4. situated human-computer dialogue formula spoken language interactive learning method according to claim 1, it is characterised in that: step S4
In, the scoring element in the link that scores are as follows: syllable judgement, fluency, integrity degree, accuracy, word speed.
5. situated human-computer dialogue formula spoken language interactive learning method according to claim 1, it is characterised in that: step S4
In, feedback element are as follows: according to appraisal result, provide word in oral expression content, syllable it is excellent, good, in, difference as a result, according to
The matching degree of oral expression content and preceding sentence content and integrity degree and accuracy feedback be repeat practice be also to continue with it is next
Dialogue training under contextual model.
6. situated human-computer dialogue formula spoken language interactive learning method according to claim 1, it is characterised in that: in step S3
In, for the content that embodies of preceding sentence, it can also provide corresponding keyword prompt and/or the prompt of whole sentence.
7. situated human-computer dialogue formula spoken language interactive learning method according to claim 6, it is characterised in that: keyword mentions
Show and/or the prompt of whole sentence is instantly prompting.
8. situated human-computer dialogue formula spoken language interactive learning method according to claim 6 or 7, it is characterised in that: crucial
Word prompt and/or the prompt of whole sentence have two kinds of Chinese and English.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910014790.1A CN109493658A (en) | 2019-01-08 | 2019-01-08 | Situated human-computer dialogue formula spoken language interactive learning method |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910014790.1A CN109493658A (en) | 2019-01-08 | 2019-01-08 | Situated human-computer dialogue formula spoken language interactive learning method |
Publications (1)
Publication Number | Publication Date |
---|---|
CN109493658A true CN109493658A (en) | 2019-03-19 |
Family
ID=65714154
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201910014790.1A Pending CN109493658A (en) | 2019-01-08 | 2019-01-08 | Situated human-computer dialogue formula spoken language interactive learning method |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN109493658A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110148430A (en) * | 2019-04-30 | 2019-08-20 | 腾讯科技(深圳)有限公司 | Method for recording and device, the storage medium and electronic device of audio |
CN110444056A (en) * | 2019-08-15 | 2019-11-12 | 湖北纽云教育科技发展有限公司 | A kind of application method of English conversational system |
CN110992754A (en) * | 2019-12-02 | 2020-04-10 | 王言之 | High-efficiency pre-examination, self-learning and teaching method for oral English |
CN111625717A (en) * | 2020-05-15 | 2020-09-04 | 广东小天才科技有限公司 | Task recommendation method and device in learning scene and electronic equipment |
CN111833671A (en) * | 2020-08-03 | 2020-10-27 | 张晶 | Circulation feedback type English teaching demonstration device |
CN115798513A (en) * | 2023-01-31 | 2023-03-14 | 新励成教育科技股份有限公司 | Talent expression management method, system and computer readable storage medium |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20040002176A (en) * | 2002-06-29 | 2004-01-07 | 인벤텍 코오포레이션 | Computer-assisted language listening and speaking teaching system and method with circumstantial shadow and assessment functions |
CN101366065A (en) * | 2005-11-30 | 2009-02-11 | 语文交流企业公司 | Interactive language education system and method |
CN101739870A (en) * | 2009-12-03 | 2010-06-16 | 深圳先进技术研究院 | Interactive language learning system and method |
CN103928023A (en) * | 2014-04-29 | 2014-07-16 | 广东外语外贸大学 | Voice scoring method and system |
US20170069226A1 (en) * | 2015-09-03 | 2017-03-09 | Amy Spinelli | System and method for diarization based dialogue analysis |
CN106558252A (en) * | 2015-09-28 | 2017-04-05 | 百度在线网络技术(北京)有限公司 | By computer implemented spoken language exercise method and device |
CN107172157A (en) * | 2017-05-22 | 2017-09-15 | 浙江精益佰汇数字技术有限公司 | The interactive English teaching system platform of many people and implementation method |
CN108153415A (en) * | 2017-12-22 | 2018-06-12 | 歌尔科技有限公司 | Virtual reality language teaching interaction method and virtual reality device |
-
2019
- 2019-01-08 CN CN201910014790.1A patent/CN109493658A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
KR20040002176A (en) * | 2002-06-29 | 2004-01-07 | 인벤텍 코오포레이션 | Computer-assisted language listening and speaking teaching system and method with circumstantial shadow and assessment functions |
CN101366065A (en) * | 2005-11-30 | 2009-02-11 | 语文交流企业公司 | Interactive language education system and method |
CN101739870A (en) * | 2009-12-03 | 2010-06-16 | 深圳先进技术研究院 | Interactive language learning system and method |
CN103928023A (en) * | 2014-04-29 | 2014-07-16 | 广东外语外贸大学 | Voice scoring method and system |
US20170069226A1 (en) * | 2015-09-03 | 2017-03-09 | Amy Spinelli | System and method for diarization based dialogue analysis |
CN106558252A (en) * | 2015-09-28 | 2017-04-05 | 百度在线网络技术(北京)有限公司 | By computer implemented spoken language exercise method and device |
CN107172157A (en) * | 2017-05-22 | 2017-09-15 | 浙江精益佰汇数字技术有限公司 | The interactive English teaching system platform of many people and implementation method |
CN108153415A (en) * | 2017-12-22 | 2018-06-12 | 歌尔科技有限公司 | Virtual reality language teaching interaction method and virtual reality device |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110148430A (en) * | 2019-04-30 | 2019-08-20 | 腾讯科技(深圳)有限公司 | Method for recording and device, the storage medium and electronic device of audio |
CN110444056A (en) * | 2019-08-15 | 2019-11-12 | 湖北纽云教育科技发展有限公司 | A kind of application method of English conversational system |
CN110992754A (en) * | 2019-12-02 | 2020-04-10 | 王言之 | High-efficiency pre-examination, self-learning and teaching method for oral English |
CN110992754B (en) * | 2019-12-02 | 2021-09-28 | 王言之 | High-efficiency pre-examination, self-learning and teaching method for oral English |
CN111625717A (en) * | 2020-05-15 | 2020-09-04 | 广东小天才科技有限公司 | Task recommendation method and device in learning scene and electronic equipment |
CN111625717B (en) * | 2020-05-15 | 2024-03-19 | 广东小天才科技有限公司 | Task recommendation method and device under learning scene and electronic equipment |
CN111833671A (en) * | 2020-08-03 | 2020-10-27 | 张晶 | Circulation feedback type English teaching demonstration device |
CN115798513A (en) * | 2023-01-31 | 2023-03-14 | 新励成教育科技股份有限公司 | Talent expression management method, system and computer readable storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN109493658A (en) | Situated human-computer dialogue formula spoken language interactive learning method | |
CN107203953B (en) | Teaching system based on internet, expression recognition and voice recognition and implementation method thereof | |
CN108335543A (en) | A kind of English dialogue training learning system | |
CN112307742A (en) | Session type human-computer interaction spoken language evaluation method, device and storage medium | |
CN101488121A (en) | Language learning method | |
WO2005099414A2 (en) | Comprehensive spoken language learning system | |
Alemi et al. | Virtual reality assisted pronunciation training (VRAPT) for young EFL learners | |
CN107967844A (en) | A kind of patients with Chinese aphasia mental language training method and training system | |
KR101438087B1 (en) | Method for providing language training service based on consecutive and simultaneous interpretation test using speech recognition engine | |
JP2018180299A (en) | Word learning training system | |
Myers | Voice recognition software and a hand-held translation machine for second-language learning | |
KR101967849B1 (en) | Foreign language acquisition practice method through the combination of shadowing and speed listening based on the processes of mother language acquisition, apparatus and computer readable program medium thereof | |
Chen | Gestures as tone markers in multilingual communication | |
CN201600791U (en) | Electronic device for learning Chinese characters | |
CN109473007B (en) | English natural spelling teaching method and system combining phonemes with sound side | |
Strik et al. | Developing a CALL system for practicing oral proficiency: How to design for speech technology, pedagogy and learners | |
Goodenough-Trepagnier et al. | Communication systems for the nonvocal based on frequent phoneme sequences | |
CN109741643A (en) | The word processing method of text-oriented big data | |
CN108717854A (en) | Method for distinguishing speek person based on optimization GFCC characteristic parameters | |
Listiawati et al. | Speaking Habit as a Strategy to Improve the Students' Speaking Skill | |
CN112863268B (en) | Spoken English man-machine conversation device | |
CN103021226A (en) | Voice evaluating method and device based on pronunciation rhythms | |
TW583610B (en) | System and method using computer to train listening comprehension and pronunciation | |
Çekiç | The effects of computer assisted pronunciation teaching on the listening comprehension of Intermediate learners | |
Zhang | The Practice of College English Teaching Reform with Intelligent Voice Evaluation in the Context of Artificial Intelligence Empowering Education |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20190319 |
|
RJ01 | Rejection of invention patent application after publication |