CN111540351A - Method for controlling interactive live broadcast classroom by using voice instruction - Google Patents

Method for controlling interactive live broadcast classroom by using voice instruction Download PDF

Info

Publication number
CN111540351A
CN111540351A CN202010269224.8A CN202010269224A CN111540351A CN 111540351 A CN111540351 A CN 111540351A CN 202010269224 A CN202010269224 A CN 202010269224A CN 111540351 A CN111540351 A CN 111540351A
Authority
CN
China
Prior art keywords
voice
compiled
information
characters
executed
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010269224.8A
Other languages
Chinese (zh)
Other versions
CN111540351B (en
Inventor
袭普照
曹鑫宇
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Able Digital & Tech Co ltd
Original Assignee
Shanghai Able Digital & Tech Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Able Digital & Tech Co ltd filed Critical Shanghai Able Digital & Tech Co ltd
Priority to CN202010269224.8A priority Critical patent/CN111540351B/en
Publication of CN111540351A publication Critical patent/CN111540351A/en
Application granted granted Critical
Publication of CN111540351B publication Critical patent/CN111540351B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L21/00Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
    • G10L21/02Speech enhancement, e.g. noise reduction or echo cancellation
    • G10L21/0208Noise filtering

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Signal Processing (AREA)
  • Electrically Operated Instructional Devices (AREA)

Abstract

The invention relates to the field of network online live broadcast control, in particular to a method for controlling an interactive live broadcast classroom by using a voice instruction. The live classroom based on the Internet comprises a voice initiating end and an executed end, and the method comprises the following steps: s1, the voice initiating terminal receives the external voice and decodes and translates the external voice; s2, generating an interaction control instruction according to the decoded and translated compiled voice characters; and S3, issuing the interaction control instruction to the corresponding executed end for execution. According to the invention, through acquiring voice and sending out the corresponding interactive control instruction after intelligent recognition, intelligent control is realized, manual operation in the control process is reduced, and the control of the electronic equipment is more intelligent and flexible.

Description

Method for controlling interactive live broadcast classroom by using voice instruction
Technical Field
The invention relates to the field of network online live broadcast control, in particular to a method for controlling an interactive live broadcast classroom by using a voice instruction.
Background
The control panel and various button switches of the equipment occupy a considerable space, so that the number of executable commands of the whole live broadcast classroom is limited, and the upgrading of programs and functions of the live broadcast classroom is inconvenient; the control panel and the button switch of the devices are very complicated, manual instruction input step by step has high requirements on operators, the manual instruction input devices are scattered at all places of a classroom, the operation is not very convenient, the efficiency of class taking in the interactive live classroom is greatly reduced, and meanwhile, the whole live classroom is also short of the function of communicating with people. In order to improve the efficiency of taking class for the interactive live-broadcast class-chamber, reduce the operation steps of teachers and students and reduce the operation difficulty of teachers and students, the increasingly complete voice recognition technology is considered, the efficiency of taking class for the interactive live-broadcast class-chamber is expected to be improved through the voice recognition technology, and the interactive experience of the class is further improved.
Disclosure of Invention
Aiming at the defects of the prior art, the invention provides a method for controlling an interactive live broadcast classroom by using a voice instruction.
The specific technical scheme of the invention is as follows: a method for controlling an interactive live classroom by using a voice instruction is used for a live classroom based on the Internet and comprises a voice initiating end and an executed end, and the method comprises the following steps:
s1, the voice initiating terminal receives the external voice and decodes and translates the external voice;
s2, generating an interaction control instruction according to the decoded and translated compiled voice characters;
and S3, issuing the interaction control instruction to the corresponding executed end for execution.
Preferably, the step S1 includes:
s1.1a, the voice signal initiating end receives the external voice through a radio device and converts the external voice into a second voice signal which can be recognized by the voice initiating end;
s1.2a, the speech signal obtains the compiled speech character through decoding and translation.
Preferably, step S1 includes: and receiving the external voice and performing noise filtration through a voice recognition module to generate a first voice signal.
Preferably, the step S1 includes:
s1.1b, receiving the external voice, performing noise filtering through a voice recognition module, and converting the external voice into a third voice signal which can be recognized by a voice initiating end;
s1.2b, the speech signal obtains the compiled speech character through decoding and translation.
Preferably, the compiled phonetic characters include identity authority information of a speech initiating terminal, corresponding identity information of the executed terminal, and command information.
Preferably, step s1.2b includes:
judging whether the compiled voice characters comprise voice initiating terminal identity authority information, corresponding executed terminal identity information and command information or not, and if the compiled voice characters are all possessed by the voice initiating terminal identity authority information, obtaining the compiled voice characters through decoding and translation; and if the judgment result shows that the information is not completely provided, generating first feedback information for error prompt.
Preferably, the step S2 includes:
s2.1, acquiring and storing the compiled phonetic characters;
and S2.2, re-compiling according to the compiled voice characters to generate the interaction control instruction.
Preferably, step S2.2 includes:
s2.21, extracting the compiled voice characters, judging whether the voice initiating terminal has the authority to execute the command on the corresponding executed terminal or not according to the identity authority information of the voice initiating terminal, the corresponding executed terminal identity information and the command information, and if the judgment result is that the voice initiating terminal has the authority, performing the next step; if the judgment result is that no authority exists, second feedback information refusing to be executed is generated;
and S2.22, generating the interactive control instruction.
Preferably, the second feedback information includes reject reason information.
Preferably, the method further includes, after step S3:
and S4, judging whether the received interaction control instruction is successfully executed, and if not, generating third feedback information of failed execution to the voice initiating terminal.
In conclusion, the invention has the following beneficial effects:
1. based on voice recognition, the compiled voice characters are obtained after parsing and compiling, an interaction control command is generated, and finally the command is issued to the corresponding executed end to execute control, so that the corresponding executed end completes corresponding actions, the efficiency of interactive live broadcast class administration is improved, and the interactive experience of a class is further improved.
2. And a noise filtering link is added, so that the probability of generating an error instruction due to noise interference in the instruction issuing process is removed, and the accuracy of voice control is improved.
Drawings
FIG. 1 is a diagram illustrating a method for controlling an interactive live classroom using voice commands according to the present invention;
FIG. 2 is a flow chart of a method for controlling an interactive live classroom using voice commands without feedback from the execution of the method according to the present invention;
FIG. 3 is a flowchart of step S1 in FIG. 2;
fig. 4 is another flowchart of step S1 in fig. 2;
FIG. 5 is a flowchart of step S2 in FIG. 2;
FIG. 6 is a flowchart of step S2.2 of FIG. 2;
fig. 7 is a flowchart illustrating a method for controlling an interactive live classroom using voice commands according to the present invention with feedback of execution status.
Detailed Description
The invention will be further explained by means of specific embodiments with reference to the drawings.
Embodiment 1, as shown in fig. 1, a method for controlling an interactive live classroom by using voice commands is used for an internet-based live classroom and comprises a voice initiating end and an executed end. The execution main body of the voice interaction method is a voice initiating end and an executed end, and the voice initiating end and the executed end are a teacher use terminal and a student use terminal. When a teacher wants to control a terminal of the teacher through voice, the voice initiating terminal and the executed terminal are both teacher use terminals; when a teacher wants to control students to use a terminal through voice, the voice initiating end is the teacher use terminal, and the executed end is the student use terminal; when a student needs to simply feed back a teacher, the voice initiating end is a student use terminal, and the executed end is a teacher use terminal; when the student needs to control own equipment through voice, the voice initiating end and the executed end are both student use terminals.
As shown in fig. 2, the method of the present invention comprises the steps of:
s1, the voice initiating terminal receives the external voice and decodes and interprets it. Firstly, an external voice is recognized through a voice recognition module and then transmitted to the voice initiating end, and the compiled voice characters are generated after analysis and compilation. The voice recognition module is used for recognizing the whole language, and the preferred default language in the scheme is Chinese Putonghua.
And S2, generating an interaction control command according to the decoded and translated compiled phonetic characters. The compiled phonetic characters are stored in a terminal (preferably a computer) processor of the voice initiating terminal, and the interactive control instruction is generated by a voice recognition module program.
And S3, issuing the interaction control instruction to the corresponding executed end for execution.
Based on voice recognition, the compiled voice characters are obtained after parsing and compiling, an interaction control command is generated, and finally the command is issued to the corresponding executed end to execute control, so that the corresponding executed end completes corresponding actions, the efficiency of interactive live broadcast class administration is improved, and the interactive experience of a class is further improved.
Example 2
Based on embodiment 1, the difference from the embodiment is that:
as shown in fig. 3, the step S1 includes:
s1.1a, the voice signal initiating end receives the external voice through a radio device and converts the external voice into a second voice signal which can be recognized by the voice initiating end.
S1.2a, the speech signal obtains the compiled speech character through decoding and translation.
Furthermore, the products of voice input are more and more popular nowadays, and the application scenes of voice input are more and more common. However, the speech input always affects the recognition rate due to noise interference in the application scene, and how to eliminate the scene noise and improve the speech recognition rate is a very critical problem. As shown in fig. 4, the step S1 includes: and receiving the external voice and performing noise filtration through a voice recognition module to generate a first voice signal.
The sound collection is carried out in five levels, namely, a loud language, a normal language, a small language, a fine language and a whisper language, thresholds of various technical parameters of amplitude, frequency, wavelength, sound, tone, timbre, sound source vector and the like of sound of each level are set, a sound spectrum of a quiet environment authority user is generated, and the system senses and memorizes the sound spectrum of the authority user. After the sound spectrum under the quiet environment is generated, the system automatically synthesizes the sound spectrum of the authority user with various application scene noises collected in advance to generate the sound spectrum of the authority user under the noise background, senses and memorizes the sound spectrum of the authority user under the noise background, and automatically trains the sensitivity of voice input for detecting, identifying and discriminating the sound spectrum of the authority user under the actual application scene.
The embodiment can avoid the voice equipment from executing the voice instruction in the surrounding environment mistakenly, thereby improving the voice interaction efficiency and improving the experience of the user in using the voice equipment.
As shown in fig. 4, the step S1 includes:
s1.1b, receiving the external voice, performing noise filtering through a voice recognition module, and converting the external voice into a third voice signal which can be recognized by a voice initiating end;
s1.2b, the speech signal obtains the compiled speech character through decoding and translation.
Wherein the step S1.2b comprises:
judging whether the compiled voice characters comprise voice initiating terminal identity authority information, corresponding executed terminal identity information and command information or not, and if the compiled voice characters are all possessed by the voice initiating terminal identity authority information, obtaining the compiled voice characters through decoding and translation; and if the judgment result shows that the information is not completely provided, generating first feedback information for error prompt. For example, the language teacher sends a command of opening the student end MATLAB software, the language teacher does not control the authority of the student end MATLAB software switch, and after judging that the authority is not controlled, the language teacher returns the first feedback information of 'no authority to perform the operation'.
Example 3
Based on embodiment 2, the difference from embodiment 2 is that:
the compiled phonetic characters comprise identity authority information of a phonetic initiating terminal, corresponding identity information of the executed terminal and command information.
As shown in fig. 5 to 6, the step S2 includes:
and S2.1, acquiring and storing the compiled phonetic characters.
And S2.2, re-compiling according to the compiled voice characters to generate the interaction control instruction.
Said step S2.2 comprises:
s2.21, extracting the compiled voice characters, judging whether the voice initiating terminal has the authority to execute the command on the corresponding executed terminal or not according to the identity authority information of the voice initiating terminal, the corresponding executed terminal identity information and the command information, and if the judgment result is that the voice initiating terminal has the authority, performing the next step; and if the judgment result is that no authority exists, generating second feedback information refusing to be executed.
Wherein the second feedback information includes reason information for rejection.
And S2.22, generating the interactive control instruction.
Example 4
Based on embodiment 1, the difference from embodiment 1 is that:
as shown in fig. 7, the step S3 is followed by:
and S4, judging whether the received interaction control instruction is successfully executed, and if not, generating third feedback information of failed execution to the voice initiating terminal.
When the voice initiating terminal receives the third feedback information containing the execution failure, the voice initiating terminal can issue the voice command again and control the corresponding executed terminal again.
The above-described embodiments are merely illustrative of the preferred embodiments of the present invention and do not limit the spirit and scope of the present invention. Various modifications and improvements of the technical solutions of the present invention may be made by those skilled in the art without departing from the design concept of the present invention, and the technical contents of the present invention are all described in the claims.

Claims (10)

1. A method for controlling an interactive live classroom by using a voice instruction is characterized by being used for a live classroom based on the Internet and comprising a voice initiating end and an executed end, and the method comprises the following steps:
s1, the voice initiating terminal receives the external voice and decodes and translates the external voice;
s2, generating an interaction control instruction according to the decoded and translated compiled voice characters;
and S3, issuing the interaction control instruction to the corresponding executed end for execution.
2. The method for controlling an interactive live classroom according to claim 1, wherein said step S1 includes:
s1.1a, the voice signal initiating end receives the external voice through a radio device and converts the external voice into a second voice signal which can be recognized by the voice initiating end;
s1.2a, the speech signal obtains the compiled speech character through decoding and translation.
3. The method as claimed in claim 2, wherein the step S1 includes: and receiving the external voice and performing noise filtration through a voice recognition module to generate a first voice signal.
4. The method for controlling an interactive live classroom according to claim 3, wherein said step S1 includes:
s1.1b, receiving the external voice, performing noise filtering through a voice recognition module, and converting the external voice into a third voice signal which can be recognized by a voice initiating end;
s1.2b, the speech signal obtains the compiled speech character through decoding and translation.
5. The method for controlling an interactive live classroom according to claim 4, wherein said step S1.2b includes:
judging whether the compiled voice characters comprise voice initiating terminal identity authority information, corresponding executed terminal identity information and command information or not, and if the compiled voice characters are all possessed by the voice initiating terminal identity authority information, obtaining the compiled voice characters through decoding and translation; and if the judgment result shows that the information is not completely provided, generating first feedback information for error prompt.
6. The method as claimed in claim 2 or 4, wherein the compiled phonetic characters include identity authority information of the speech initiator, identity information of the corresponding executed terminal, and command information.
7. The method as claimed in claim 5, wherein the step S2 includes:
s2.1, acquiring and storing the compiled phonetic characters;
and S2.2, re-compiling according to the compiled voice characters to generate the interaction control instruction.
8. A method for controlling an interactive live classroom according to claim 6, wherein said step S2.2 includes:
s2.21, extracting the compiled voice characters, judging whether the voice initiating terminal has the authority to execute the command on the corresponding executed terminal or not according to the identity authority information of the voice initiating terminal, the corresponding executed terminal identity information and the command information, and if the judgment result is that the voice initiating terminal has the authority, performing the next step; if the judgment result is that no authority exists, second feedback information refusing to be executed is generated;
and S2.22, generating the interactive control instruction.
9. The method as claimed in claim 7, wherein the second feedback information includes reason for rejection information.
10. The method as claimed in claim 1, wherein the step S3 is further followed by:
and S4, judging whether the received interaction control instruction is successfully executed, and if not, generating third feedback information of failed execution to the voice initiating terminal.
CN202010269224.8A 2020-04-08 2020-04-08 Method for controlling interactive live broadcast classroom by using voice instruction Active CN111540351B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010269224.8A CN111540351B (en) 2020-04-08 2020-04-08 Method for controlling interactive live broadcast classroom by using voice instruction

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010269224.8A CN111540351B (en) 2020-04-08 2020-04-08 Method for controlling interactive live broadcast classroom by using voice instruction

Publications (2)

Publication Number Publication Date
CN111540351A true CN111540351A (en) 2020-08-14
CN111540351B CN111540351B (en) 2021-05-07

Family

ID=71978603

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010269224.8A Active CN111540351B (en) 2020-04-08 2020-04-08 Method for controlling interactive live broadcast classroom by using voice instruction

Country Status (1)

Country Link
CN (1) CN111540351B (en)

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN105895098A (en) * 2016-06-08 2016-08-24 乐视控股(北京)有限公司 Play control method and device
CN205788812U (en) * 2016-02-03 2016-12-07 成都中智融通科技有限公司 A kind of smart classroom speech control system
CN106297454A (en) * 2016-09-07 2017-01-04 苏州国锦智承信息科技有限公司 A kind of mobile classroom interaction teaching system
CN107341747A (en) * 2017-07-14 2017-11-10 安徽智星交通科技股份有限公司 Class management method and system
CN107517109A (en) * 2016-06-17 2017-12-26 上海理优教育科技有限公司 A kind of more people's real-time voice scheme in communication suitable for Mobile Online's education
US9911420B1 (en) * 2014-12-02 2018-03-06 Google Llc Behavior adjustment using speech recognition system
CN108389578A (en) * 2018-02-09 2018-08-10 深圳市鹰硕技术有限公司 Smart classroom speech control system
CN108806376A (en) * 2018-06-16 2018-11-13 佛山市贝思特科技有限公司 A kind of remote interactive teaching method and device of Mobile Teaching resource
CN109256126A (en) * 2018-10-16 2019-01-22 视联动力信息技术股份有限公司 A kind of view networking service execution method and apparatus
CN109686155A (en) * 2019-01-30 2019-04-26 重庆康普达科技有限公司 A kind of authority distributing method for children education system
US20190266999A1 (en) * 2018-02-27 2019-08-29 Microsoft Technology Licensing, Llc Empathetic personal virtual digital assistant

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9911420B1 (en) * 2014-12-02 2018-03-06 Google Llc Behavior adjustment using speech recognition system
CN205788812U (en) * 2016-02-03 2016-12-07 成都中智融通科技有限公司 A kind of smart classroom speech control system
CN105895098A (en) * 2016-06-08 2016-08-24 乐视控股(北京)有限公司 Play control method and device
CN107517109A (en) * 2016-06-17 2017-12-26 上海理优教育科技有限公司 A kind of more people's real-time voice scheme in communication suitable for Mobile Online's education
CN106297454A (en) * 2016-09-07 2017-01-04 苏州国锦智承信息科技有限公司 A kind of mobile classroom interaction teaching system
CN107341747A (en) * 2017-07-14 2017-11-10 安徽智星交通科技股份有限公司 Class management method and system
CN108389578A (en) * 2018-02-09 2018-08-10 深圳市鹰硕技术有限公司 Smart classroom speech control system
US20190266999A1 (en) * 2018-02-27 2019-08-29 Microsoft Technology Licensing, Llc Empathetic personal virtual digital assistant
CN108806376A (en) * 2018-06-16 2018-11-13 佛山市贝思特科技有限公司 A kind of remote interactive teaching method and device of Mobile Teaching resource
CN109256126A (en) * 2018-10-16 2019-01-22 视联动力信息技术股份有限公司 A kind of view networking service execution method and apparatus
CN109686155A (en) * 2019-01-30 2019-04-26 重庆康普达科技有限公司 A kind of authority distributing method for children education system

Also Published As

Publication number Publication date
CN111540351B (en) 2021-05-07

Similar Documents

Publication Publication Date Title
US9916830B1 (en) Altering audio to improve automatic speech recognition
CN110800046B (en) Speech recognition and translation method and translation device
CN109618202B (en) Method for controlling peripheral equipment, television and readable storage medium
CN108831469B (en) Voice command customizing method, device and equipment and computer storage medium
KR101213835B1 (en) Verb error recovery in speech recognition
JP2019185062A (en) Voice interaction method, terminal apparatus, and computer readable recording medium
CN109360563B (en) Voice control method and device, storage medium and air conditioner
US20060206335A1 (en) Method for remote control of an audio device
CN107909998B (en) Voice instruction processing method and device, computer equipment and storage medium
US9087520B1 (en) Altering audio based on non-speech commands
CN111627436B (en) Voice control method and device
CN111933108B (en) Automatic testing method for intelligent voice interaction system of intelligent network terminal
US11823662B2 (en) Control method and control apparatus for speech interaction, storage medium and system
US11404052B2 (en) Service data processing method and apparatus and related device
CN107016993A (en) The voice interactive system and method for a kind of smart home
EP3422344B1 (en) Electronic device for performing operation corresponding to voice input
US8509396B2 (en) Automatic creation of complex conversational natural language call routing system for call centers
CN112102828A (en) Voice control method and system for automatically broadcasting content on large screen
CN109460548B (en) Intelligent robot-oriented story data processing method and system
CN111540351B (en) Method for controlling interactive live broadcast classroom by using voice instruction
CN105426154A (en) Voice input control method, apparatus, and terminal
CN109065018B (en) Intelligent robot-oriented story data processing method and system
CN110164020A (en) Ballot creation method, device, computer equipment and computer readable storage medium
CN212624795U (en) Interactive system, voice interaction equipment and control equipment
CN113205810A (en) Voice signal processing method, device, medium, remote controller and server

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB02 Change of applicant information
CB02 Change of applicant information

Address after: Room 901, 902, 903, 904, building 1, 1188 Qinzhou North Road, Xuhui District, Shanghai

Applicant after: Shanghai Zhuoyue Ruixin Digital Technology Co.,Ltd.

Address before: Room 901, 902, 903, 904, building 1, 1188 Qinzhou North Road, Xuhui District, Shanghai

Applicant before: SHANGHAI ABLE DIGITAL & TECH Co.,Ltd.

GR01 Patent grant
GR01 Patent grant