CN111639224A - Method for acquiring spoken language practice content and electronic equipment - Google Patents
Method for acquiring spoken language practice content and electronic equipment Download PDFInfo
- Publication number
- CN111639224A CN111639224A CN202010465305.5A CN202010465305A CN111639224A CN 111639224 A CN111639224 A CN 111639224A CN 202010465305 A CN202010465305 A CN 202010465305A CN 111639224 A CN111639224 A CN 111639224A
- Authority
- CN
- China
- Prior art keywords
- content
- sparring
- audio
- spoken language
- practice
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 62
- 238000004590 computer program Methods 0.000 claims description 11
- 230000000694 effects Effects 0.000 abstract description 11
- 238000010586 diagram Methods 0.000 description 6
- 238000004891 communication Methods 0.000 description 4
- 238000001514 detection method Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000001413 cellular effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011156 evaluation Methods 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 238000007654 immersion Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/63—Querying
- G06F16/632—Query formulation
- G06F16/634—Query by example, e.g. query by humming
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F16/00—Information retrieval; Database structures therefor; File system structures therefor
- G06F16/60—Information retrieval; Database structures therefor; File system structures therefor of audio data
- G06F16/68—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
- G06F16/683—Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06Q—INFORMATION AND COMMUNICATION TECHNOLOGY [ICT] SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES; SYSTEMS OR METHODS SPECIALLY ADAPTED FOR ADMINISTRATIVE, COMMERCIAL, FINANCIAL, MANAGERIAL OR SUPERVISORY PURPOSES, NOT OTHERWISE PROVIDED FOR
- G06Q50/00—Information and communication technology [ICT] specially adapted for implementation of business processes of specific business sectors, e.g. utilities or tourism
- G06Q50/10—Services
- G06Q50/20—Education
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B19/00—Teaching not covered by other main groups of this subclass
- G09B19/04—Speaking
-
- G—PHYSICS
- G09—EDUCATION; CRYPTOGRAPHY; DISPLAY; ADVERTISING; SEALS
- G09B—EDUCATIONAL OR DEMONSTRATION APPLIANCES; APPLIANCES FOR TEACHING, OR COMMUNICATING WITH, THE BLIND, DEAF OR MUTE; MODELS; PLANETARIA; GLOBES; MAPS; DIAGRAMS
- G09B5/00—Electrically-operated educational appliances
- G09B5/06—Electrically-operated educational appliances with both visual and audible presentation of the material to be studied
- G09B5/065—Combinations of audio and video presentations, e.g. videotapes, videodiscs, television systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L25/00—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
- G10L25/48—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use
- G10L25/51—Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 specially adapted for particular use for comparison or discrimination
Landscapes
- Engineering & Computer Science (AREA)
- Business, Economics & Management (AREA)
- Theoretical Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Multimedia (AREA)
- Educational Administration (AREA)
- Health & Medical Sciences (AREA)
- Educational Technology (AREA)
- Tourism & Hospitality (AREA)
- Databases & Information Systems (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Data Mining & Analysis (AREA)
- General Engineering & Computer Science (AREA)
- Library & Information Science (AREA)
- Economics (AREA)
- Entrepreneurship & Innovation (AREA)
- Human Resources & Organizations (AREA)
- Marketing (AREA)
- Primary Health Care (AREA)
- Strategic Management (AREA)
- General Business, Economics & Management (AREA)
- Mathematical Physics (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
The embodiment of the application discloses a method for acquiring spoken language practice content and electronic equipment, wherein the method comprises the following steps: collecting historical exercise audio of a specified user; determining pronunciation information to be corrected of the designated user by analyzing historical practice audio of the designated user; and acquiring the spoken language practice content according to the pronunciation information to be corrected of the specified user. Through implementing this application embodiment, can promote spoken language exercise effect.
Description
Technical Field
The application relates to the technical field of computers, in particular to a method for acquiring spoken language practice content and electronic equipment.
Background
The 'speaking' is one of key skills which need to be mastered when people learn a voice, and the 'speaking' usually requires students to frequently practice spoken language.
In practice, when students use the home education machine to practice spoken language, the practice content is usually acquired by the students from the database of the home education machine, and the students usually have difficulty in recognizing pronunciation defects in spoken language practice, so that the practice content acquired by the students mostly lacks pertinence, and the spoken language practice effect is often poor.
Disclosure of Invention
The embodiment of the application discloses a method for acquiring spoken language practice content and electronic equipment, and the method and the electronic equipment can improve the spoken language practice effect.
The first aspect of the embodiment of the present application discloses a method for acquiring spoken language practice content, including:
collecting historical exercise audio of a specified user;
determining pronunciation information to be corrected of the designated user by analyzing the historical practice audios;
and acquiring oral practice contents according to the pronunciation information to be corrected.
As an optional implementation manner, in the first aspect of this embodiment of this application, the acquiring spoken language practice content according to the pronunciation information to be corrected includes:
acquiring grade information of the specified user;
and acquiring oral practice contents according to the grade information and the pronunciation information to be corrected.
As an optional implementation manner, in the first aspect of this embodiment of the present application, before the acquiring the historical exercise audio of the specified user, the method further includes:
starting a sparring customization mode;
collecting customized audio in the sparring customization mode;
determining a first sound characteristic of a designated user according to the customized audio;
the collecting historical exercise audio of a specified user comprises:
and acquiring historical exercise audio of the specified user according to the first sound characteristic.
As an optional implementation manner, in the first aspect of this embodiment of this application, after the obtaining of the spoken language practice content according to the pronunciation information to be corrected, the method further includes:
determining standard audio corresponding to the spoken language practice content according to the first sound characteristic;
and playing the standard audio.
As an optional implementation manner, in the first aspect of this embodiment of the present application, after determining, according to the first sound characteristic, a standard audio corresponding to the spoken language practice content, the method further includes:
acquiring current practice audio of the specified user for the spoken language practice content;
and analyzing the current practice audio according to the standard audio to obtain a scoring result of the current practice audio.
As an optional implementation manner, in the first aspect of this embodiment of the present application, when the spoken language practice content corresponds to a plurality of conversations, after the spoken language practice content is acquired according to the pronunciation information to be corrected, the method further includes:
determining target exercise content selected by the specified user from the spoken exercise content when a sparring request is detected;
acquiring content information of a first conversation corresponding to the target exercise content; wherein the target exercise content is content information of a first character of a plurality of sparring characters of the first conversation;
and starting a sparring mode corresponding to the first character according to the content information of the first dialogue so as to enable the appointed user to carry out spoken language sparring in the sparring mode.
As an optional implementation manner, in the first aspect of the embodiment of the present application, the starting, according to the content information of the first dialog, a sparring mode corresponding to the first character includes:
when the sparring mode is a virtual sparring mode, acquiring preset audio of the sparring roles except the first role according to the content information of the first dialogue; starting a first dialogue mode corresponding to the first character according to the preset audio;
determining a second character matched with the second sound feature from the plurality of sparring characters except the first character according to a second sound feature of a sparring user of the specified user when the sparring mode is a real sparring mode; acquiring the content information of the second role from the content information of the first conversation; and sending the content information of the second role to the terminal equipment of the sparring user so as to start a second sparring mode corresponding to the first role.
A second aspect of an embodiment of the present application discloses an electronic device, including:
the acquisition unit is used for acquiring historical exercise audio of a specified user;
a determining unit, configured to determine pronunciation information to be corrected of the specified user by analyzing the historical practice audio;
and the acquisition unit is used for acquiring the oral practice content according to the pronunciation information to be corrected.
As an optional implementation manner, in a second aspect of an embodiment of the present application, the obtaining unit is specifically configured to obtain the rank information of the specified user; and acquiring oral practice content according to the grade information and the pronunciation information to be corrected.
As an optional implementation manner, in the second aspect of the embodiments of the present application, the electronic device further includes:
the customization unit is used for starting a sparring customization mode before the acquisition unit acquires the historical exercise audio of the specified user; and, collecting customized audio in the sparring customization mode; determining a first sound characteristic of a designated user according to the customized audio;
the acquisition unit is specifically configured to acquire the historical exercise audio of the specified user according to the first sound feature.
As an optional implementation manner, in the second aspect of the embodiment of the present application, the determining unit is further configured to determine a standard audio corresponding to the spoken language practice content according to the first sound feature after the obtaining unit obtains the spoken language practice content according to the pronunciation information to be corrected;
the electronic device further includes:
and the playing unit is used for playing the standard audio.
As an optional implementation manner, in a second aspect of this embodiment of this application, the acquiring unit is further configured to acquire a current practice audio of the specified user for the spoken language practice content after the determining unit determines, according to the first sound feature, a standard audio corresponding to the spoken language practice content;
the electronic device further includes:
and the scoring unit is used for analyzing the current practice audio according to the standard audio to obtain a scoring result of the current practice audio.
As an optional implementation manner, in the second aspect of the embodiment of the present application, when the spoken language practice content corresponds to a plurality of dialogs, the electronic device further includes:
the practice unit is used for determining target practice contents selected by the appointed user from the oral practice contents when a practice request is detected after the oral practice contents are acquired by the acquisition unit according to the pronunciation information to be corrected; acquiring content information of a first conversation corresponding to the target exercise content; wherein the target exercise content is content information of a first character of a plurality of sparring characters of the first conversation; and starting a dialogue mode corresponding to the first character according to the content information of the first dialogue so as to enable the appointed user to carry out spoken language dialogue.
As an optional implementation manner, in the second aspect of the embodiment of the present application, the manner in which the sparring unit starts the sparring mode corresponding to the first character is specifically:
the sparring unit is used for acquiring preset audios of the sparring roles except the first character according to the content information of the first dialogue when the sparring mode is a virtual sparring mode; starting a first dialogue mode corresponding to the first character according to the preset audio; and when the sparring mode is a real sparring mode, determining a second role matched with the second sound characteristic from the plurality of sparring roles except the first role according to the second sound characteristic of the sparring user of the specified user; acquiring the content information of the second role from the content information of the first conversation; and sending the content information of the second role to the terminal equipment of the sparring user so as to start a second sparring mode corresponding to the first role.
A third aspect of the embodiments of the present application discloses an electronic device, including:
a memory storing executable program code;
a processor coupled with the memory;
the processor calls the executable program code stored in the memory to perform part or all of the steps of any one of the methods of the first aspect of the present application.
A fourth aspect of embodiments of the present application discloses a computer-readable storage medium storing a computer program comprising a program code for performing some or all of the steps of any one of the methods of the first aspect of the present application.
A fifth aspect of embodiments of the present application discloses a computer program product, which, when run on a computer, causes the computer to perform part or all of the steps of any one of the methods of the first aspect.
A sixth aspect of embodiments of the present application discloses an application issuing system, configured to issue a computer program product, where the computer program product is configured to, when run on a computer, cause the computer to perform part or all of the steps of any one of the methods of the first aspect.
Compared with the prior art, the embodiment of the application has the following beneficial effects:
by implementing the embodiment of the application, historical exercise audio of a specified user is collected; determining pronunciation information to be corrected of the designated user by analyzing historical practice audio of the designated user; and acquiring the spoken language practice content according to the pronunciation information to be corrected of the specified user. By implementing the method, the spoken language weak points of the designated user are mined based on the historical practice audios of the designated user, and the spoken language practice content for improving the spoken language weak points of the designated user is recommended to the designated user, so that the spoken language practice content is stronger in pertinence, and the spoken language practice effect is favorably improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present application, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art that other drawings can be obtained according to these drawings without making a creative effort.
Fig. 1 is a schematic flowchart of a method for acquiring spoken language practice content disclosed in an embodiment of the present application;
FIG. 2 is a flow chart illustrating another method for acquiring spoken language practice content disclosed in the embodiments of the present application;
fig. 3 is a schematic structural diagram of an electronic device disclosed in an embodiment of the present application;
fig. 4 is a schematic structural diagram of another electronic device disclosed in the embodiments of the present application;
fig. 5 is a schematic structural diagram of an electronic device disclosed in an embodiment of the present application.
Detailed Description
The technical solutions in the embodiments of the present application will be clearly and completely described below with reference to the drawings in the embodiments of the present application, and it is obvious that the described embodiments are only a part of the embodiments of the present application, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present application.
It should be noted that the terms "comprises," "comprising," and any variations thereof in the embodiments and drawings of the present application are intended to cover non-exclusive inclusions. For example, a process, method, system, article, or apparatus that comprises a list of steps or elements is not limited to only those steps or elements listed, but may alternatively include other steps or elements not listed, or inherent to such process, method, article, or apparatus.
The method for acquiring spoken language practice content disclosed in the embodiment of the present application may be applied to an electronic device, where the electronic device may be a family education machine, and an operating system of the family education machine may include, but is not limited to, an Android operating system, an IOS operating system, a Symbian operating system, a Black Berry operating system, a Windows Phone8 operating system, and the like.
The electronic device may be a terminal device or other electronic devices. The terminal device may be referred to as a User Equipment (UE), a Mobile Station (MS), a mobile terminal (mobile terminal), an intelligent terminal, and the like, and may communicate with one or more core networks through a Radio Access Network (RAN). For example, the terminal equipment may be a mobile phone (or so-called "cellular" phone), a computer with a mobile terminal, etc., and the terminal equipment may also be a portable, pocket, hand-held, computer-included or vehicle-mounted mobile device and terminal equipment in future NR networks, which exchange voice or data with a radio access network.
The embodiment of the application discloses a method for acquiring spoken language practice content and electronic equipment, and the method and the electronic equipment can improve the spoken language practice effect. The details will be described below.
Example one
Referring to fig. 1, fig. 1 is a schematic flowchart illustrating a method for acquiring spoken language practice content according to an embodiment of the present application. The method for acquiring spoken language practice content shown in fig. 1 may specifically include the following steps:
101. historical exercise audio is collected for a given user.
In this embodiment of the present application, the historical practice audio of the specified user may be collected only by a sound pickup of the electronic device, may also be collected only by a sound pickup of a terminal device communicatively connected to the electronic device, may also be collected partially by a sound pickup of the electronic device, and partially by a sound pickup of a terminal device communicatively connected to the electronic device, which is not limited in this embodiment of the present application.
When the historical practice audio of the specified user is collected through a sound pick-up of the electronic equipment, before the step 101 is executed, the following steps can be further executed: starting a sparring customization mode; collecting customized audio in a sparring customization mode; a first sound characteristic of the designated user is determined from the customized audio. Based on the above, acquiring historical exercise audio of a specific user may include: and acquiring historical exercise audio of the specified user according to the first sound characteristic.
When the historical exercise audio of the specified user is collected through a sound collector of a terminal device which is in communication connection with the electronic device, the following steps can be further executed before the historical exercise audio of the specified user is collected: when a connection request of a first terminal device is detected, a first identifier of the first terminal device is obtained; searching user information associated with the first identifier in an equipment-user relation table preset by the electronic equipment, wherein the equipment-user relation table stores the corresponding relation between the user information and the equipment identifier; judging whether the user of the first terminal equipment is a designated user or not according to the user information associated with the first identifier; when the user of the first terminal equipment is the designated user, establishing connection with the first terminal equipment; capturing historical exercise audio for a given user may include: and receiving first exercise audio sent by the first terminal equipment, wherein the first exercise audio is used as historical exercise audio of a specified user.
It should be noted that the number of the first terminal devices may be one or more, and when the number of the first terminal devices is multiple, the efficiency of acquiring the historical exercise audio of the specified user may be improved.
102. And determining the pronunciation information to be corrected of the specified user by analyzing the historical practice audio.
In an embodiment of the present application, the pronunciation information to be corrected may include pronunciation weakness information and accent problem information, and determining the pronunciation information to be corrected for the specific user by analyzing the historical practice audio may include: acquiring exercise content corresponding to historical exercise audio; obtaining a first standard audio according to a first sound characteristic of a designated user and exercise content corresponding to a historical exercise audio; and comparing the first standard audio with the historical practice audio to obtain the pronunciation information to be corrected of the appointed user. By implementing the method, the first standard audio is obtained based on the first sound characteristic of the specified user, and the accuracy of the pronunciation information to be corrected can be improved.
103. And acquiring the oral practice content according to the pronunciation information to be corrected.
Optionally, in this embodiment of the present application, the obtaining of the spoken language practice content according to the pronunciation information to be corrected includes, but is not limited to, the following implementation manners:
mode 1: acquiring grade information of a designated user; and acquiring oral practice contents according to the grade information and the pronunciation information to be corrected.
Mode 2: acquiring grade information of a designated user; acquiring preset exercise time of a designated user; and acquiring spoken language practice contents according to the grade information, the pronunciation information to be corrected and the preset practice duration.
By implementing the method, the difficulty level of the spoken language practice content can be matched with the appointed user, quantitative recommendation can be performed on the appointed user according to the preset practice duration, and the use experience of the user can be enhanced.
After step 103, the following steps may also be performed:
determining standard audio corresponding to the spoken language practice content according to the first sound characteristic of the specified user; and playing standard audio corresponding to the spoken language practice content.
Optionally, after determining the standard audio corresponding to the spoken language practice content according to the first sound feature of the specified user, the noise index of the current environment may also be detected; when the noise index of the current environment is larger than the preset noise index, controlling a loudspeaker of the electronic equipment to start a directional play mode; playing standard audio corresponding to the spoken language practice content may include: and playing standard audio corresponding to the spoken language practice content in the directional mode. By implementing the method, the playing mode of the loudspeaker of the electronic equipment is adjusted according to the noise index of the current environment, and the standard audio corresponding to the spoken language practice content can be effectively played.
For example, when the noise index of the current environment is greater than the preset noise index, it may be further detected whether the current electric quantity of the electronic device is greater than an electric quantity threshold, and if the current electric quantity is greater than the electric quantity threshold, a speaker of the electronic device is controlled to start a directional play mode; if the electric quantity is smaller than or equal to the electric quantity threshold value, whether the electronic equipment is in communication connection with the earphone is detected, and if not, prompt information for prompting a specified user to wear the earphone is output. By implementing the method, when the standard audio corresponding to the spoken language practice content is played, the current electric quantity of the electronic equipment is taken as one of the consideration factors, so that the occurrence probability of influencing the playing of the standard audio due to the problem of the electric quantity can be reduced.
Optionally, after determining a standard audio corresponding to the spoken language practice content according to the first sound feature of the specified user, the current practice audio of the specified user for the spoken language practice content may also be acquired; and analyzing the current practice audio according to the standard audio corresponding to the spoken language practice content to obtain a scoring result of the current practice audio. By implementing the method, the objective evaluation of the current practice audio of the specified user is realized according to the standard audio corresponding to the spoken language practice content.
Optionally, when the electronic device is a family education machine, after the standard audio corresponding to the spoken language practice content is determined according to the first sound feature of the specified user, the obtained spoken language practice content and the standard audio corresponding to the spoken language practice content can be sent to the wearable device of the specified user, and based on the portability feature of the wearable device, the spoken language practice requirement of the specified user at any time and any place can be met.
Optionally, according to the first sound characteristic of the specified user, after the standard audio corresponding to the spoken language practice content is determined, whether the electronic device is connected with a sound box or not can be detected, if the electronic device is connected with the sound box, the standard audio corresponding to the spoken language practice content is sent to the sound box, so that the standard audio is played through the sound box, and the playing effect of the standard audio can be improved.
In the embodiment of the present application, after step 103, the following steps may also be performed:
acquiring historical spoken language practice time information of a specified user;
obtaining an exercise schedule of a specified user aiming at the spoken language exercise content according to historical spoken language exercise time information of the user and the spoken language exercise content; wherein, the exercise schedule records the time information and the content information of each time of spoken language exercise. By implementing the method, the appointed user can be assisted to carry out planned spoken language practice.
By implementing the method, the spoken language practice effect can be improved, the acquisition efficiency of historical practice audio of the appointed user can be improved, the accuracy of pronunciation information to be corrected can be improved, the standard audio corresponding to spoken language practice content can be effectively played, and the spoken language practice requirement of the appointed user at any time and any place can be met.
Example two
Referring to fig. 2, fig. 2 is a schematic flow chart of another method for acquiring spoken language practice content according to an embodiment of the present application. The method for acquiring spoken language practice content shown in fig. 2 may specifically include the following steps:
for detailed descriptions of step 201 to step 203, please refer to the descriptions of step 101 to step 103 in the first embodiment, which is not described again in this embodiment. It should be noted that the spoken language practice content acquired in step 103 corresponds to a plurality of dialogs.
204. Upon detection of a sparring request, target exercise content selected by a specified user is determined from spoken exercise content.
In the embodiment of the present application, the spoken language exercise content may include exercise content from a dialog or exercise content from a non-dialog, and the embodiment of the present application is not limited.
Determining target exercise content, which is selected by a designated user, from the spoken exercise content upon detection of the sparring request may include:
determining first exercise content from the dialog from the spoken exercise content upon detecting the sparring request;
displaying the first exercise content on a display screen of the electronic device;
target exercise content selected by a specified user is determined from the first exercise content.
Wherein, the determination of the target exercise content selected by the designated user from the first exercise content includes but is not limited to the following implementation manners:
mode 1: detecting a touch track of a designated user on a display screen; acquiring first exercise content corresponding to the touch track; determining first exercise content corresponding to the touch track as target exercise content;
mode 2: detecting the checking operation of a designated user for the dominant frame on the display screen, wherein each first exercise content corresponds to one dominant frame; determining first exercise content corresponding to the selected dominant frame as target exercise content;
205. acquiring content information of a first conversation corresponding to target exercise content; the target exercise content is content information of a first character in a plurality of sparring characters of the first conversation.
Acquiring content information of a first dialog corresponding to target exercise content may include: acquiring directory information of target exercise content; wherein the directory information may include identification information of the first dialogue; acquiring content information of the first conversation according to the identification information of the first conversation; the content information of the first session may be stored in the electronic device, and may also be stored in the server, which is not limited in the embodiment of the present application.
206. And starting a dialogue mode corresponding to the first character according to the content information of the first dialogue so as to enable the appointed user to carry out spoken language dialogue.
In this embodiment, the sparring mode corresponding to the first character may be divided into a virtual sparring mode in which sparring audios of the plurality of sparring characters other than the first character are preset audios and a real sparring mode in which sparring audios of the plurality of sparring characters other than the first character are input by a sparring user of a designated user in real time. Specifically, the starting of the dialogue mode corresponding to the first character according to the content information of the first dialogue includes but is not limited to the following implementation manners:
mode 1: when the sparring mode is the virtual sparring mode, acquiring preset audio of the sparring roles except the first role according to the content information of the first dialogue; starting a first dialogue mode corresponding to a first character according to a preset audio;
mode 2: when the sparring mode is the real sparring mode, according to the second sound characteristic of the sparring user of the appointed user, determining a second role matched with the second sound characteristic from the plurality of sparring roles except the first role; acquiring content information of a second role from the content information of the first conversation; and sending the content information of the second role to the terminal equipment of the sparring user so as to start a second sparring mode corresponding to the first role.
By executing steps 204 to 206, the designated user can be made more immersive in the practice of spoken language by the sparring mode, which contributes to further improvement of the practice effect.
By implementing the method, the spoken language practice effect can be favorably improved, the acquisition efficiency of historical practice audio of the appointed user can be improved, the accuracy of pronunciation information to be corrected can be improved, the standard audio corresponding to spoken language practice content can be effectively played, the spoken language practice requirement of the appointed user at any time and any place can be met, the immersion feeling of the appointed user during spoken language practice can be stronger, and the practice effect can be further improved.
EXAMPLE III
Referring to fig. 3, fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure. As shown in fig. 3, the electronic device may include:
a collecting unit 301 for collecting historical exercise audio of a specified user.
In an embodiment of the present application, when the historical practice audio of the specified user is collected by a microphone of the electronic device, the electronic device further includes: the customization unit is used for starting a sparring customization mode before the acquisition unit acquires 301 the historical exercise audio of the specified user; and, collecting customized audio in the sparring customization mode; determining a first sound characteristic of a designated user according to the customized audio; the collecting unit 301 is specifically configured to collect historical exercise audio of a specified user according to the first sound feature.
The acquisition unit 301 is further configured to acquire a first identifier of a first terminal device when a connection request of the first terminal device is detected, when the historical exercise audio of a specified user is acquired through a sound pickup of the terminal device in communication connection with the electronic device; searching user information associated with the first identifier in an equipment-user relation table preset by the electronic equipment, wherein the equipment-user relation table stores the corresponding relation between the user information and the equipment identifier; judging whether the user of the first terminal equipment is a designated user or not according to the user information associated with the first identifier; when the user of the first terminal equipment is the designated user, establishing connection with the first terminal equipment; and receiving a first exercise audio sent by the first terminal device, wherein the first exercise audio is used as a historical exercise audio of a specified user.
A determining unit 302, configured to determine pronunciation information to be corrected of the specified user by analyzing the historical practice audio.
In this embodiment of the application, the manner that the determining unit 302 is configured to determine the pronunciation information to be corrected of the specified user by analyzing the historical practice audio may specifically be: a determining unit 302, configured to obtain exercise content corresponding to historical exercise audio; obtaining a first standard audio according to a first sound characteristic of a designated user and exercise content corresponding to a historical exercise audio; and comparing the first standard audio with the historical practice audio to obtain the pronunciation information to be corrected of the appointed user.
The obtaining unit 303 is configured to obtain the spoken language practice content according to the pronunciation information to be corrected.
An obtaining unit 303, specifically configured to obtain information of a grade of a specified user; acquiring oral practice contents according to the grade information and the pronunciation information to be corrected;
or,
an obtaining unit 303, specifically configured to obtain information of a grade of a specified user; acquiring preset exercise time of a designated user; acquiring spoken language practice content according to the grade information, the pronunciation information to be corrected and preset practice duration;
optionally, the determining unit 302 is further configured to determine a standard audio corresponding to the spoken language practice content according to the first sound feature after the obtaining unit 303 obtains the spoken language practice content according to the pronunciation information to be corrected; the electronic device further includes: and the playing unit is used for playing the standard audio.
Further optionally, the determining unit 302 is further configured to detect a noise index of the current environment after determining a standard audio corresponding to the spoken language practice content according to the first sound feature of the specified user; when the noise index of the current environment is larger than the preset noise index, controlling a loudspeaker of the electronic equipment to start a directional play mode;
the manner of the playing unit for playing the standard audio corresponding to the spoken language practice content may specifically be: the playing unit plays standard audio corresponding to the spoken language practice content in the directional mode.
For example, when the noise index of the current environment is greater than the preset noise index, the determining unit 302 may further detect whether the current electric quantity of the electronic device is greater than an electric quantity threshold, and if the current electric quantity is greater than the electric quantity threshold, control a speaker of the electronic device to start a directional play mode; if the electric quantity is smaller than or equal to the electric quantity threshold value, whether the electronic equipment is in communication connection with the earphone is detected, and if not, prompt information for prompting a specified user to wear the earphone is output.
Optionally, the acquisition unit 301 is further configured to acquire a current practice audio of the specified user for the spoken language practice content after the determination unit 302 determines the standard audio corresponding to the spoken language practice content according to the first sound feature;
the electronic device further includes: and the scoring unit is used for analyzing the current practice audio according to the standard audio to obtain a scoring result of the current practice audio.
Optionally, when the electronic device is a family education machine, the determining unit 302 is further configured to, after determining the standard audio corresponding to the spoken language practice content according to the first sound feature of the specified user, send the acquired spoken language practice content and the standard audio corresponding to the spoken language practice content to the wearable device of the specified user.
Optionally, the determining unit 302 is further configured to determine, according to a first sound characteristic of the specified user, whether the electronic device is connected to a sound box after determining the standard audio corresponding to the spoken language practice content, and if the electronic device is connected to the sound box, send the standard audio corresponding to the spoken language practice content to the sound box, so that the standard audio is played through the sound box, and a playing effect of the standard audio can be improved.
In this embodiment of the application, the obtaining unit 303 is further configured to obtain historical spoken language practice time information of a specified user after obtaining spoken language practice content according to pronunciation information to be corrected; obtaining an exercise schedule of a specified user aiming at the spoken language exercise content according to historical spoken language exercise time information of the user and the spoken language exercise content; wherein, the exercise schedule records the time information and the content information of each time of spoken language exercise.
Example four
Referring to fig. 4, fig. 4 is a schematic structural diagram of another electronic device disclosed in the embodiment of the present application. The electronic device shown in fig. 4 is optimized from the electronic device shown in fig. 3, wherein when the spoken language practice content corresponds to a plurality of dialogs, as shown in fig. 4, the electronic device may further include:
a sparring unit 304, configured to determine, when a sparring request is detected after the acquisition unit 303 acquires spoken language practice contents according to pronunciation information to be corrected, a target practice content selected by a designated user from the spoken language practice contents; acquiring content information of a first conversation corresponding to the target exercise content; the target exercise content is content information of a first character in a plurality of sparring characters of a first conversation; and starting a dialogue mode corresponding to the first character according to the content information of the first dialogue so as to enable the appointed user to carry out spoken language dialogue.
Optionally, when the sparring request is detected, the manner for determining the target exercise content selected by the designated user from the spoken language exercise contents by the sparring unit 304 may specifically be:
a sparring unit 304 for determining first exercise content from the dialog from the spoken exercise content when a sparring request is detected; displaying the first exercise content on a display screen of the electronic device; target exercise content selected by a specified user is determined from the first exercise content.
Further, the manner in which the sparring unit 304 determines the targeted exercise content selected by the designated user from the first exercise content includes, but is not limited to, the following:
a sparring unit 304 for detecting a touch trajectory of a specified user on the display screen; acquiring first exercise content corresponding to the touch track; determining first exercise content corresponding to the touch track as target exercise content;
the sparring unit 304 is used for detecting the checking operation of a designated user for the dominant frames on the display screen, wherein each first exercise content corresponds to one dominant frame; determining first exercise content corresponding to the selected dominant frame as target exercise content;
the way of acquiring the content information of the first dialog corresponding to the target exercise content by the sparring unit 304 may specifically be: a sparring unit 304 for acquiring directory information of the target exercise content; wherein the directory information may include identification information of the first dialogue; acquiring content information of the first conversation according to the identification information of the first conversation; the content information of the first session may be stored in the electronic device, and may also be stored in the server, which is not limited in the embodiment of the present application.
In the embodiment of the present application, the sparring mode corresponding to the first character can be divided into a virtual sparring mode and a real sparring mode, and the modes of the sparring unit 304 for starting the sparring mode corresponding to the first character according to the content information of the first dialog include, but are not limited to, the following modes:
a sparring unit 304, configured to obtain preset audios of the plurality of sparring characters except the first character according to content information of the first dialog when the sparring mode is the virtual sparring mode; starting a first dialogue mode corresponding to a first character according to a preset audio;
a sparring unit 304 for determining a second character matching the second sound feature from the plurality of sparring characters other than the first character according to a second sound feature of the sparring user of the designated user when the sparring mode is the real sparring mode; acquiring content information of a second role from the content information of the first conversation; and sending the content information of the second role to the terminal equipment of the sparring user so as to start a second sparring mode corresponding to the first role.
Referring to fig. 5, fig. 5 is a schematic structural diagram of an electronic device according to an embodiment of the present disclosure. As shown in fig. 5, the electronic device may include:
a memory 501 in which executable program code is stored;
a processor 502 coupled to a memory 501;
the processor 502 calls the executable program code stored in the memory 501 to execute the method for acquiring the spoken language practice content shown in any one of fig. 1 to 2.
An embodiment of the application discloses a computer-readable storage medium, which stores a computer program, wherein the computer program enables a computer to execute a method for acquiring spoken language practice content shown in any one of fig. 1 to fig. 2.
The embodiment of the application discloses a computer program product, which enables a computer to execute any one of the methods for acquiring spoken language practice contents in fig. 1-2 when the computer program product runs on the computer.
The embodiment of the application discloses an application issuing system, which is used for issuing a computer program product, wherein when the computer program product runs on a computer, the computer is enabled to execute any one of the methods for acquiring the spoken language practice content shown in the figures 1 to 2.
It will be understood by those skilled in the art that all or part of the steps in the methods of the embodiments described above may be implemented by instructions associated with a program, which may be stored in a computer-readable storage medium, where the storage medium includes Read-Only Memory (ROM), Random Access Memory (RAM), Programmable Read-Only Memory (PROM), Erasable Programmable Read-Only Memory (EPROM), One-time Programmable Read-Only Memory (OTPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), compact disc-Read-Only Memory (CD-ROM), or other Memory, magnetic disk, magnetic tape, or magnetic tape, Or any other medium which can be used to carry or store data and which can be read by a computer.
The foregoing describes in detail a method for acquiring spoken language practice content and an electronic device disclosed in the embodiments of the present application, and specific examples are applied in the present application to explain the principles and implementations of the present application, and the step numbers in the specific examples do not mean the order of execution, and the execution order of each process should be determined by its function and inherent logic, but should not be limited to the implementation process of the embodiments of the present application. The units described as separate parts may or may not be physically separate, and some or all of the units may be selected according to actual needs to achieve the purpose of the embodiment.
The character "/" herein generally indicates that the former and latter associated objects are in an "or" relationship. In the embodiments provided herein, it should be understood that "B corresponding to a" means that B is associated with a from which B can be determined. It should also be understood, however, that determining B from a does not mean determining B from a alone, but may also be determined from a and/or other information. In addition, functional units in the embodiments of the present application may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit. If the integrated unit is implemented as a software functional unit and sold or used as a stand-alone product, it may be stored in a memory accessible to a computer. Based on such understanding, the technical solution of the present application, which is a part of or contributes to the prior art in essence, or all or part of the technical solution, may be embodied in the form of a software product, stored in a memory, including several requests for causing a computer device (which may be a personal computer, a server, a network device, or the like, and may specifically be a processor in the computer device) to execute part or all of the steps of the above-described method of the embodiments of the present application.
The above description of the embodiments is only for the purpose of helping to understand the method of the present application and its core ideas; meanwhile, for a person skilled in the art, according to the idea of the present application, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present application.
Claims (10)
1. A method for acquiring spoken language practice content, the method comprising:
collecting historical exercise audio of a specified user;
determining pronunciation information to be corrected of the designated user by analyzing the historical practice audios;
and acquiring oral practice contents according to the pronunciation information to be corrected.
2. The method according to claim 1, wherein the obtaining of the spoken language practice content according to the pronunciation information to be corrected comprises:
acquiring grade information of the specified user;
and acquiring oral practice contents according to the grade information and the pronunciation information to be corrected.
3. The method of claim 1 or 2, wherein prior to said capturing historical exercise audio for a given user, the method further comprises:
starting a sparring customization mode;
collecting customized audio in the sparring customization mode;
determining a first sound characteristic of a designated user according to the customized audio;
the collecting historical exercise audio of a specified user comprises:
and acquiring historical exercise audio of the specified user according to the first sound characteristic.
4. The method according to claim 3, wherein after the obtaining of the spoken language practice content according to the pronunciation information to be corrected, the method further comprises:
determining standard audio corresponding to the spoken language practice content according to the first sound characteristic;
and playing the standard audio.
5. The method of claim 4, wherein after determining the standard audio corresponding to the spoken language practice content according to the first sound characteristic, the method further comprises:
acquiring current practice audio of the specified user for the spoken language practice content;
and analyzing the current practice audio according to the standard audio to obtain a scoring result of the current practice audio.
6. The method according to claim 1, wherein after the spoken language practice content is obtained according to the pronunciation information to be corrected when the spoken language practice content corresponds to a plurality of dialogues, the method further comprises:
determining target exercise content selected by the specified user from the spoken exercise content when a sparring request is detected;
acquiring content information of a first conversation corresponding to the target exercise content; wherein the target exercise content is content information of a first character of a plurality of sparring characters of the first conversation;
and starting a sparring mode corresponding to the first character according to the content information of the first dialogue so as to enable the appointed user to carry out spoken language sparring.
7. The method according to claim 6, wherein the starting of the first dialog mode corresponding to the first character according to the content information of the first dialog comprises:
when the sparring mode is a virtual sparring mode, acquiring preset audio of the sparring roles except the first role according to the content information of the first dialogue; starting a first dialogue mode corresponding to the first character according to the preset audio;
determining a second character matched with the second sound feature from the plurality of sparring characters except the first character according to a second sound feature of a sparring user of the specified user when the sparring mode is a real sparring mode; acquiring the content information of the second role from the content information of the first conversation; and sending the content information of the second role to the terminal equipment of the sparring user so as to start a second sparring mode corresponding to the first role.
8. An electronic device, comprising:
the acquisition unit is used for acquiring historical exercise audio of a specified user;
a determining unit, configured to determine pronunciation information to be corrected of the specified user by analyzing the historical practice audio;
and the acquisition unit is used for acquiring the oral practice content according to the pronunciation information to be corrected.
9. An electronic device, characterized in that the electronic device comprises:
a memory storing executable program code;
a processor coupled with the memory;
the processor calls the executable program code stored in the memory to execute part or all of the steps of the method according to any one of claims 1 to 5.
10. A computer-readable storage medium having stored thereon a computer program comprising instructions for carrying out some or all of the steps of the method according to any one of claims 1 to 5.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010465305.5A CN111639224A (en) | 2020-05-28 | 2020-05-28 | Method for acquiring spoken language practice content and electronic equipment |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010465305.5A CN111639224A (en) | 2020-05-28 | 2020-05-28 | Method for acquiring spoken language practice content and electronic equipment |
Publications (1)
Publication Number | Publication Date |
---|---|
CN111639224A true CN111639224A (en) | 2020-09-08 |
Family
ID=72331056
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010465305.5A Pending CN111639224A (en) | 2020-05-28 | 2020-05-28 | Method for acquiring spoken language practice content and electronic equipment |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN111639224A (en) |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW200506764A (en) * | 2003-08-05 | 2005-02-16 | Wen-Fu Peng | Interactive language learning method with speech recognition |
CN102169642A (en) * | 2011-04-06 | 2011-08-31 | 李一波 | Interactive virtual teacher system having intelligent error correction function |
US20120308968A1 (en) * | 2009-10-20 | 2012-12-06 | Voctrainer Oy | Language training apparatus, method and computer program |
CN109410664A (en) * | 2018-12-12 | 2019-03-01 | 广东小天才科技有限公司 | Pronunciation correction method and electronic equipment |
CN109829050A (en) * | 2019-01-30 | 2019-05-31 | 龙马智芯(珠海横琴)科技有限公司 | A kind of language exercise method, apparatus and system |
CN109903594A (en) * | 2019-04-02 | 2019-06-18 | 北京儒博科技有限公司 | Spoken language exercise householder method, device, equipment and storage medium |
CN110276073A (en) * | 2019-06-14 | 2019-09-24 | 南京机电职业技术学院 | A kind of interactive mode Oral English Practice bearing calibration |
CN111048109A (en) * | 2019-12-25 | 2020-04-21 | 广州酷狗计算机科技有限公司 | Acoustic feature determination method and apparatus, computer device, and storage medium |
-
2020
- 2020-05-28 CN CN202010465305.5A patent/CN111639224A/en active Pending
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
TW200506764A (en) * | 2003-08-05 | 2005-02-16 | Wen-Fu Peng | Interactive language learning method with speech recognition |
US20120308968A1 (en) * | 2009-10-20 | 2012-12-06 | Voctrainer Oy | Language training apparatus, method and computer program |
CN102169642A (en) * | 2011-04-06 | 2011-08-31 | 李一波 | Interactive virtual teacher system having intelligent error correction function |
CN109410664A (en) * | 2018-12-12 | 2019-03-01 | 广东小天才科技有限公司 | Pronunciation correction method and electronic equipment |
CN109829050A (en) * | 2019-01-30 | 2019-05-31 | 龙马智芯(珠海横琴)科技有限公司 | A kind of language exercise method, apparatus and system |
CN109903594A (en) * | 2019-04-02 | 2019-06-18 | 北京儒博科技有限公司 | Spoken language exercise householder method, device, equipment and storage medium |
CN110276073A (en) * | 2019-06-14 | 2019-09-24 | 南京机电职业技术学院 | A kind of interactive mode Oral English Practice bearing calibration |
CN111048109A (en) * | 2019-12-25 | 2020-04-21 | 广州酷狗计算机科技有限公司 | Acoustic feature determination method and apparatus, computer device, and storage medium |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
EP3611895B1 (en) | Method and device for user registration, and electronic device | |
CN102568478B (en) | Video play control method and system based on voice recognition | |
CN103679452A (en) | Payment authentication method, device thereof and system thereof | |
CN104123115A (en) | Audio information processing method and electronic device | |
CN103841272B (en) | A kind of method and device sending speech message | |
EP4002363A1 (en) | Method and apparatus for detecting an audio signal, and storage medium | |
CN111343028A (en) | Distribution network control method and device | |
CN111161746B (en) | Voiceprint registration method and system | |
CN107135452A (en) | Audiphone adaptation method and device | |
CN104851423B (en) | Sound information processing method and device | |
CN107977187B (en) | Reverberation adjusting method and electronic equipment | |
CN108322770A (en) | Video frequency program recognition methods, relevant apparatus, equipment and system | |
CN107767862B (en) | Voice data processing method, system and storage medium | |
CN112700767A (en) | Man-machine conversation interruption method and device | |
CN109271480B (en) | Voice question searching method and electronic equipment | |
CN110010135A (en) | A kind of voice-based personal identification method, device and electronic equipment | |
CN111639224A (en) | Method for acquiring spoken language practice content and electronic equipment | |
CN111986680A (en) | Method and device for evaluating spoken language of object, storage medium and electronic device | |
CN109086448B (en) | Voice question searching method based on gender characteristic information and family education equipment | |
CN115565518B (en) | Method for processing player dubbing in interactive game and related device | |
CN113270099B (en) | Intelligent voice extraction method and device, electronic equipment and storage medium | |
CN108491183B (en) | Information processing method and electronic equipment | |
CN113920996A (en) | Voice interaction processing method and device, electronic equipment and storage medium | |
CN115620713A (en) | Dialog intention recognition method, device, equipment and storage medium | |
CN111785280A (en) | Identity authentication method and device, storage medium and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
RJ01 | Rejection of invention patent application after publication |
Application publication date: 20200908 |
|
RJ01 | Rejection of invention patent application after publication |