CN109460209B - Control method for dictation and reading progress and electronic equipment - Google Patents

Control method for dictation and reading progress and electronic equipment Download PDF

Info

Publication number
CN109460209B
CN109460209B CN201811560158.9A CN201811560158A CN109460209B CN 109460209 B CN109460209 B CN 109460209B CN 201811560158 A CN201811560158 A CN 201811560158A CN 109460209 B CN109460209 B CN 109460209B
Authority
CN
China
Prior art keywords
dictation
audio
time point
user
playing time
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201811560158.9A
Other languages
Chinese (zh)
Other versions
CN109460209A (en
Inventor
韦肖莹
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guangdong Genius Technology Co Ltd
Original Assignee
Guangdong Genius Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guangdong Genius Technology Co Ltd filed Critical Guangdong Genius Technology Co Ltd
Priority to CN201811560158.9A priority Critical patent/CN109460209B/en
Publication of CN109460209A publication Critical patent/CN109460209A/en
Application granted granted Critical
Publication of CN109460209B publication Critical patent/CN109460209B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F3/00Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
    • G06F3/16Sound input; Sound output
    • G06F3/167Audio in a user interface, e.g. using voice commands for navigating, audio feedback
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G11INFORMATION STORAGE
    • G11BINFORMATION STORAGE BASED ON RELATIVE MOVEMENT BETWEEN RECORD CARRIER AND TRANSDUCER
    • G11B27/00Editing; Indexing; Addressing; Timing or synchronising; Monitoring; Measuring tape travel
    • G11B27/10Indexing; Addressing; Timing or synchronising; Measuring tape travel
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The embodiment of the invention relates to the technical field of voice control, and discloses a control method for dictation and reading progress and electronic equipment, wherein the method comprises the following steps: when a first voice instruction input by a user is detected, if the user intention is recognized as the adjustment of the reading progress, extracting search content from the first voice instruction; intercepting a plurality of short sentence materials containing search contents from the dictation materials, and numbering the short sentence materials; circularly reading each short sentence material and the audio number thereof until a target audio number input by the user through a second voice instruction is detected; performing text positioning on a target short sentence material corresponding to a target audio number in the dictation material to obtain a first text positioning position; and determining a first playing time point corresponding to the first text positioning position in the dictation audio, and starting playing the dictation audio from the first playing time point. By implementing the embodiment of the invention, the accuracy of dictation broadcasting progress control can be improved.

Description

Control method for dictation and reading progress and electronic equipment
Technical Field
The invention relates to the technical field of voice control, in particular to a control method of dictation and reading progress and electronic equipment.
Background
Dictation is a common teaching mode when learning human language. Currently, most people mainly use players provided by electronic devices (such as smart phones, notebook computers, tablet computers, and the like) to play dictation audios and write down characters reported and read in the dictation audios to perform dictation exercises. In practice, it is found that when a dictating person wants to listen to a certain piece of content in the dictation audio again, the player progress bar needs to be dragged manually to adjust the position of the player progress bar to a playing time point corresponding to the piece of content, and the playing is resumed. Therefore, the mode makes it difficult for the user to manually adjust the position of the progress bar of the player to a certain precise playing time point, and further the progress of the dictation broadcast cannot be accurately controlled.
Disclosure of Invention
The embodiment of the invention discloses a control method of dictation broadcast progress and electronic equipment, which can improve the accuracy of dictation broadcast progress control.
The first aspect of the embodiments of the present invention discloses a method for controlling a dictation and reading progress, where the method includes:
when a first voice instruction input by a user is detected, identifying user intention of the first voice instruction;
if the user intention of the first voice instruction is recognized to be the adjustment of the reading progress, extracting search content from the first voice instruction;
according to the search content, intercepting a plurality of short sentence materials containing the search content from dictation materials corresponding to dictation audio, and numbering the short sentence materials to obtain an audio number of each short sentence material in the short sentence materials;
circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until a target audio number input by a user through a second voice instruction is detected; the target audio number is the audio number of any one of the plurality of short sentence materials;
according to the target audio number, performing text positioning on a target short sentence material corresponding to the target audio number in the dictation material to obtain a first text positioning position;
and determining a first playing time point corresponding to the first text positioning position in the dictation audio, and starting to play the dictation audio from the first playing time point.
As an optional implementation manner, in the first aspect of the embodiment of the present invention, after the recognizing a user intention of the first voice instruction when the first voice instruction input by the user is detected, the method further includes:
if the user intention of the first voice instruction is identified as a reading progress mark, recording the current playing time point of the dictation audio;
determining a second text positioning position corresponding to the current playing time point in the dictation material;
extracting a complete sentence where the second text positioning position is located from the dictation material, and positioning a third text positioning position where the first character of the complete sentence is located;
determining a second playing time point corresponding to the third text positioning position in the dictation audio;
and when the dictation audio is detected to be played completely, the dictation audio is played from the second playing time point again.
As an optional implementation manner, in the first aspect of the embodiment of the present invention, after the beginning of playing the dictation audio from the first playing time point, the method further includes:
when a second voice instruction input by a user is detected, recognizing the user intention of the second voice instruction;
if the user intention of the second voice instruction is to circularly play, pausing the playing of the dictation audio, and recording the pause playing time point of the dictation audio;
and intercepting an audio clip corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio, and playing the audio clip.
As an optional implementation manner, in the first aspect of this embodiment of the present invention, the method further includes:
and acquiring the marking information input by the user when marking the first playing time point or the second playing time point, and displaying the marking information at the progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
As an optional implementation manner, in the first aspect of this embodiment of the present invention, the method further includes:
and attaching a label corresponding to the dictation audio to the labeling information, storing the labeling information after attaching the label corresponding to the dictation audio to a preset database, deriving the labeling information from the preset database according to the label corresponding to the dictation audio and displaying the labeling information at a progress bar corresponding to the dictation audio when the dictation audio is played each time.
A second aspect of an embodiment of the present invention discloses an electronic device, including:
the first recognition module is used for recognizing the user intention of a first voice instruction when the first voice instruction input by a user is detected;
the extraction module is used for extracting search contents from the first voice instruction when the user intention of the first voice instruction is recognized to be the adjustment of the reading progress;
the intercepting module is used for intercepting a plurality of short sentence materials containing the search content from dictation materials corresponding to dictation audio according to the search content and numbering the short sentence materials to obtain the audio number of each short sentence material in the short sentence materials;
the reading module is used for circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until a target audio number input by a user through a second voice instruction is detected; the target audio number is the audio number of any one of the plurality of short sentence materials;
the first positioning module is used for performing text positioning on a target short sentence material corresponding to the target audio number in the dictation material according to the target audio number so as to obtain a first text positioning position;
and the first control module is used for determining a first playing time point corresponding to the first text positioning position in the dictation audio and starting playing the dictation audio from the first playing time point.
As an optional implementation manner, in the second aspect of the embodiment of the present invention, the electronic device further includes:
the first recording module is used for recording the current playing time point of the dictation audio when the user intention of the first voice instruction is recognized as a reading degree mark after the first recognition module recognizes the user intention of the first voice instruction;
the second positioning module is used for determining a second text positioning position corresponding to the current playing time point in the dictation material;
the third positioning module is used for extracting the complete sentence where the second text positioning position is located from the dictation material and positioning a third text positioning position where the first character of the complete sentence is located;
and the second control module is used for determining a second playing time point corresponding to the third text positioning position in the dictation audio and playing the dictation audio from the second playing time point again when the dictation audio is detected to be played completely.
As an optional implementation manner, in the second aspect of the embodiment of the present invention, the electronic device further includes:
the second recognition module is used for recognizing the user intention of a second voice instruction when the second voice instruction input by the user is detected after the first control module starts to play the dictation audio from the first playing time point;
the second recording module is used for pausing the playing of the dictation audio and recording the pause playing time point of the dictation audio when the user intention of the second voice instruction is to circularly play;
and the third control module is used for intercepting an audio clip corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio and playing the audio clip.
As an optional implementation manner, in the second aspect of the embodiment of the present invention, the electronic device further includes:
and the marking module is used for acquiring marking information input by a user when marking the first playing time point or the second playing time point, and displaying the marking information at a progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
As an optional implementation manner, in the second aspect of the embodiment of the present invention, the electronic device further includes:
and the storage module is used for attaching a label corresponding to the dictation audio to the labeling information, storing the labeling information after the label corresponding to the dictation audio is attached to a preset database, deriving the labeling information from the preset database according to the label corresponding to the dictation audio and displaying the labeling information at a progress bar corresponding to the dictation audio when the dictation audio is played each time.
A third aspect of an embodiment of the present invention discloses an electronic device, including:
a memory storing executable program code;
a processor coupled with the memory;
the processor calls the executable program code stored in the memory to execute the control method for dictation, report and read progress disclosed in the first aspect of the embodiment of the invention.
A fourth aspect of the embodiments of the present invention discloses a computer-readable storage medium, which stores a computer program, where the computer program enables a computer to execute the method for controlling a dictation and reading progress disclosed in the first aspect of the embodiments of the present invention.
A fifth aspect of embodiments of the present invention discloses a computer program product, which, when run on a computer, causes the computer to perform some or all of the steps of any one of the methods of the first aspect.
A sixth aspect of the present embodiment discloses an application publishing platform, where the application publishing platform is configured to publish a computer program product, where the computer program product is configured to, when running on a computer, cause the computer to perform part or all of the steps of any one of the methods in the first aspect.
Compared with the prior art, the embodiment of the invention has the following beneficial effects:
by implementing the embodiment of the invention, when the user intention indicated by the voice instruction input by the user is identified as the adjustment of the reading progress, the search content is extracted from the voice instruction, a plurality of short sentence materials containing the search content can be intercepted from the dictation material corresponding to the dictation audio according to the search content, and the user does not need to manually drag the progress bar of the player to find out the audio content to be played again from the dictation audio, so that the unnecessary search time is greatly reduced, and the efficiency of dictation exercise is improved; in addition, according to the target short sentence material selected by the user from the plurality of short sentence materials through voice, the target short sentence material is automatically subjected to text positioning in the dictation material, the dictation audio is played from the playing time point corresponding to the text positioning position, the position of the progress bar of the player does not need to be manually adjusted to the accurate playing time point by the user, and the accuracy of dictation broadcasting progress control can be improved.
Drawings
In order to more clearly illustrate the technical solutions in the embodiments of the present invention, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and it is obvious for those skilled in the art that other drawings can be obtained according to these drawings without creative efforts.
Fig. 1 is a schematic flow chart of a method for controlling dictation and reading progress according to an embodiment of the present invention;
fig. 2 is a schematic flow chart of another method for controlling dictation and newspaper reading progress disclosed in the embodiment of the present invention;
fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure;
FIG. 4 is a schematic structural diagram of another electronic device disclosed in the embodiments of the present invention;
fig. 5 is a schematic structural diagram of another electronic device disclosed in the embodiment of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments. All other embodiments, which can be derived by a person skilled in the art from the embodiments given herein without making any creative effort, shall fall within the protection scope of the present invention.
It should be noted that the terms "first", "second", "third", "fourth", and the like in the description and the claims of the present invention are used for distinguishing different objects, and are not used for describing a specific order. The terms "comprises," "comprising," and any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, system, article, or apparatus that comprises a list of steps or elements is not necessarily limited to those steps or elements expressly listed, but may include other steps or elements not expressly listed or inherent to such process, method, article, or apparatus.
The embodiment of the invention discloses a control method of dictation broadcast progress and electronic equipment, which can improve the accuracy of dictation broadcast progress control. The following detailed description is made with reference to the accompanying drawings.
Example one
Referring to fig. 1, fig. 1 is a schematic flow chart illustrating a method for controlling a dictation and reading progress according to an embodiment of the present invention. As shown in fig. 1, the dictation and reading progress control method may include the following steps.
101. When a first voice instruction input by a user is detected, the user intention of the first voice instruction is identified.
The method for controlling the progress of dictation and reading disclosed by the embodiment of the invention is suitable for smart phones (such as Android phones, iOS phones and the like), tablet computers, palm computers, UMPCs (Ultra-mobile Personal computers), netbooks, mobile internet equipment, PADs and other electronic equipment comprising touch screens, and is particularly suitable for Web applications, APPs or special software in the electronic equipment.
102. And if the user intention of the first voice instruction is recognized to be the adjustment of the reading progress, extracting the search content from the first voice instruction.
As an optional implementation manner, if the user intention of the voice instruction cannot be recognized, a function prompt list may be output, where the function prompt list includes control instructions for triggering different functions, for example, adjustment of the reading progress, a reading progress flag, and the like, and can guide the user to input a corresponding control instruction according to the intention by voice.
103. And according to the search content, intercepting a plurality of short sentence materials containing the search content from the dictation materials corresponding to the dictation audio, and numbering the short sentence materials to obtain the audio number of each short sentence material in the short sentence materials.
In the embodiment of the present invention, the dictation material may be a chinese material, an english material, or a material of other specified languages, which is not limited herein. The specific content of the dictation material may be words, sentences or articles, etc., and is not limited thereto.
104. And circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until the target audio number input by the user through the second voice instruction is detected.
In the embodiment of the invention, the target audio number is the audio number of any one of the short sentence materials.
105. And according to the target audio number, performing text positioning on the target short sentence material corresponding to the target audio number in the dictation material to obtain a first text positioning position.
106. And determining a first playing time point corresponding to the first text positioning position in the dictation audio, and starting playing the dictation audio from the first playing time point.
In the embodiment of the invention, optionally, the intention keywords included in the voice instruction can be extracted by performing syllable segmentation, stress judgment, grammar analysis and semantic analysis on the voice instruction input by the user; and executing a corresponding control instruction according to the preset command keyword when the intention keyword is judged to be matched with the preset command keyword. For example, when listening to a lesson, "Guilin mountain and Water", the user inputs a voice command "replay" the mountain of the Guilin "or" listen again "the mountain of the Guilin", and the keyword "replay" or "listen again" is extracted from the voice command, indicating that the user intends to make an adjustment for the progress of reading. Furthermore, search contents can be extracted from the voice instruction, and a plurality of short sentence materials containing the search contents are reported and read for the user to select. In combination with the voice instruction "replay the mountain of the Guilin"; if three short sentence materials are intercepted from the text (dictation materials) according to the search content and respectively correspond to different audio labels, the reading content is as follows: segment 1, mountain luqio of Guilin, standing up one seat, "segment 2, mountain luxio of Guilin, like a green barrier," and mountain luzhen of segment 3 Guilin, dangerous peak. If the user inputs a voice instruction of selecting the segment 2, text positioning is carried out on a sentence "mountain reality show of Guilin, like a dark green barrier" in a text, a corresponding playing time point in the dictation audio is obtained according to the position of the text positioning, and the dictation audio is played again from the playing time point.
It can be seen that, with the method described in fig. 1, when it is recognized that the user intention indicated by the voice instruction input by the user is adjustment of the reading progress, by extracting the search content from the voice instruction, several short sentence materials containing the search content can be intercepted from the dictation material corresponding to the dictation audio according to the search content, and the user does not need to manually drag the progress bar of the player to find out the audio content to be played again from the dictation audio, thereby greatly reducing unnecessary search time and improving the efficiency of dictation practice; in addition, according to the target short sentence material selected by the user from the plurality of short sentence materials through voice, the target short sentence material is automatically subjected to text positioning in the dictation material, the dictation audio is played from the playing time point corresponding to the text positioning position, the position of the progress bar of the player does not need to be manually adjusted to the accurate playing time point by the user, and the accuracy of dictation broadcasting progress control can be improved.
Example two
Referring to fig. 2, fig. 2 is a schematic flow chart illustrating another method for controlling a dictation and reading progress according to an embodiment of the present invention. As shown in fig. 2, the dictation and reading progress control method may include the following steps.
201. When a first voice instruction input by a user is detected, the user intention of the first voice instruction is identified.
202. And if the user intention of the first voice instruction is recognized to be the adjustment of the reading progress, extracting the search content from the first voice instruction.
As an optional implementation manner, after step 201, the present scheme may further include:
if the user intention of the first voice instruction is identified as a reading progress mark, recording the current playing time point of the dictation audio;
determining a second text positioning position corresponding to the current playing time point in the dictation material;
extracting a complete sentence where the second text positioning position is located from the dictation material, and positioning a third text positioning position where the first character of the complete sentence is located;
determining a second playing time point corresponding to the third text positioning position in the dictation audio;
and when the fact that the dictation audio is played is detected to be finished, the dictation audio is played from the second playing time point again.
Therefore, by implementing the optional implementation mode, the user can be helped to review the dictation contents from the marked playing time point again after the dictation audio is completely played by reading and marking the specific playing time point in the dictation process so as to realize self correction and defect and leakage detection; and the complete sentence with the marked playing time point is positioned, and the dictation audio is played back from the playing time point corresponding to the first character of the complete sentence, so that the completeness of sentences in the dictation content can be ensured, and the user experience is improved.
Further, as an optional implementation, the present solution may further include:
and acquiring the marking information input by the user when marking the first playing time point or the second playing time point, and displaying the marking information at the progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
Furthermore, a label corresponding to the dictation audio can be attached to the labeling information, and the labeling information with the label corresponding to the dictation audio attached thereto is stored in a preset database, so that the labeling information is derived from the preset database according to the label corresponding to the dictation audio and is displayed on a progress bar corresponding to the dictation audio each time the dictation audio is played.
Therefore, the marking information input by the user to the adjusted playing time point is displayed at the progress bar, the user can be helped to subsequently playback any marked playing time point, unnecessary adjusting steps are reduced, the marking information can be stored in the preset database, when the dictation audio is called out next time, the marking information is simultaneously led out for the user to review the historical difficult and complicated content in the dictation material, and the learning effect of the user is favorably improved.
203. And according to the search content, intercepting a plurality of short sentence materials containing the search content from the dictation materials corresponding to the dictation audio, and numbering the short sentence materials to obtain the audio number of each short sentence material in the short sentence materials.
204. And circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until the target audio number input by the user through the second voice instruction is detected.
In the embodiment of the invention, the target audio number is the audio number of any one of the short sentence materials.
205. And according to the target audio number, performing text positioning on the target short sentence material corresponding to the target audio number in the dictation material to obtain a first text positioning position.
206. And determining a first playing time point corresponding to the first text positioning position in the dictation audio, and starting playing the dictation audio from the first playing time point.
As an alternative embodiment, before the dictation audio is played from the first play time point, the voice may output a speed-adjustment query message to query the user whether to lower the current play speed; if a confirmation instruction input by the user voice is received, reducing the current playing multiple speed to a specified low multiple speed, and starting to play the dictation audio from a first playing time point at the specified low multiple speed; otherwise, the current playing speed is kept, and the dictation audio is played from the first playing time point at the current playing speed.
Furthermore, the number of clicks of the user on the marking information at the progress bar can be counted; when detecting that a user clicks the marking information, if the accumulated clicking times of the marking information exceed the preset times, reducing the current playing multiple speed to the appointed low multiple speed, and playing the dictation audio from the playing time point corresponding to the marking information at the appointed low multiple speed.
207. And when a second voice instruction input by the user is detected, recognizing the user intention of the second voice instruction.
208. And if the user intention of the second voice instruction is to circularly play, pausing the playing of the dictation audio and recording the pause playing time point of the dictation audio.
209. And intercepting an audio clip corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio, and playing the audio clip.
It can be seen that, step 207 to step 209 are implemented, and according to the control instruction input by the user voice, the specified audio segment can be intercepted from the dictation audio, so as to narrow the scope of the dictation content, and enable the user to perform dictation exercise on the specified audio segment in a targeted manner.
As an optional implementation manner, the dictation answering content of the dictation material of the user can be received, and the dictation answering content of the user is modified according to the dictation material to obtain a modification result; extracting error contents of the user according to the correction result; acquiring a correct text corresponding to the error content of the user, and performing text positioning on the correct text in the dictation material to acquire a text positioning position corresponding to the correct text; finding out a playing starting time point and a playing ending time point corresponding to the correct text from the dictation audio according to the text positioning position of the correct text so as to intercept an audio clip corresponding to the correct text; and pushing the audio clip corresponding to the correct text to the user so that the user can autonomously correct the error content in the dictation answering content according to the audio clip corresponding to the correct text. Optionally, the handwritten dictation answering content of the user may be scanned by using a camera of the electronic device, and answering text information of the user is extracted from the dictation answering content, so as to modify the answering text information of the user and obtain a modification result. Therefore, according to the error text in the dictation answering content of the user, the audio clip corresponding to the correct text for error correction is automatically pushed to the user, and the error correction efficiency can be improved.
It can be seen that, with the implementation of the method described in fig. 2, when it is recognized that the user intention indicated by the voice instruction input by the user is adjustment of the reading progress, by extracting the search content from the voice instruction, several short sentence materials containing the search content can be intercepted from the dictation material corresponding to the dictation audio according to the search content, and the user does not need to manually drag the progress bar of the player to find out the audio content to be played again from the dictation audio, thereby greatly reducing unnecessary search time and improving the efficiency of dictation practice; in addition, according to the target short sentence material selected by the user from the plurality of short sentence materials through voice, the target short sentence material is automatically subjected to text positioning in the dictation material, and dictation audio is played from the playing time point corresponding to the text positioning position, the position of the progress bar of the player does not need to be manually adjusted to the accurate playing time point by the user, and the accuracy of dictation broadcasting progress control can be improved; in addition, by reading and marking the specific playing time point in the dictation process, the user can be helped to review the dictation content from the marked playing time point again after the dictation audio is completely played, so that self correction and defect and omission mending are realized; in addition, by positioning the complete sentence where the marked playing time point is located and starting to replay the dictation audio from the playing time point corresponding to the first character of the complete sentence, the completeness of sentences in the dictation content can be ensured, and the user experience is improved; in addition, the marking information input by the user to the adjusted playing time point is displayed at the progress bar, the user can be helped to subsequently play back any marked playing time point, unnecessary adjusting steps are reduced, the marking information can be stored in a preset database, and when the dictation audio is called out next time, the marking information is simultaneously led out for the user to review historical difficult and complicated contents in the dictation material, so that the learning effect of the user is favorably improved; in addition, according to the control instruction input by the voice of the user, the appointed audio frequency segment can be intercepted from the dictation audio frequency, so that the scope of the dictation content is narrowed, and the user can carry out dictation exercise on the appointed audio frequency segment in a targeted manner.
EXAMPLE III
Referring to fig. 3, fig. 3 is a schematic structural diagram of an electronic device according to an embodiment of the disclosure. As shown in fig. 3, the electronic device may include:
the first recognition module 301 is configured to, when a first voice instruction input by a user is detected, recognize a user intention of the first voice instruction.
As an optional implementation manner, the electronic device is further configured to output a function prompt list when the user intention of the voice instruction cannot be recognized, where the function prompt list includes control instructions triggering different functions, such as reading progress adjustment, reading progress flag, and the like, and is capable of guiding the user to input a corresponding control instruction according to the intention by voice.
The extracting module 302 is configured to extract the search content from the first voice instruction when it is recognized that the user intent of the first voice instruction is adjustment of the reading progress.
The intercepting module 303 is configured to intercept, according to the search content, a plurality of short sentence materials containing the search content from the dictation material corresponding to the dictation audio, and number the plurality of short sentence materials to obtain an audio number of each of the plurality of short sentence materials.
A reading module 304, configured to read the plurality of phrase materials and the audio number of each of the plurality of phrase materials in a circulating manner until a target audio number input by the user through the second voice instruction is detected; the target audio number is the audio number of any one of the short sentence materials.
The first positioning module 305 is configured to perform text positioning on a target short sentence material corresponding to a target audio number in the dictation material according to the target audio number, so as to obtain a first text positioning position.
The first control module 306 is configured to determine a first playing time point corresponding to the first text positioning position in the dictation audio, and start playing the dictation audio from the first playing time point.
It can be seen that, with the electronic device described in fig. 3, when it is recognized that the user intention indicated by the voice instruction input by the user is adjustment of the reading progress, by extracting the search content from the voice instruction, several short sentence materials containing the search content can be intercepted from the dictation material corresponding to the dictation audio according to the search content, and the user does not need to manually drag the progress bar of the player to find out the audio content to be played again from the dictation audio, thereby greatly reducing unnecessary search time and improving the efficiency of dictation practice; in addition, according to the target short sentence material selected by the user from the plurality of short sentence materials through voice, the target short sentence material is automatically subjected to text positioning in the dictation material, the dictation audio is played from the playing time point corresponding to the text positioning position, the position of the progress bar of the player does not need to be manually adjusted to the accurate playing time point by the user, and the accuracy of dictation broadcasting progress control can be improved.
Example four
Referring to fig. 4, fig. 4 is a schematic structural diagram of another electronic device according to an embodiment of the disclosure. The electronic device shown in fig. 4 is optimized from the electronic device shown in fig. 3. Compared to the electronic device shown in fig. 3, the electronic device shown in fig. 4 may further include:
a second recognition module 307, configured to recognize a user intention of the second voice instruction when the second voice instruction input by the user is detected after the first control module 306 starts playing the dictation audio from the first playing time point.
And the second recording module 308 is configured to pause playing of the dictation audio and record a pause playing time point of the dictation audio when the user of the second voice instruction intends to perform loop playing.
The third control module 309 is configured to intercept an audio segment corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio, and play the audio segment.
As an optional implementation, the electronic device may further include:
the first recording module 310 is configured to record a current playing time point of the dictation audio when the user intention of the first voice instruction is recognized as the reading-in degree mark after the first recognition module 301 recognizes the user intention of the first voice instruction.
And the second positioning module 311 is configured to determine a second text positioning position corresponding to the current playing time point in the dictation material.
The third positioning module 312 is configured to extract the complete sentence where the second text positioning position is located from the dictation material, and position a third text positioning position where the first character of the complete sentence is located.
The second control module 313 is configured to determine a second playing time point corresponding to the third text positioning position in the dictation audio, and start playing the dictation audio from the second playing time point again when it is detected that the dictation audio is completely played.
Further, as an optional implementation, the electronic device may further include:
and the marking module is used for acquiring marking information input by a user when marking the first playing time point or the second playing time point, and displaying the marking information at a progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
Still further, the electronic device further includes:
and the storage module is used for attaching a label corresponding to the dictation audio to the labeling information, storing the labeling information to which the label corresponding to the dictation audio is attached to a preset database, deriving the labeling information from the preset database according to the label corresponding to the dictation audio when the dictation audio is played every time, and displaying the labeling information on a progress bar corresponding to the dictation audio.
As an optional implementation manner, the electronic device may be further configured to, before the first control module 306 starts playing the dictation audio from the first playing time point, output a speed-doubling adjustment query message by voice to query the user whether to decrease the current playing speed-doubling; if a confirmation instruction input by the user voice is received, reducing the current playing multiple speed to a specified low multiple speed, and starting to play the dictation audio from a first playing time point at the specified low multiple speed; otherwise, the current playing speed is kept, and the dictation audio is played from the first playing time point at the current playing speed.
Furthermore, the number of clicks of the user on the marking information at the progress bar can be counted; when detecting that a user clicks the marking information, if the accumulated clicking times of the marking information exceed the preset times, reducing the current playing multiple speed to the appointed low multiple speed, and playing the dictation audio from the playing time point corresponding to the marking information at the appointed low multiple speed.
As an optional implementation manner, the electronic device may be further configured to receive dictation response content of the user for the dictation material, and modify the dictation response content of the user according to the dictation material to obtain a modification result; extracting error contents of the user according to the correction result; acquiring a correct text corresponding to the error content of the user, and performing text positioning on the correct text in the dictation material to acquire a text positioning position corresponding to the correct text; finding out a playing starting time point and a playing ending time point corresponding to the correct text from the dictation audio according to the text positioning position of the correct text so as to intercept an audio clip corresponding to the correct text; and pushing the audio clip corresponding to the correct text to the user so that the user can autonomously correct the error content in the dictation answering content according to the audio clip corresponding to the correct text. Optionally, the handwritten dictation answering content of the user may be scanned by using a camera of the electronic device, and answering text information of the user is extracted from the dictation answering content, so as to modify the answering text information of the user and obtain a modification result. Therefore, according to the error text in the dictation answering content of the user, the audio clip corresponding to the correct text for error correction is automatically pushed to the user, and the error correction efficiency can be improved.
It can be seen that, with the electronic device described in fig. 4, when it is recognized that the user intention indicated by the voice instruction input by the user is adjustment of the reading progress, by extracting the search content from the voice instruction, several short sentence materials containing the search content can be intercepted from the dictation material corresponding to the dictation audio according to the search content, and the user does not need to manually drag the progress bar of the player to find out the audio content to be played again from the dictation audio, thereby greatly reducing unnecessary search time and improving the efficiency of dictation practice; in addition, according to the target short sentence material selected by the user from the plurality of short sentence materials through voice, the target short sentence material is automatically subjected to text positioning in the dictation material, and dictation audio is played from the playing time point corresponding to the text positioning position, the position of the progress bar of the player does not need to be manually adjusted to the accurate playing time point by the user, and the accuracy of dictation broadcasting progress control can be improved; in addition, by reading and marking the specific playing time point in the dictation process, the user can be helped to review the dictation content from the marked playing time point again after the dictation audio is completely played, so that self correction and defect and omission mending are realized; in addition, by positioning the complete sentence where the marked playing time point is located and starting to replay the dictation audio from the playing time point corresponding to the first character of the complete sentence, the completeness of sentences in the dictation content can be ensured, and the user experience is improved; in addition, the marking information input by the user to the adjusted playing time point is displayed at the progress bar, the user can be helped to subsequently play back any marked playing time point, unnecessary adjusting steps are reduced, the marking information can be stored in a preset database, and when the dictation audio is called out next time, the marking information is simultaneously led out for the user to review historical difficult and complicated contents in the dictation material, so that the learning effect of the user is favorably improved; in addition, according to the control instruction input by the voice of the user, the appointed audio frequency segment can be intercepted from the dictation audio frequency, so that the scope of the dictation content is narrowed, and the user can carry out dictation exercise on the appointed audio frequency segment in a targeted manner.
EXAMPLE five
Referring to fig. 5, fig. 5 is a schematic structural diagram of another electronic device according to an embodiment of the disclosure. As shown in fig. 5, the electronic device may include:
a memory 501 in which executable program code is stored;
a processor 502 coupled to a memory 501;
the processor 502 calls the executable program code stored in the memory 501 to execute any one of the methods for controlling the progress of dictation and reading shown in fig. 1 to 2.
The embodiment of the invention discloses a computer-readable storage medium which stores a computer program, wherein the computer program enables a computer to execute any one control method of dictation and reading progress in figures 1-2.
Embodiments of the present invention also disclose a computer program product, wherein, when the computer program product is run on a computer, the computer is caused to execute part or all of the steps of the method as in the above method embodiments.
The embodiment of the present invention also discloses an application publishing platform, which is used for publishing a computer program product, wherein when the computer program product runs on a computer, the computer is caused to execute part or all of the steps of the method in the above method embodiments.
It will be understood by those skilled in the art that all or part of the steps in the methods of the embodiments described above may be implemented by hardware instructions of a program, and the program may be stored in a computer-readable storage medium, where the storage medium includes Read-Only Memory (ROM), Random Access Memory (RAM), Programmable Read-Only Memory (PROM), Erasable Programmable Read-Only Memory (EPROM), One-time Programmable Read-Only Memory (OTPROM), Electrically Erasable Programmable Read-Only Memory (EEPROM), Compact Disc Read-Only Memory (CD-ROM), or other Memory, such as a magnetic disk, or a combination thereof, A tape memory, or any other medium readable by a computer that can be used to carry or store data.
The method for controlling dictation and reading progress and the electronic device disclosed by the embodiment of the invention are described in detail, a specific example is applied to explain the principle and the implementation mode of the invention, and the description of the embodiment is only used for helping to understand the method and the core idea of the invention; meanwhile, for a person skilled in the art, according to the idea of the present invention, there may be variations in the specific embodiments and the application scope, and in summary, the content of the present specification should not be construed as a limitation to the present invention.

Claims (10)

1. A control method for dictation and reading progress is characterized by comprising the following steps:
when a first voice instruction input by a user is detected, identifying user intention of the first voice instruction;
if the user intention of the first voice instruction is recognized to be the adjustment of the reading progress, extracting search content from the first voice instruction;
according to the search content, intercepting a plurality of short sentence materials containing the search content from dictation materials corresponding to dictation audio, and numbering the short sentence materials to obtain an audio number of each short sentence material in the short sentence materials;
circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until a target audio number input by a user through a second voice instruction is detected; the target audio number is the audio number of any one of the plurality of short sentence materials;
according to the target audio number, performing text positioning on a target short sentence material corresponding to the target audio number in the dictation material to obtain a first text positioning position;
determining a first playing time point corresponding to the first text positioning position in the dictation audio, and outputting speed-doubling adjusting inquiry information by voice to inquire whether a user reduces the current playing speed; if a confirmation instruction input by a user voice is received, reducing the current playing multiple speed to a specified low multiple speed, and starting to play the dictation audio from the first playing time point at the specified low multiple speed; otherwise, the current playing double speed is kept and the dictation audio is played from the first playing time point.
2. The method of claim 1, wherein after identifying the user intent of the first voice instruction upon detecting the first voice instruction input by the user, the method further comprises:
if the user intention of the first voice instruction is identified as a reading progress mark, recording the current playing time point of the dictation audio;
determining a second text positioning position corresponding to the current playing time point in the dictation material;
extracting a complete sentence where the second text positioning position is located from the dictation material, and positioning a third text positioning position where the first character of the complete sentence is located;
determining a second playing time point corresponding to the third text positioning position in the dictation audio;
and when the dictation audio is detected to be played completely, the dictation audio is played from the second playing time point again.
3. The method of claim 1, wherein after the beginning of the playing of the dictation audio from the first play time point, the method further comprises:
when a second voice instruction input by a user is detected, recognizing the user intention of the second voice instruction;
if the user intention of the second voice instruction is to circularly play, pausing the playing of the dictation audio, and recording the pause playing time point of the dictation audio;
and intercepting an audio clip corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio, and playing the audio clip.
4. The method of claim 2, further comprising:
and acquiring the marking information input by the user when marking the first playing time point or the second playing time point, and displaying the marking information at the progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
5. The method of claim 4, further comprising:
and attaching a label corresponding to the dictation audio to the labeling information, storing the labeling information after attaching the label corresponding to the dictation audio to a preset database, deriving the labeling information from the preset database according to the label corresponding to the dictation audio and displaying the labeling information at a progress bar corresponding to the dictation audio when the dictation audio is played each time.
6. An electronic device, characterized in that the electronic device comprises:
the first recognition module is used for recognizing the user intention of a first voice instruction when the first voice instruction input by a user is detected;
the extraction module is used for extracting search contents from the first voice instruction when the user intention of the first voice instruction is recognized to be the adjustment of the reading progress;
the intercepting module is used for intercepting a plurality of short sentence materials containing the search content from dictation materials corresponding to dictation audio according to the search content and numbering the short sentence materials to obtain the audio number of each short sentence material in the short sentence materials;
the reading module is used for circularly reading the plurality of short sentence materials and the audio number of each short sentence material in the plurality of short sentence materials until a target audio number input by a user through a second voice instruction is detected; the target audio number is the audio number of any one of the plurality of short sentence materials;
the first positioning module is used for performing text positioning on a target short sentence material corresponding to the target audio number in the dictation material according to the target audio number so as to obtain a first text positioning position;
the first control module is used for determining a first playing time point corresponding to the first text positioning position in the dictation audio, and outputting speed-doubling adjustment inquiry information by voice to inquire whether a user reduces the current playing speed-doubling; if a confirmation instruction input by a user voice is received, reducing the current playing multiple speed to a specified low multiple speed, and starting to play the dictation audio from the first playing time point at the specified low multiple speed; otherwise, the current playing double speed is kept and the dictation audio is played from the first playing time point.
7. The electronic device of claim 6, further comprising:
the first recording module is used for recording the current playing time point of the dictation audio when the user intention of the first voice instruction is recognized as a reading degree mark after the first recognition module recognizes the user intention of the first voice instruction;
the second positioning module is used for determining a second text positioning position corresponding to the current playing time point in the dictation material;
the third positioning module is used for extracting the complete sentence where the second text positioning position is located from the dictation material and positioning a third text positioning position where the first character of the complete sentence is located;
and the second control module is used for determining a second playing time point corresponding to the third text positioning position in the dictation audio and playing the dictation audio from the second playing time point again when the dictation audio is detected to be played completely.
8. The electronic device of claim 6, further comprising:
the second recognition module is used for recognizing the user intention of a second voice instruction when the second voice instruction input by the user is detected after the first control module starts to play the dictation audio from the first playing time point;
the second recording module is used for pausing the playing of the dictation audio and recording the pause playing time point of the dictation audio when the user intention of the second voice instruction is to circularly play;
and the third control module is used for intercepting an audio clip corresponding to the playing time period from the first playing time point to the pause playing time point from the dictation audio and playing the audio clip.
9. The electronic device of claim 7, further comprising:
and the marking module is used for acquiring marking information input by a user when marking the first playing time point or the second playing time point, and displaying the marking information at a progress bar corresponding to the dictation audio so as to play the dictation audio from the playing time point corresponding to the marking information when detecting that the user clicks the marking information.
10. The electronic device of claim 9, further comprising:
and the storage module is used for attaching a label corresponding to the dictation audio to the labeling information, storing the labeling information after the label corresponding to the dictation audio is attached to a preset database, deriving the labeling information from the preset database according to the label corresponding to the dictation audio and displaying the labeling information at a progress bar corresponding to the dictation audio when the dictation audio is played each time.
CN201811560158.9A 2018-12-20 2018-12-20 Control method for dictation and reading progress and electronic equipment Active CN109460209B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201811560158.9A CN109460209B (en) 2018-12-20 2018-12-20 Control method for dictation and reading progress and electronic equipment

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201811560158.9A CN109460209B (en) 2018-12-20 2018-12-20 Control method for dictation and reading progress and electronic equipment

Publications (2)

Publication Number Publication Date
CN109460209A CN109460209A (en) 2019-03-12
CN109460209B true CN109460209B (en) 2022-03-01

Family

ID=65613919

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201811560158.9A Active CN109460209B (en) 2018-12-20 2018-12-20 Control method for dictation and reading progress and electronic equipment

Country Status (1)

Country Link
CN (1) CN109460209B (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111026786B (en) * 2019-04-11 2023-06-27 广东小天才科技有限公司 Dictation list generation method and home education equipment
CN109887349B (en) * 2019-04-12 2021-05-11 广东小天才科技有限公司 Dictation auxiliary method and device
CN111833846B (en) * 2019-04-12 2023-06-02 广东小天才科技有限公司 Method and device for starting dictation state according to intention, and storage medium
CN111046220A (en) * 2019-04-29 2020-04-21 广东小天才科技有限公司 Method for replaying reading voice in dictation process and electronic equipment
CN111081103B (en) * 2019-05-17 2022-03-01 广东小天才科技有限公司 Dictation answer obtaining method, family education equipment and storage medium
CN112013294B (en) * 2019-05-29 2022-06-21 广东小天才科技有限公司 Intelligent dictation table lamp and dictation assisting method thereof
CN111079500B (en) * 2019-07-11 2023-10-27 广东小天才科技有限公司 Method and system for correcting dictation content
CN111081082B (en) * 2019-07-11 2022-04-29 广东小天才科技有限公司 Dictation intelligent control method based on user intention and electronic equipment
CN111079423A (en) * 2019-08-02 2020-04-28 广东小天才科技有限公司 Method for generating dictation, reading and reporting audio, electronic equipment and storage medium
CN111077982B (en) * 2019-08-02 2023-11-24 广东小天才科技有限公司 Man-machine interaction method under dictation environment and electronic equipment
CN111899575A (en) * 2020-07-21 2020-11-06 北京字节跳动网络技术有限公司 Dictation content distribution method, dictation content distribution device, dictation content distribution equipment and storage medium
CN112712806A (en) * 2020-12-31 2021-04-27 南方科技大学 Auxiliary reading method and device for visually impaired people, mobile terminal and storage medium
CN113194380B (en) * 2021-04-26 2022-08-23 读书郎教育科技有限公司 Control system and method for dictation progress of English new words
CN113488045A (en) * 2021-07-01 2021-10-08 读书郎教育科技有限公司 System and method for controlling dictation content
CN113568529A (en) * 2021-07-13 2021-10-29 读书郎教育科技有限公司 Intelligent desk lamp and method for realizing dictation broadcasting progress control based on infrared technology
CN116847164B (en) * 2023-07-04 2024-02-13 广东景行教育科技有限公司 Course content feedback system and method based on big data

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101796516A (en) * 2007-09-04 2010-08-04 苹果公司 navigation systems and methods
CN101996195A (en) * 2009-08-28 2011-03-30 中国移动通信集团公司 Searching method and device of voice information in audio files and equipment
CN103226950A (en) * 2012-01-29 2013-07-31 特克特朗尼克公司 Speech processing in telecommunication network
CN103841465A (en) * 2012-11-28 2014-06-04 上海斐讯数据通信技术有限公司 Intelligent terminal and system and method for controlling playback progress thereof
CN104219628A (en) * 2014-08-29 2014-12-17 南京信息职业技术学院 Blind person information service method and blind person information service system based on enhancement of combination of reality with smart phone
CN104765821A (en) * 2015-04-07 2015-07-08 合肥芯动微电子技术有限公司 Voice frequency ordering method and device
CN107919127A (en) * 2017-11-27 2018-04-17 北京地平线机器人技术研发有限公司 Method of speech processing, device and electronic equipment
CN108959282A (en) * 2017-05-17 2018-12-07 北京博瑞彤芸文化传播股份有限公司 A kind of data query method

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8825482B2 (en) * 2005-09-15 2014-09-02 Sony Computer Entertainment Inc. Audio, video, simulation, and user interface paradigms

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN101796516A (en) * 2007-09-04 2010-08-04 苹果公司 navigation systems and methods
CN101996195A (en) * 2009-08-28 2011-03-30 中国移动通信集团公司 Searching method and device of voice information in audio files and equipment
CN103226950A (en) * 2012-01-29 2013-07-31 特克特朗尼克公司 Speech processing in telecommunication network
CN103841465A (en) * 2012-11-28 2014-06-04 上海斐讯数据通信技术有限公司 Intelligent terminal and system and method for controlling playback progress thereof
CN104219628A (en) * 2014-08-29 2014-12-17 南京信息职业技术学院 Blind person information service method and blind person information service system based on enhancement of combination of reality with smart phone
CN104765821A (en) * 2015-04-07 2015-07-08 合肥芯动微电子技术有限公司 Voice frequency ordering method and device
CN108959282A (en) * 2017-05-17 2018-12-07 北京博瑞彤芸文化传播股份有限公司 A kind of data query method
CN107919127A (en) * 2017-11-27 2018-04-17 北京地平线机器人技术研发有限公司 Method of speech processing, device and electronic equipment

Also Published As

Publication number Publication date
CN109460209A (en) 2019-03-12

Similar Documents

Publication Publication Date Title
CN109460209B (en) Control method for dictation and reading progress and electronic equipment
CN109346059B (en) Dialect voice recognition method and electronic equipment
CN109635096B (en) Dictation prompting method and electronic equipment
CN106710592B (en) Voice recognition error correction method and device in intelligent hardware equipment
US9158753B2 (en) Data processing method, presentation method, and corresponding apparatuses
CN102568478B (en) Video play control method and system based on voice recognition
CN109960809B (en) Dictation content generation method and electronic equipment
CN108986564B (en) Reading control method based on intelligent interaction and electronic equipment
US9818450B2 (en) System and method of subtitling by dividing script text into two languages
CN111081084B (en) Method for broadcasting dictation content and electronic equipment
CN109165336B (en) Information output control method and family education equipment
CN111081080B (en) Voice detection method and learning device
CN111077996A (en) Information recommendation method based on point reading and learning equipment
US20220269724A1 (en) Audio playing method, electronic device, and storage medium
CN111026786A (en) Dictation list generation method and family education equipment
CN111723235A (en) Music content identification method, device and equipment
CN111079501B (en) Character recognition method and electronic equipment
CN111079504A (en) Character recognition method and electronic equipment
CN111081227B (en) Recognition method of dictation content and electronic equipment
CN111026839B (en) Method for detecting mastering degree of dictation word and electronic equipment
US10970910B2 (en) Animation of concepts in printed materials
CN111091008A (en) Method for determining dictation prompt content and electronic equipment
CN111079486A (en) Method for starting dictation detection and electronic equipment
CN111755026B (en) Voice recognition method and system
CN111028591A (en) Dictation control method and learning equipment

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant