CN110718225A - Voice control method, terminal and storage medium - Google Patents

Voice control method, terminal and storage medium Download PDF

Info

Publication number
CN110718225A
CN110718225A CN201911177576.4A CN201911177576A CN110718225A CN 110718225 A CN110718225 A CN 110718225A CN 201911177576 A CN201911177576 A CN 201911177576A CN 110718225 A CN110718225 A CN 110718225A
Authority
CN
China
Prior art keywords
voice
instruction
terminal
determining whether
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911177576.4A
Other languages
Chinese (zh)
Inventor
同超
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen Konka Electronic Technology Co Ltd
Original Assignee
Shenzhen Konka Electronic Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen Konka Electronic Technology Co Ltd filed Critical Shenzhen Konka Electronic Technology Co Ltd
Priority to CN201911177576.4A priority Critical patent/CN110718225A/en
Publication of CN110718225A publication Critical patent/CN110718225A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/24Speech recognition using non-acoustical features
    • G10L15/25Speech recognition using non-acoustical features using position of the lips, movement of the lips or face analysis
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The invention discloses a voice control method, a terminal and a storage medium, wherein the voice control method comprises the following steps: acquiring a voice instruction and an image in a preset range in front of a terminal, and determining whether the voice instruction is an effective instruction according to the characteristics of the voice instruction and the image; and when the voice command is determined to be an effective command, controlling the terminal according to the voice command. The voice control method determines whether the voice instruction sent by the user is an effective instruction or not through the voice instruction sent by the user and the image of the user, corresponding voice control is carried out on the terminal only when the voice instruction sent by the user is determined to be the effective instruction, the user does not need to execute complex operation, the convenience of using the voice control by the user is improved, and meanwhile the false response triggering probability is reduced.

Description

Voice control method, terminal and storage medium
Technical Field
The present invention relates to the field of voice control technologies, and in particular, to a voice control method, a terminal, and a storage medium.
Background
With the development of voice recognition technology, voice has been widely used for controlling various terminals, and in the process of performing voice control, how to determine whether the voice uttered by the user is used for controlling the terminal becomes a difficult problem to be solved.
In the prior art, a voice collected after a certain operation is performed by a user, for example, the user presses a button or the distance from the user to the terminal is within a preset distance, is determined as a voice for controlling the terminal. However, this voice control method requires the user to perform corresponding operations, and the operations such as pressing a button and approaching the terminal cause a user to have trouble in use, and the user preferably performs voice control without performing complicated operations. Meanwhile, after the user presses the button carelessly or the distance close to the terminal is a preset distance, all voices sent by the user can be received by the terminal and judged as the control instruction, and the triggering probability of false response is high.
Thus, there is a need for improvements and enhancements in the art.
Disclosure of Invention
The technical problem to be solved by the present invention is to provide a voice control method, a terminal and a storage medium for solving the above-mentioned drawbacks of the prior art, and to solve the problem that the voice control method in the prior art is inconvenient for a user to operate.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
a method of voice control, wherein the method comprises:
acquiring a voice instruction and an image in a preset range in front of a terminal, and determining whether the voice instruction is an effective instruction according to the characteristics of the voice instruction and the image;
and when the voice command is determined to be an effective command, controlling the terminal according to the voice command.
The voice control method, wherein the determining whether the voice command is an effective command according to the feature of the voice command and the image specifically includes:
acquiring a first face feature in the image;
determining whether the user is a registered user or not according to a first voiceprint feature and/or the first face feature corresponding to the voice instruction;
when the user is a registered user, determining the face orientation of the user according to the first facial features;
and when the face faces to the direction of the terminal, converting the voice instruction into a text, and determining whether the voice instruction is an effective instruction according to the text.
The voice control method, wherein the determining whether the user is a registered user according to the first voiceprint feature and/or the first face feature corresponding to the voice instruction specifically includes:
determining whether the first voiceprint feature matches a second prestored voiceprint feature;
and/or determining whether the first facial features are matched with second facial features stored in advance.
The voice control method, wherein the determining whether the voice command is an effective command according to the text specifically includes:
and when the voice control system of the terminal is in a sleep state, determining whether the text is consistent with a pre-stored awakening word.
The voice control method, wherein the determining whether the voice command is an effective command according to the text specifically includes:
and when the terminal is in an awakening state, determining whether the voice instruction is an effective instruction according to a pre-trained instruction tag model.
The voice control method, wherein the determining whether the voice command is an effective command according to a pre-trained command classification model specifically includes:
inputting the text into the instruction label model, and determining whether the instruction label model outputs an instruction label corresponding to the text;
the instruction label model is trained according to a data set with a plurality of texts, the data set is provided with a plurality of groups of training samples, and each group of training samples comprises texts and instruction labels corresponding to the texts.
The voice control method, wherein the controlling the terminal according to the voice instruction specifically includes:
and when the terminal is in the sleep state, controlling the terminal to be switched from the sleep state to the awakening state.
The voice control method, wherein the controlling the terminal according to the effective instruction specifically includes:
and when the terminal is in an awakening state, controlling the terminal to execute the voice instruction according to the instruction tag corresponding to the text.
A terminal, wherein the terminal comprises: the voice control device comprises a processor and a storage medium which is in communication connection with the processor, wherein the storage medium is suitable for storing a plurality of instructions, and the processor is suitable for calling the instructions in the storage medium to execute the steps of realizing the voice control method.
A storage medium, wherein the storage medium stores one or more programs, which are executable by one or more processors to implement the steps of the voice control method of any one of the above.
Has the advantages that: compared with the prior art, the voice control method, the terminal and the storage medium are provided, whether the voice instruction sent by the user is an effective instruction is determined through the voice instruction sent by the user and the image of the user, the terminal is subjected to corresponding voice control only when the voice instruction sent by the user is determined to be the effective instruction, the user does not need to execute complex operation, the convenience of using voice control by the user is improved, and meanwhile the false response triggering probability is reduced.
Drawings
FIG. 1 is a flowchart illustrating a first embodiment of a voice control method according to the present invention;
FIG. 2 is a flowchart illustrating sub-steps S100 of an embodiment of a voice control method provided in the present invention;
fig. 3 is a schematic structural diagram of a preferred embodiment of the terminal provided in the present invention.
Detailed Description
In order to make the objects, technical solutions and effects of the present invention clearer and clearer, the present invention is further described in detail below with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
Referring to fig. 1, fig. 1 is a flowchart illustrating a voice control method according to a preferred embodiment of the present invention. The method comprises the following steps:
s100, acquiring a voice instruction and an image in a preset range in front of a terminal, and determining whether the voice instruction is an effective instruction or not according to the voice instruction and the image information.
The voice instruction is a voice instruction issued by a user, and the voice instruction can be issued when the user wants to control the terminal by voice. The voice instruction can be obtained through a preset microphone, and the microphone can be installed on the terminal or can be arranged independently.
When a user controls a terminal, the user is often in front of the terminal, and therefore, in this embodiment, the image in a preset range in front of the terminal is obtained, the preset range may be specifically set according to an actual situation, for example, when the terminal is a television, the preset range may be set according to a size and an applicable space of the television, for example, the television is a large-sized television, and then the television can be applicable to a larger living room, and then the preset range may be set larger, conversely, if the television is smaller in size, then the range in which the content displayed by the television can be viewed may be smaller, and then the preset range may be set smaller, and a person skilled in the art may specifically set the preset range according to an actual situation.
The image can be acquired by a preset camera, and the camera can be installed on the terminal or can be independently set.
The microphone and the camera can be in a working state all the time, or the microphone can be in a working state all the time, and after the microphone acquires the voice command, the camera is controlled to start to collect images, so that energy consumption is reduced.
After the voice instruction and the image are obtained, determining whether the voice instruction is an effective instruction according to the feature of the voice instruction and the image, as shown in fig. 2, specifically including:
and S110, acquiring a first face feature in the image.
After the image is acquired, face recognition can be performed on the image to acquire a first face feature in the image.
And S120, determining whether the user is a registered user according to the first voiceprint feature and/or the first face feature corresponding to the voice instruction.
In this embodiment, in order to prevent a security risk brought by other people who control the terminal arbitrarily, it is set that only a registered user can control the terminal, and specifically, it is determined whether the user is a registered user according to the first voiceprint and/or the first facial feature corresponding to the voice instruction.
Specifically, the determining whether the user is a registered user according to the first voiceprint feature and/or the first facial feature corresponding to the voice instruction includes:
s121, determining whether the first voiceprint feature is matched with a second voiceprint feature stored in advance; and/or the presence of a gas in the gas,
and S122, determining whether the first face features are matched with second face features stored in advance.
The voiceprint is a sound wave frequency spectrum which is displayed by an electro-acoustic instrument and carries speech information, and the voiceprint not only has specificity, but also has the characteristic of relative stability, namely, the voiceprint characteristics of different voices emitted by the same person are consistent. In this embodiment, after the voice instruction is obtained, the voice instruction may be analyzed to obtain the first voiceprint feature corresponding to the voice instruction, and whether the user is a registered user is determined according to whether the first voiceprint feature is consistent with a voiceprint feature of the registered user. Specifically, the terminal may collect and store a second voiceprint feature of a registered user in advance, match the first voiceprint feature with the second voiceprint feature after the terminal acquires the first voiceprint feature, and determine that the user is a registered user if the first voiceprint feature can be matched with the second voiceprint feature. Of course, it may be understood that one terminal may have more than one registered user, that is, there may be a plurality of second voiceprint features, and after the first voiceprint feature is obtained, as long as the first voiceprint feature can be matched with one of the second voiceprint features, the user is determined to be a registered user.
The first facial feature in the image may also be used to determine whether the user is a registered user, and specifically, in this embodiment, after the image is obtained, the image is analyzed to extract the first facial feature in the image, and it is easy to see that the first facial feature is a facial feature of the user who is using the terminal. Determining whether the user is a registered user by judging whether the first facial features are consistent with the facial features of the registered user. Specifically, the terminal may collect and store a second face feature of a registered user in advance, match the first face feature with the second face feature after the terminal acquires the first face feature, and determine that the user is a registered user if the first face feature and the second face feature can be matched. Similarly, there may be a plurality of second facial features, and as described above for the second voiceprint feature, details are not repeated here.
It should be noted that, in the step S111, determining whether the first voiceprint feature matches with the second prestored voiceprint feature and the step S112, determining whether the first face feature matches with the second prestored face feature may be performed simultaneously, or may be performed separately, that is, a relationship between the two is and/or. When the first voiceprint feature and the first face feature are executed simultaneously, namely, whether the user is a registered user is determined simultaneously according to the first voiceprint feature and the first face feature, the identification accuracy of the registered user can be improved.
If the user is determined to be the registered user, the next operation can be executed, and if the user is determined not to be the registered user, the current process is ended, and the next voice command and image are waited to be obtained.
Referring to fig. 2 again, in the present embodiment, after determining whether the user is a registered user, the method includes:
and S130, when the user is a registered user, determining the face orientation of the user according to the first face features.
In the embodiment, in order to facilitate the user to use the voice control terminal, whether the voice command sent by the user is sent to the terminal is determined by the face orientation of the user. Specifically, the determining of the face orientation of the user according to the first face feature may be determined by judging whether the first face feature is a front face, or may be determined by extracting an eye part in the first face feature and determining whether the implementation is looking at the terminal. Of course, the present invention is not limited to the above-mentioned exemplary methods, and those skilled in the art may select different methods for determining the face orientation according to actual situations.
S140, when the face faces to the direction of the terminal, the voice instruction is converted into a text, and whether the voice instruction is an effective instruction or not is determined according to the text.
Specifically, in the prior art, a plurality of terminals are provided with a plurality of turns of conversation functions, once the plurality of turns of conversation functions are started, all voices sent by a user are collected and judged as instructions, at this time, the voices sent by the user and not directed at the terminal cause wrong responses of the terminal, the triggering rate of the wrong responses is high, the user cannot perform other language behaviors during the use of the terminal, and experience is poor. In this embodiment, if the face is oriented toward the terminal, it is determined that the voice instruction sent by the user is sent by the terminal, at this time, the voice instruction is converted into a text, and the semantic instruction is converted into the text by using an existing voice recognition technology, such as an asr (automatic speech recognition) technology, and the like, which is not described herein any more, after the text is obtained, it is determined whether the voice instruction is an effective instruction according to the text, the terminal performs a corresponding operation when the voice instruction is an effective instruction, and the user can perform voice control on the terminal without a complicated operation, and meanwhile, triggering false response is also effectively avoided. And if the face direction is not the direction towards the terminal, ending the current process and waiting for acquiring a new voice command and an image.
Specifically, the determining whether the voice instruction is a valid instruction according to the text includes:
when the terminal is in a sleep state, determining whether the text is consistent with a pre-stored awakening word;
and when the terminal is in an awakening state, determining whether the voice instruction is an effective instruction according to a pre-trained instruction tag model.
Specifically, in order to save energy consumption, most terminals are provided with a sleep mode, and when the terminals are not operated for a period of time, the terminals enter a sleep state, where the sleep state refers to a state in which the terminals are in a standby state with low energy consumption, at this time, most functions of the terminals are turned off, only some systems such as a voice receiving system and the like are in operation, and are ready to receive a voice instruction, at this time, the terminals need to be awakened by a specific awakening word, and when the terminals are in the awakening state, corresponding operations can be executed according to the voice instruction.
In this embodiment, when the terminal is in a sleep state, only the determination of whether the voice instruction is the wake-up instruction is performed, at this time, the text is matched with a pre-stored wake-up word, if the text is consistent with the wake-up word, the voice instruction is an effective instruction, a wake-up operation can be performed on the terminal, and if the text is inconsistent with the wake-up word, the terminal is maintained in the sleep state. It can be seen that, with the voice control method provided in this embodiment, only when a registered user looks at the terminal and speaks a preset wake-up word, the terminal can be woken up, so that the terminal is prevented from being woken up by mistake when the user speaks the wake-up word when the user does not wake up the terminal by accident.
And when the terminal is in an awakening state, determining whether the voice instruction is an effective instruction according to a pre-trained instruction label model. Specifically, the instruction label model is trained according to a data set with a plurality of texts, the data set has a plurality of groups of training samples, and each group of training samples includes a text and an instruction label corresponding to the text. The instruction tag corresponds to a specific category of the voice instruction corresponding to the text, for example: the terminal comprises a local instruction (increasing volume, reducing brightness and the like) for local operation of the terminal, a video type instruction (i want to watch XX movie, open XX TV play and the like) for the terminal to operate through cloud service, a music type (i want to listen to XXX song, search XX album) and the like. That is, the instruction tag model actually has a function of classifying the voice instruction, and the instruction tag model can be realized by a Natural Language Processing (NLP) technique.
Specifically, the determining whether the voice command is a valid command according to a pre-trained command label model includes: and inputting the text into the instruction label model, and determining whether the instruction label model outputs an instruction label corresponding to the text.
After the text corresponding to the voice command is acquired, the text is input into the command classification model, and as can be seen from the foregoing, the command label model has a function of classifying the text, and when the command label model can output the command label corresponding to the text, it indicates that the voice command can be recognized as a specific category, and the terminal can execute a corresponding operation according to the voice command. And when the instruction tag model cannot output the instruction tag corresponding to the text, the voice instruction cannot be identified as a specific category, the terminal cannot execute corresponding operation according to the voice instruction, and the voice instruction is an invalid instruction.
And when the voice command is an invalid command, the terminal does not respond to the voice command, ends the current flow and waits for acquiring a new voice command and a new head portrait.
Referring to fig. 1 again, the voice control method further includes:
and S200, controlling the terminal according to the voice command when the voice command is determined to be an effective command.
When the terminal is in a sleep state, the controlling the terminal according to the voice instruction specifically includes: and controlling the terminal to be switched from a sleep state to an awakening state.
As already explained above, when the terminal is in the sleep state and the text of the voice command is consistent with the wakeup word, it is determined that the voice command is an effective command, and at this time, the terminal is woken up to be switched from the sleep state to the wakeup state.
When the terminal is in the wake-up state, the controlling the terminal according to the voice instruction specifically includes: and controlling the terminal to execute the voice instruction according to the instruction label corresponding to the text.
Specifically, when the voice control system of the terminal is in an awake state, if the instruction tag model can output the instruction tag corresponding to the text according to the text corresponding to the voice instruction, it is determined that the voice instruction is an effective instruction. And after the instruction tag corresponding to the text is obtained, controlling the terminal to execute the voice instruction according to the instruction tag. For example, when the voice command is: and when the instruction label is 'video type', controlling the terminal to search the XX in the video type resource.
In summary, the present embodiment provides a voice control method, which determines whether a user is a registered user through voiceprint matching and/or face feature matching, determines whether a face of the user is towards a terminal through face features, determines whether a received voice instruction is an effective instruction through steps of determining an instruction tag through a natural language processing model, and the like, controls the terminal to execute a corresponding action only when the received voice instruction is the effective instruction, and the user only needs to look at the terminal to speak the voice instruction without other operations, thereby not only facilitating the user to use a voice control system, but also reducing a false response rate of the terminal to voice.
It should be understood that, although the steps in the flowcharts shown in the figures of the present specification are shown in order as indicated by the arrows, the steps are not necessarily performed in order as indicated by the arrows. The steps are not performed in the exact order shown and described, and may be performed in other orders, unless explicitly stated otherwise. Moreover, at least a portion of the steps in the flowchart may include multiple sub-steps or multiple stages, which are not necessarily performed at the same time, but may be performed at different times, and the order of performing the sub-steps or stages is not necessarily sequential, but may be performed alternately or alternately with other steps or at least a portion of the sub-steps or stages of other steps.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, databases, or other media used in embodiments provided herein may include non-volatile and/or volatile memory. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
Example two
Based on the above embodiments, the present invention further provides a terminal, and a schematic block diagram thereof may be as shown in fig. 3. The terminal comprises a processor, a memory, a network interface, a display screen and a temperature sensor which are connected through a system bus. Wherein the processor of the terminal is configured to provide computing and control capabilities. The memory of the terminal comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the terminal is used for connecting and communicating with an external terminal through a network. The computer program is executed by a processor to implement a speech control method. The display screen of the terminal can be a liquid crystal display screen or an electronic ink display screen, and the temperature sensor of the terminal is arranged in the terminal in advance and used for detecting the current operating temperature of internal equipment.
It will be understood by those skilled in the art that the block diagram shown in fig. 3 is a block diagram of only a portion of the structure associated with the inventive arrangements and is not intended to limit the terminals to which the inventive arrangements may be applied, and that a particular terminal may include more or less components than those shown, or may have some components combined, or may have a different arrangement of components.
In one embodiment, a terminal is provided, comprising a memory and a processor, the memory having a computer program stored therein, the processor when executing the computer program implementing at least the following steps:
acquiring a voice instruction and an image in a preset range in front of a terminal, and determining whether the voice instruction is an effective instruction according to the characteristics of the voice instruction and the image;
and when the voice command is determined to be an effective command, controlling the terminal according to the voice command.
Wherein, the determining whether the voice instruction is an effective instruction according to the feature of the voice instruction and the image specifically includes:
acquiring a first face feature in the image;
determining whether the user is a registered user or not according to a first voiceprint feature and/or the first face feature corresponding to the voice instruction;
when the user is a registered user, determining the face orientation of the user according to the first facial features;
and when the face faces to the direction of the terminal, converting the voice instruction into a text, and determining whether the voice instruction is an effective instruction according to the text.
Wherein the determining whether the user is a registered user according to the first voiceprint feature and/or the first facial feature corresponding to the voice instruction specifically includes:
determining whether the first voiceprint feature matches a second prestored voiceprint feature;
and/or determining whether the first facial features are matched with second facial features stored in advance.
Wherein the determining whether the voice instruction is an effective instruction according to the text specifically includes:
and when the terminal is in a sleep state, determining whether the text is consistent with a pre-stored awakening word.
Wherein the determining whether the voice instruction is an effective instruction according to the text specifically includes:
and when the terminal is in an awakening state, determining whether the voice instruction is an effective instruction according to a pre-trained instruction tag model.
Wherein the determining whether the voice instruction is an effective instruction according to the pre-trained instruction classification model specifically comprises:
inputting the text into the instruction label model, and determining whether the instruction label model outputs an instruction label corresponding to the text;
the instruction label model is trained according to a data set with a plurality of texts, the data set is provided with a plurality of groups of training samples, and each group of training samples comprises texts and instruction labels corresponding to the texts.
The voice control method, wherein the controlling the terminal according to the voice instruction specifically includes:
and when the terminal is in the sleep state, controlling the terminal to be switched from the sleep state to the awakening state.
Wherein the controlling the terminal according to the voice instruction specifically includes:
and when the terminal is in an awakening state, controlling the terminal to execute the voice instruction according to the instruction tag corresponding to the text.
EXAMPLE III
The present invention also provides a storage medium storing one or more programs executable by one or more processors to implement the steps of the voice control method described in the above embodiments.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (10)

1. A voice control method, characterized in that the voice control method comprises:
acquiring a voice instruction and an image in a preset range in front of a terminal, and determining whether the voice instruction is an effective instruction according to the characteristics of the voice instruction and the image;
and when the voice command is determined to be an effective command, controlling the terminal according to the voice command.
2. The voice control method according to claim 1, wherein the determining whether the voice command is a valid command according to the feature of the voice command and the image specifically comprises:
acquiring a first face feature in the image;
determining whether the user is a registered user or not according to a first voiceprint feature and/or the first face feature corresponding to the voice instruction;
when the user is a registered user, determining the face orientation of the user according to the first facial features;
and when the face faces to the direction of the terminal, converting the voice instruction into a text, and determining whether the voice instruction is an effective instruction according to the text.
3. The voice control method according to claim 2, wherein the determining whether the user is a registered user according to the first voiceprint feature and/or the first facial feature corresponding to the voice instruction specifically includes:
determining whether the first voiceprint feature matches a second prestored voiceprint feature;
and/or determining whether the first facial features are matched with second facial features stored in advance.
4. The voice control method according to claim 2, wherein the determining whether the voice command is a valid command according to the text specifically includes:
and when the terminal is in a sleep state, determining whether the text is consistent with a pre-stored awakening word.
5. The voice control method according to claim 2, wherein the determining whether the voice command is a valid command according to the text specifically includes:
and when the terminal is in an awakening state, determining whether the voice instruction is an effective instruction according to a pre-trained instruction tag model.
6. The method according to claim 5, wherein the determining whether the voice command is a valid command according to a pre-trained command classification model specifically comprises:
inputting the text into the instruction label model, and determining whether the instruction label model outputs an instruction label corresponding to the text;
the instruction label model is trained according to a data set with a plurality of texts, the data set is provided with a plurality of groups of training samples, and each group of training samples comprises texts and instruction labels corresponding to the texts.
7. The voice control method according to claim 4, wherein the controlling the terminal according to the voice instruction specifically comprises:
and when the terminal is in the sleep state, controlling the terminal to be switched from the sleep state to the awakening state.
8. The voice control method according to claim 6, wherein the controlling the terminal according to the voice instruction specifically comprises:
and when the terminal is in an awakening state, controlling the terminal to execute the voice instruction according to the instruction tag corresponding to the text.
9. A terminal, characterized in that the terminal comprises: a processor, a storage medium communicatively coupled to the processor, the storage medium adapted to store a plurality of instructions, the processor adapted to invoke the instructions in the storage medium to perform the steps of implementing the voice control method of any of claims 1-8.
10. A storage medium storing one or more programs, the one or more programs being executable by one or more processors to implement the steps of the voice control method of any one of claims 1-8.
CN201911177576.4A 2019-11-25 2019-11-25 Voice control method, terminal and storage medium Pending CN110718225A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911177576.4A CN110718225A (en) 2019-11-25 2019-11-25 Voice control method, terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911177576.4A CN110718225A (en) 2019-11-25 2019-11-25 Voice control method, terminal and storage medium

Publications (1)

Publication Number Publication Date
CN110718225A true CN110718225A (en) 2020-01-21

Family

ID=69216511

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911177576.4A Pending CN110718225A (en) 2019-11-25 2019-11-25 Voice control method, terminal and storage medium

Country Status (1)

Country Link
CN (1) CN110718225A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111783892A (en) * 2020-07-06 2020-10-16 广东工业大学 Robot instruction identification method and device, electronic equipment and storage medium
CN111968633A (en) * 2020-07-16 2020-11-20 深圳市鸿合创新信息技术有限责任公司 Management control method and device of interactive equipment
CN113539252A (en) * 2020-04-22 2021-10-22 庄连豪 Barrier-free intelligent voice system and control method thereof

Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103456299A (en) * 2013-08-01 2013-12-18 百度在线网络技术(北京)有限公司 Method and device for controlling speech recognition
CN106843882A (en) * 2017-01-20 2017-06-13 联想(北京)有限公司 A kind of information processing method, device and information processing system
CN206805695U (en) * 2017-05-27 2017-12-26 杨倩旖 A kind of intelligence reads financial special purpose device
CN107665708A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Intelligent sound exchange method and system
CN107679506A (en) * 2017-10-12 2018-02-09 Tcl通力电子(惠州)有限公司 Awakening method, intelligent artifact and the computer-readable recording medium of intelligent artifact
CN108469966A (en) * 2018-03-21 2018-08-31 北京金山安全软件有限公司 Voice broadcast control method and device, intelligent device and medium
CN108538298A (en) * 2018-04-04 2018-09-14 科大讯飞股份有限公司 voice awakening method and device
CN109166575A (en) * 2018-07-27 2019-01-08 百度在线网络技术(北京)有限公司 Exchange method, device, smart machine and the storage medium of smart machine
CN109192204A (en) * 2018-08-31 2019-01-11 广东小天才科技有限公司 A kind of sound control method and smart machine based on smart machine camera
CN109710080A (en) * 2019-01-25 2019-05-03 华为技术有限公司 A kind of screen control and sound control method and electronic equipment
US20190138330A1 (en) * 2017-11-08 2019-05-09 Alibaba Group Holding Limited Task Processing Method and Device
CN110335609A (en) * 2019-06-26 2019-10-15 四川大学 A kind of air-ground communicating data analysis method and system based on speech recognition
CN110335600A (en) * 2019-07-09 2019-10-15 四川长虹电器股份有限公司 The multi-modal exchange method and system of household appliance

Patent Citations (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN103456299A (en) * 2013-08-01 2013-12-18 百度在线网络技术(北京)有限公司 Method and device for controlling speech recognition
CN107665708A (en) * 2016-07-29 2018-02-06 科大讯飞股份有限公司 Intelligent sound exchange method and system
CN106843882A (en) * 2017-01-20 2017-06-13 联想(北京)有限公司 A kind of information processing method, device and information processing system
CN206805695U (en) * 2017-05-27 2017-12-26 杨倩旖 A kind of intelligence reads financial special purpose device
CN107679506A (en) * 2017-10-12 2018-02-09 Tcl通力电子(惠州)有限公司 Awakening method, intelligent artifact and the computer-readable recording medium of intelligent artifact
US20190138330A1 (en) * 2017-11-08 2019-05-09 Alibaba Group Holding Limited Task Processing Method and Device
CN108469966A (en) * 2018-03-21 2018-08-31 北京金山安全软件有限公司 Voice broadcast control method and device, intelligent device and medium
CN108538298A (en) * 2018-04-04 2018-09-14 科大讯飞股份有限公司 voice awakening method and device
CN109166575A (en) * 2018-07-27 2019-01-08 百度在线网络技术(北京)有限公司 Exchange method, device, smart machine and the storage medium of smart machine
CN109192204A (en) * 2018-08-31 2019-01-11 广东小天才科技有限公司 A kind of sound control method and smart machine based on smart machine camera
CN109710080A (en) * 2019-01-25 2019-05-03 华为技术有限公司 A kind of screen control and sound control method and electronic equipment
CN110335609A (en) * 2019-06-26 2019-10-15 四川大学 A kind of air-ground communicating data analysis method and system based on speech recognition
CN110335600A (en) * 2019-07-09 2019-10-15 四川长虹电器股份有限公司 The multi-modal exchange method and system of household appliance

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
王文 等: ""人-服务机器人交互中自然语言理解研究"", 《微型电脑应用》 *

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113539252A (en) * 2020-04-22 2021-10-22 庄连豪 Barrier-free intelligent voice system and control method thereof
CN111783892A (en) * 2020-07-06 2020-10-16 广东工业大学 Robot instruction identification method and device, electronic equipment and storage medium
CN111783892B (en) * 2020-07-06 2021-10-01 广东工业大学 Robot instruction identification method and device, electronic equipment and storage medium
CN111968633A (en) * 2020-07-16 2020-11-20 深圳市鸿合创新信息技术有限责任公司 Management control method and device of interactive equipment
WO2022011965A1 (en) * 2020-07-16 2022-01-20 深圳市鸿合创新信息技术有限责任公司 Management control method and apparatus for interaction device

Similar Documents

Publication Publication Date Title
KR102293063B1 (en) Customizable wake-up voice commands
US11289100B2 (en) Selective enrollment with an automated assistant
US20200302913A1 (en) Electronic device and method of controlling speech recognition by electronic device
CN113327609B (en) Method and apparatus for speech recognition
CN110718225A (en) Voice control method, terminal and storage medium
CN111045639A (en) Voice input method, device, electronic equipment and storage medium
CN108154878A (en) Control the method and device of monitoring device
CN111326154B (en) Voice interaction method and device, storage medium and electronic equipment
US20200380971A1 (en) Method of activating voice assistant and electronic device with voice assistant
CN111210829A (en) Speech recognition method, apparatus, system, device and computer readable storage medium
CN109785834B (en) Voice data sample acquisition system and method based on verification code
KR20200007530A (en) Method for processing user voice input and electronic device supporting the same
CN108595406B (en) User state reminding method and device, electronic equipment and storage medium
CN112700782A (en) Voice processing method and electronic equipment
CN111784971B (en) Alarm processing method and system, computer readable storage medium and electronic device
CN112420044A (en) Voice recognition method, voice recognition device and electronic equipment
CN111862943A (en) Speech recognition method and apparatus, electronic device, and storage medium
CN110782886A (en) System, method, television, device and medium for speech processing
CN113611316A (en) Man-machine interaction method, device, equipment and storage medium
WO2020073839A1 (en) Voice wake-up method, apparatus and system, and electronic device
CN110400568B (en) Awakening method of intelligent voice system, intelligent voice system and vehicle
CN114999496A (en) Audio transmission method, control equipment and terminal equipment
CN113870857A (en) Voice control scene method and voice control scene system
CN112151028A (en) Voice recognition method and device
CN112885341A (en) Voice wake-up method and device, electronic equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200121

RJ01 Rejection of invention patent application after publication