CN110706705A - Voice control method, terminal and computer storage medium - Google Patents

Voice control method, terminal and computer storage medium Download PDF

Info

Publication number
CN110706705A
CN110706705A CN201911007522.3A CN201911007522A CN110706705A CN 110706705 A CN110706705 A CN 110706705A CN 201911007522 A CN201911007522 A CN 201911007522A CN 110706705 A CN110706705 A CN 110706705A
Authority
CN
China
Prior art keywords
terminal
trigger condition
control
control information
voice data
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201911007522.3A
Other languages
Chinese (zh)
Inventor
李凯
崔薇
赵云
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hisense Mobile Communications Technology Co Ltd
Original Assignee
Hisense Mobile Communications Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hisense Mobile Communications Technology Co Ltd filed Critical Hisense Mobile Communications Technology Co Ltd
Priority to CN201911007522.3A priority Critical patent/CN110706705A/en
Publication of CN110706705A publication Critical patent/CN110706705A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The invention relates to the field of computers, in particular to a voice control method, a terminal and a computer storage medium, which solve the problem that the voice control method is not flexible enough when the terminal is required to execute control operation at present, the terminal of the embodiment of the invention collects voice data input by a user and used for carrying out delay control on the terminal through a microphone, after identifying a trigger condition contained in the voice data and determining control information used for controlling the terminal according to the voice data, stores the corresponding relation between the trigger condition and the control information into an operation set to be executed, after determining that the trigger condition in the operation set to be executed is satisfied, the terminal acquires the control information corresponding to the satisfied trigger condition from the operation set to be executed and executes the control operation corresponding to the control information, thereby solving the problem that the current voice control method can only execute the operation corresponding to the control instruction immediately, but not yet flexible enough.

Description

Voice control method, terminal and computer storage medium
Technical Field
The present invention relates to the field of computers, and in particular, to a voice control method, a terminal, and a computer storage medium.
Background
With the development of the voice recognition technology, the intelligent voice assistant enters the sight of people, when a user cannot use two hands to operate at present, the intelligent voice assistant can quickly recognize voice input by the user and execute some control instructions included in the voice, such as setting an alarm clock, sending short messages, dialing a telephone and the like, the user can execute the desired operation only by inputting the voice, and convenience is added to daily life.
At present, after a terminal collects a voice containing a control instruction input by a user, the terminal recognizes the control instruction in the voice and immediately executes an operation corresponding to the control instruction, for example, the user inputs a voice "call to zhang san" and immediately dials a call of zhang san. If the user wants to perform some operations at other time, the user can only perform voice control on the terminal at the time when the terminal is required to perform the operations, which is inconvenient, and the user may forget the operations he wants to perform after the time arrives.
In summary, currently, the terminal can only be controlled by voice when the terminal is required to execute the control operation, and the voice control method is not flexible.
Disclosure of Invention
The invention provides a voice control method, a terminal and a computer storage medium, which are used for solving the problems that the voice control can only be carried out on the terminal when the terminal is required to execute control operation at present and the voice control method is not flexible enough.
Based on the foregoing problem, in a first aspect, an embodiment of the present invention provides a voice control method, including:
the method comprises the steps that a terminal collects voice data input by a user and used for carrying out delay control on the terminal through a microphone;
the terminal identifies a trigger condition contained in the voice data and determines control information for controlling the terminal according to the voice data;
the terminal stores the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and after determining that the trigger condition is met in the operation set to be executed, the terminal acquires control information corresponding to the met trigger condition from the operation set to be executed and executes the control operation corresponding to the control information.
In a second aspect, an embodiment of the present invention provides a voice control terminal, including a microphone, a memory, and a processor;
the microphone is used for collecting voice data which are input by a user and used for carrying out delay control on the terminal;
the memory storing program code that, when executed by the processor, causes the processor to perform the following:
recognizing a trigger condition contained in the voice data, and determining control information for controlling the terminal according to the voice data;
storing the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and after determining that the trigger condition is met in the operation set to be executed, acquiring control information corresponding to the met trigger condition from the operation set to be executed, and executing the control operation corresponding to the control information.
In a third aspect, an embodiment of the present invention provides a voice control apparatus, including:
the acquisition module is used for acquiring voice data which is input by a user and used for carrying out delay control on the terminal;
the recognition module is used for recognizing the trigger condition contained in the voice data and determining control information for controlling the terminal according to the voice data;
the storage module is used for storing the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and the execution module is used for acquiring the control information corresponding to the satisfied trigger condition from the operation set to be executed and executing the control operation corresponding to the control information after determining that the trigger condition is satisfied in the operation set to be executed.
In a fourth aspect, embodiments of the present invention provide a computer-readable medium on which a computer program is stored, which when executed by a processor, performs the steps of the method according to the first aspect.
According to the voice control method provided by the embodiment of the invention, the terminal acquires and identifies the voice data which is input by the user and used for controlling the terminal, and if the trigger condition is identified from the voice data, the terminal executes the control operation corresponding to the control instruction identified from the voice data after determining that the trigger condition is met, so that the non-real-time control of the voice instruction is realized, and the problem that the existing voice control method only can realize the real-time control on the control instruction and is not flexible is solved.
Drawings
The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments consistent with the invention and together with the description, serve to explain the principles of the invention and are not to be construed as limiting the invention.
FIG. 1 is a flow chart of a voice control method according to an embodiment of the present invention;
fig. 2 is a schematic diagram of a terminal to which the embodiment of the present invention is applicable;
FIG. 3 is a diagram of a first speech input interface according to an embodiment of the present invention;
FIG. 4 is a diagram illustrating a second exemplary voice input interface according to the present invention;
FIG. 5 is a diagram illustrating a first set of operations to be performed according to the embodiment of the present invention;
FIG. 6 is a diagram illustrating a modification command in response to a user trigger according to an embodiment of the present invention;
FIG. 7 is a diagram illustrating a second set of operations to be performed according to the embodiment of the present invention;
FIG. 8 is a diagram of a third exemplary embodiment of a speech input interface;
fig. 9 is a schematic diagram illustrating that the terminal deletes the trigger condition and the control information corresponding to the user control instruction from the operation set to be executed according to the embodiment of the present invention;
FIG. 10 is a diagram illustrating a list of operations to be performed being jumped by an operation option to be performed according to an embodiment of the present invention;
FIG. 11 is a flowchart of a first method for controlling speech according to the present invention;
FIG. 12 is a flowchart of a second method for controlling speech integrity according to an embodiment of the present invention;
fig. 13 is a schematic structural diagram of a first voice control terminal according to an embodiment of the present invention;
fig. 14 is a schematic structural diagram of a second voice-controlled terminal according to an embodiment of the present invention;
fig. 15 is a schematic structural diagram of a voice control apparatus according to an embodiment of the present invention.
Detailed Description
In order to make those skilled in the art better understand the technical solution of the present invention, the technical solution in the embodiments of the present invention will be clearly and completely described below with reference to the accompanying drawings.
It should be noted that the terms "first," "second," and the like in the description and claims of the present invention and in the drawings described above are used for distinguishing between similar elements and not necessarily for describing a particular sequential or chronological order. It is to be understood that the data so used is interchangeable under appropriate circumstances such that the embodiments of the invention described herein are capable of operation in sequences other than those illustrated or described herein. The embodiments described in the following exemplary embodiments do not represent all embodiments consistent with the present invention. Rather, they are merely examples of apparatus and methods consistent with certain aspects of the invention, as detailed in the appended claims.
The term "and/or" in the embodiments of the present invention describes an association relationship of associated objects, and indicates that three relationships may exist, for example, a and/or B may indicate: a exists alone, A and B exist simultaneously, and B exists alone. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
The term "terminal" in the embodiment of the present invention is a device that can install various applications and can display an object provided in an installed application, and the terminal device may be mobile or fixed. For example, a mobile phone, a tablet computer, various wearable devices, a vehicle-mounted device, a Personal Digital Assistant (PDA), a point of sale (POS), or other terminal devices capable of implementing the above functions.
At present, a terminal collects voice data which is input by a user and contains a control instruction, and immediately executes a control operation corresponding to the control instruction after the control instruction contained in the voice data is recognized; for example, if a user inputs voice "call to zhang san", he can immediately call zhang san, and if he wants to perform some operations at other time, he can only need the time for the terminal to perform operations to perform voice control on the terminal, which is inconvenient, and he may forget the operations he wants to perform after the time arrives, so that there is a problem that the voice control can be performed on the terminal only when he needs to perform control operations on the terminal, and the voice control method is not flexible enough.
Based on the above problem, an embodiment of the present invention provides a voice control method, as shown in fig. 1, the method includes the following steps:
s101, a terminal collects voice data input by a user and used for carrying out delay control on the terminal through a microphone;
step S102, the terminal identifies a trigger condition contained in the voice data and determines control information for controlling the terminal according to the voice data;
step S103, the terminal stores the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and step S104, after determining that the trigger condition in the operation set to be executed is met, the terminal acquires control information corresponding to the met trigger condition from the operation set to be executed and executes the control operation corresponding to the control information.
It should be noted here that the control operation corresponding to the control information is preset by a person skilled in the art, for example, the control operation corresponding to the control information may be "call," "send a message," "update the system," or the like.
Fig. 2 illustrates a schematic diagram of a terminal suitable for use in embodiments of the present invention; the terminal comprises a microphone, and the terminal collects voice data input by a user through the microphone.
In the embodiment of the invention, the voice data input by the user for carrying out delay control on the terminal can be the voice data input by the user once or the voice data input by the user for multiple times.
Specifically, after a user inputs a piece of voice data, the terminal collects the voice data through a microphone, and if the terminal recognizes a trigger condition from the voice data and determines that the voice data contains control information for controlling the terminal, the terminal stores a corresponding relation between the trigger condition and the control information into an operation set to be executed;
the operation set to be executed comprises the corresponding relation between the triggering condition which is input by the terminal and is not executed and the control information;
after the terminal recognizes the trigger condition from the voice data, an optional implementation manner is that the terminal also recognizes a control instruction contained in the voice data, if it is determined that the voice data contains the control instruction, determines control information for controlling the terminal according to the voice data, and stores a corresponding relationship between the trigger condition and the control information in an operation set to be executed.
For example, the user inputs a voice "eight night helps me to explain a meal in the day for three WeChat letters", the terminal recognizes the trigger condition "eight night" from the voice data, and recognizes a control instruction "explain a meal in the day for three WeChat letters" contained in the voice data, the terminal further determines control information for controlling the terminal according to the voice data, and the control information and the trigger condition "eight night" are correspondingly stored in the operation set to be executed.
If the terminal recognizes only the trigger condition from a piece of voice data input by the user and does not recognize the control instruction, an optional implementation manner is that the terminal returns a recognition result of the trigger condition.
For example, as shown in the voice input interface shown in fig. 3, the user inputs a voice "eight night", the terminal only recognizes that the voice data includes the trigger condition "eight night", and if no control instruction for the terminal is recognized, the recognition result for "eight night" is returned.
Another optional implementation manner is that, if the terminal recognizes only the trigger condition from one piece of voice data input by the user and does not recognize that the voice data contains the control instruction, the terminal receives the voice data containing the control instruction input again by the user through the microphone, determines control information for controlling the terminal according to the voice data input by the user for multiple times, and stores the corresponding relationship between the trigger condition and the control information into the operation set to be executed.
In an implementation, after only recognizing a trigger condition from a piece of voice data input by a user, the terminal may prompt the user whether to input voice data containing a control instruction, for example, as shown in fig. 4, the user inputs a voice "eight night", the terminal may prompt the user "what do i need to do for you at eight night", the user inputs a voice data "help me make a call to zhang san, the terminal determines control information for the terminal to control after receiving the voice data, and the trigger condition" eight night "and the control information are correspondingly stored in an operation set to be executed.
After the terminal collects the voice data input by the user and identifies the trigger condition contained in the voice data, an optional implementation manner is that the voice data is used as control information, and the corresponding relation between the trigger condition and the voice data is stored in an operation set to be executed.
Another optional implementation manner is that a control instruction contained in the voice data is recognized, the control instruction is used as control information, and the corresponding relationship between the trigger condition and the control instruction is stored in the operation set to be executed.
The following describes different ways of specifying control information.
In the first mode, the terminal uses voice data as control information.
The terminal identifies a trigger condition contained in voice data which is input by a user and used for carrying out delay control on the terminal, and stores the corresponding relation between the trigger condition and the voice data into an operation set to be executed after the voice data input by the user is used as control information.
Here, the terminal may use all the voice data input by the user for performing the delay control on the terminal as the control information, or may use part of the voice data including the control instruction in the voice data input by the user for performing the delay control on the terminal as the control information.
In implementation, for example, a user inputs a voice "eight-point-in-morning update system" for performing delay control on a terminal, after the terminal recognizes a trigger condition "eight-point-in-morning" in the voice, if the terminal uses all voice data input by the user for performing delay control on the terminal as control information, the terminal uses all the voice data "eight-point-in-morning update system" as the control information, and stores the trigger condition "eight-point-in-morning" and the voice data "eight-point-in-morning update system" in a corresponding manner in an operation set to be executed; if the terminal uses the voice data which only contains the control instruction in the voice data which is input by the user and used for carrying out the delay control on the terminal as the control information, the terminal uses the voice data 'updating system' which only contains the control instruction in the voice of the delay control as the control information, and correspondingly stores the triggering condition 'eight morning spots' and the voice data 'updating system' in the operation set to be executed.
And after determining that the trigger condition is met in the operation set to be executed, the terminal acquires the voice data corresponding to the met trigger condition from the operation set to be executed and executes the control operation corresponding to the voice data.
An optional implementation manner is that after the terminal acquires the voice data corresponding to the satisfied trigger condition from the operation set to be executed, before executing the control operation corresponding to the voice data, the terminal further recognizes the control instruction contained in the voice data.
Specifically, after determining that the trigger condition is satisfied in the operation set to be executed, the terminal acquires the voice data corresponding to the trigger condition from the operation set to be executed, re-identifies the control instruction contained in the voice data, and executes the control operation corresponding to the control instruction.
For example, the operation set to be executed includes a trigger condition "seven night" and corresponding voice data "seven night open WeChat", a trigger condition "eight night" and corresponding voice data "eight night gives three-to-three short messages and says night; the terminal determines that the trigger condition ' seven night ' in the operation set to be executed is met after the terminal determines that the time is seven night ', acquires the voice data corresponding to the trigger condition ' seven night ' from the operation set to be executed, and recognizes the control instruction ' open WeChat ' contained in the voice data again, and then the terminal executes the operation of opening the WeChat. After the terminal determines that the time is eight nights, it is determined that a triggering condition of 'eight nights' is met in the operation set to be executed, voice data corresponding to the triggering condition of 'eight nights' is acquired from the operation set to be executed and is sent to the third party for sending the short message with night, a control instruction 'sending the short message with night' contained in the voice is identified again, and then the terminal executes the operation of sending the short message with night to the third party.
In the second method, the terminal recognizes the control command included in the voice data, and uses the recognized control command as control information.
After the terminal collects voice data input by a user, identifying a trigger condition and a control instruction which are contained in the voice data input by the user and used for carrying out delay control on the terminal; and the identified control instruction is used as control information, and the corresponding relation between the trigger condition and the control instruction is stored in the operation set to be executed.
For example, if the voice data input by the user to perform delay control on the terminal is "update system after wifi connection", the terminal recognizes the triggering condition "wifi connection" and the control instruction "update system" in the voice data input by the user, then uses the control instruction "update system" as the control information, and correspondingly stores the triggering condition "wifi connection" and the control instruction "update system" in the operation set to be executed.
And after determining that the trigger condition is met in the to-be-operated set, the terminal acquires an operation instruction corresponding to the met trigger condition from the to-be-executed operation set and executes the control operation corresponding to the operation instruction.
For example, if the to-be-executed operation set includes a trigger condition of "wifi connection" and a corresponding control instruction of "call for lie IV", the terminal determines that the trigger condition of "wifi connection" in the to-be-executed operation set is satisfied after determining that wifi connection is performed, obtains the control instruction of "wifi connection" corresponding to the trigger condition of "call for lie IV" from the to-be-executed operation set, and then executes an operation of calling for lie IV.
The trigger condition included in the operation set to be executed in the embodiment of the present invention may include one or more pieces of condition information, and when determining whether the trigger condition in the operation set to be executed is satisfied, the terminal needs to determine according to the condition information included in the trigger condition.
An optional implementation manner is that the terminal determines that the trigger condition is satisfied according to the following partial or all manners:
in the mode 1, for a trigger condition containing a time point, after the time point is reached, the trigger condition is determined to be met;
in the mode 1, the condition information included in the trigger condition is a time point; for example, the time points may be "eight evening," "nine tomorrow morning," "six weekly," and so forth.
If the condition information included in the trigger condition is only the time point, the terminal determines that the trigger condition is satisfied after the time point included in the trigger condition is reached.
Specifically, after recognizing that voice data input by a user and used for performing delay control on the terminal comprises a trigger condition of a time point, the terminal sets a notification message in the system according to a preset notification format and corresponding time, and after the corresponding time point arrives, the terminal receives the notification message sent by the system and determines that the trigger condition is met.
Mode 2, for a trigger condition containing state information, when the trigger condition is in a state corresponding to the state information, determining that the trigger condition is met;
in the embodiment 2, the condition information included in the trigger condition is only the state information; for example, the status information may be "connect wifi", "disconnect wifi", "connect bluetooth", "disconnect bluetooth", and so on.
If the condition information included in the trigger condition is only the state information, the terminal determines that the trigger condition is satisfied when the terminal is in a state corresponding to the state information included in the trigger condition.
Specifically, when a state corresponding to the state information included in the trigger condition is reached, the terminal receives a system notification that the state is reached, and after receiving the system notification, it is determined that the trigger condition is satisfied.
Mode 3, for a trigger condition including location information, when the trigger condition is located at a location corresponding to the location information, it is determined that the trigger condition is satisfied.
In the aspect 3, the condition information included in the trigger condition is only the position information; for example, the location information may be "after xx company", "after xx cell", and the like.
If the condition information included in the trigger condition is only the location information, the terminal determines that the trigger condition is satisfied when the terminal is located at the location corresponding to the location information included in the trigger condition.
Mode 4, for the trigger condition containing the message type and the object, after receiving the message of the message type sent by the object, determining that the trigger condition is satisfied.
In the mode 4, the condition information included in the trigger condition is a message type and an object; for example, the message type and object may be "WeChat message received xx", "QQ message received xx", and so on.
If the condition information included in the trigger condition is only the message type and the object, the terminal determines that the trigger condition is satisfied after receiving the message type sent by the object included in the trigger condition.
Specifically, the terminal monitors a system notification message after recognizing that voice data input by a user and used for performing delay control on the terminal includes a message type and a trigger condition of an object, acquires notification message content and object information after monitoring a system notification message corresponding to the message type included in the trigger condition, and further determines whether the object is an object included in the trigger condition, and if so, determines that the trigger condition is satisfied.
If the trigger condition includes a plurality of pieces of condition information, the terminal determines that the trigger condition is satisfied after determining that all pieces of condition information included in the trigger condition are satisfied.
In example 1, if the condition information included in the trigger condition is time point and state information "after wifi is connected at eight night", the terminal determines that the trigger condition is satisfied after determining that the time point reaches at eight night and the state of the terminal is wifi, and if the terminal determines that the time point reaches at eight night but the state of the terminal is wifi is disconnected or the terminal determines that the state is wifi is connected but the time point does not reach eight night, the terminal determines that the trigger condition is not satisfied.
Example 2, if the condition information included in the trigger condition is status information, location information, a message type, and an object, "a fourth WeChat message is received in the XX cell at seven am", the terminal determines that the seven morning time arrives and determines that the current location of the user is in the XX cell, and determines that the trigger condition is satisfied after receiving the WeChat message sent by the fourth WeChat message.
In the embodiment of the present invention, after the terminal stores the corresponding relationship between the trigger condition and the control information in the to-be-executed operation set, before determining that the trigger condition is satisfied, the terminal may further modify and delete the trigger condition and/or the control information in the to-be-executed operation set according to the actual requirement of the user, so as to ensure that the trigger condition and the corresponding control information in the to-be-executed operation set are both the trigger condition and the control information required by the user.
1. And modifying the trigger condition and/or the control information in the operation set to be executed.
In an optional implementation manner, the terminal responds to a modification instruction triggered by a user, and modifies a trigger condition and/or control information corresponding to the modification instruction in the operation set to be executed.
For example, as shown in fig. 5, the user selects the control information "call three at seven points in today's evening" in the operation set interface to be executed to modify, as shown in fig. 6, the terminal responds to a modification instruction that the user modifies the trigger condition of "seven points in today's evening" contained in the control information to "eight points in today's evening" and modifies the control information corresponding to the instruction.
2. And deleting the trigger condition and/or the control information in the operation set to be executed.
In an optional implementation manner, the terminal responds to a deletion instruction triggered by the user, and deletes a corresponding relationship between a trigger condition corresponding to the deletion instruction and the control information in the operation set to be executed.
For example, as shown in the operation set interface to be executed in fig. 7, if the user selects the trigger condition and the corresponding control information of "leave the duet-four-letter WeChat together in the operation set interface to be executed" to delete, the terminal responds to the deletion instruction triggered by the user to delete the trigger condition and the corresponding control information from the operation set interface to be executed.
According to the embodiment of the invention, a user can input the deleting instruction or the modifying instruction through voice, and after the deleting instruction or the modifying instruction input by the user is identified, the terminal executes modification or deletion of the corresponding trigger condition and/or control information in the operation set to be executed.
As shown in fig. 8, in the voice input interface, if the user inputs a voice "call is not made three times a night" including a deletion instruction, after the terminal recognizes the deletion instruction included in the voice data, it determines that the trigger condition and the corresponding control information are in the operation set to be executed, and then the terminal responds to the deletion instruction, and deletes the trigger condition and the control information corresponding to the control instruction as shown in fig. 9.
In the embodiment of the invention, the operation set to be executed can be stored in the terminal in a mode of an operation list to be executed; as shown in fig. 10, the setting display interface of the terminal includes an operation option to be executed, and after clicking the operation option to be executed, the user may jump to the display interface of the operation list to be executed, where the operation list to be executed includes the correspondence between all trigger conditions and control information in the operation set to be executed, so that the user may modify and delete the trigger conditions and the corresponding control information in the operation set to be executed.
It should be noted here that, the position of the operation option to be executed preset in the terminal may be preset in a setting interface of the terminal, or may be set in a security and privacy setting interface of the terminal system, and is not limited specifically.
Fig. 11 is a flowchart illustrating a first voice control completion method according to an exemplary embodiment, where a terminal uses voice data as control information, and specifically includes the following steps:
step S1101, inputting voice data for performing delay control on the terminal by a user;
step S1102, the terminal identifies the trigger condition contained in the acquired voice data of the delay control, and stores the identified trigger condition and the voice data into an operation set to be executed;
step S1103, after determining that the trigger condition in the operation set to be executed is satisfied, the terminal acquires voice data corresponding to the satisfied trigger condition from the operation set to be executed;
step S1104, the terminal identifies the control instruction included in the voice data, and executes a control operation corresponding to the control instruction.
Fig. 12 is a flowchart illustrating a second method for controlling speech according to an exemplary embodiment, where a terminal uses a recognized control command as control information, and specifically includes the following steps:
step S1201, inputting voice data for performing delay control on the terminal by a user;
step S1202, the terminal identifies the trigger condition and the control instruction contained in the acquired voice data of the delay control;
step S1203, the terminal stores the identified trigger condition and the control instruction corresponding to the trigger condition into an operation set to be executed;
step S1204, after determining that the trigger condition is satisfied in the operation set to be executed, the terminal acquires a control instruction corresponding to the satisfied trigger condition from the operation set to be executed;
and step S1205, the terminal executes the control operation corresponding to the control instruction.
Based on the same inventive concept, the embodiment of the present invention further provides a voice control terminal, and as the principle of the terminal for voice control is similar to the voice control method provided by the embodiment of the present invention, the implementation of the terminal may refer to the implementation of the method, and repeated details are not described again.
As shown in fig. 13, a voice control terminal 1300 according to an embodiment of the present invention includes a microphone 1301, a memory 1302, and a processor 1303;
the microphone 1301 is used for acquiring voice data which is input by a user and used for carrying out delay control on the terminal;
the memory 1302 is configured to store program codes, and when the program codes are executed by the processor 1303, the processor 1303 executes the following processes:
recognizing a trigger condition contained in the voice data, and determining control information for controlling the terminal according to the voice data;
storing the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and after determining that the trigger condition is met in the operation set to be executed, acquiring control information corresponding to the met trigger condition from the operation set to be executed, and executing the control operation corresponding to the control information.
Optionally, the processor 1303 is specifically configured to:
using the voice data as the control information;
and identifying a control instruction contained in the control information, and executing a control operation corresponding to the control instruction.
Optionally, the processor 1303 is specifically configured to:
and recognizing a control instruction contained in the voice data, and using the recognized control instruction as the control information.
Optionally, after the terminal stores the corresponding relationship between the trigger condition and the control information in the operation set to be executed, before determining that the trigger condition is satisfied, the processor 1303 is further configured to:
responding to a modification instruction triggered by the user, and modifying the trigger condition and/or the control information corresponding to the modification instruction in the operation set to be executed; or the like, or, alternatively,
and responding to a deleting instruction triggered by the user, and deleting the corresponding relation between the triggering condition corresponding to the deleting instruction in the operation set to be executed and the control information.
Optionally, the processor 1303 is configured to determine that the trigger condition is met according to some or all of the following manners:
in the mode 1, for a trigger condition containing a time point, after the time point is reached, the trigger condition is determined to be met;
mode 2, for a trigger condition containing state information, when the trigger condition is in a state corresponding to the state information, determining that the trigger condition is met;
mode 3, for a trigger condition containing position information, when the trigger condition is located at a position corresponding to the position information, determining that the trigger condition is met;
mode 4, for the trigger condition containing the message type and the object, after receiving the message of the message type sent by the object, determining that the trigger condition is satisfied.
As shown in fig. 14, another voice control terminal 1400 according to an embodiment of the present invention includes: a Radio Frequency (RF) circuit 1410, a power supply 1420, a processor 1430, a memory 1440, an input unit 1450, a display unit 1460, a camera 1470, a communication interface 1480, and a Wireless Fidelity (WiFi) module 1490, a microphone 14100, and the like. Those skilled in the art will appreciate that the configuration of the terminal shown in fig. 14 is not intended to be limiting, and that the terminal provided by the embodiments of the present application may include more or less components than those shown, or some components may be combined, or a different arrangement of components may be provided.
The following describes the various components of the terminal 1400 in detail with reference to fig. 14:
the RF circuitry 1410 may be used for receiving and transmitting data during a communication or conversation. Specifically, the RF circuit 1410 sends downlink data of a base station to the processor 1430 for processing; and in addition, sending the uplink data to be sent to the base station. In general, the RF circuit 1410 includes, but is not limited to, an antenna, at least one Amplifier, a transceiver, a coupler, a Low Noise Amplifier (LNA), a duplexer, and the like.
In addition, the RF circuit 1410 may also communicate with a network and other terminals through wireless communication. The wireless communication may use any communication standard or protocol, including but not limited to Global System for mobile communications (GSM), General Packet Radio Service (GPRS), Code Division Multiple Access (CDMA), Wideband Code Division Multiple Access (WCDMA), Long Term Evolution (LTE), email, Short Messaging Service (SMS), and the like.
The WiFi technology belongs to a short-distance wireless transmission technology, and the terminal 1400 is connected to an Access Point (AP) through a WiFi module 1490, so as to achieve Access to a data network. The WiFi module 1490 can be used for receiving and transmitting data during communication.
The terminal 1400 can be physically connected to other terminals through the communication interface 1480. Optionally, the communication interface 1480 is connected to a communication interface of the other terminal through a cable, so as to implement data transmission between the terminal 1400 and the other terminal.
In the embodiment of the present application, the terminal 1400 is capable of implementing a communication service and sending information to other contacts, so that the terminal 1400 needs to have a data transmission function, that is, the terminal 1400 needs to include a communication module inside. Although fig. 14 shows communication modules such as the RF circuit 1410, the WiFi module 1490, and the communication interface 1480, it is to be understood that at least one of the above components or other communication modules (e.g., bluetooth module) for enabling communication exists in the terminal 1400 for data transmission.
For example, when the terminal 1400 is a mobile phone, the terminal 1400 may include the RF circuit 1410, and may further include the WiFi module 1490; when the terminal 1400 is a computer, the terminal 1400 may include the communication interface 1480 and may further include the WiFi module 1490; when the terminal 1400 is a tablet computer, the terminal 1400 may include the WiFi module.
The memory 1440 may be used to store software programs and modules. The processor 1430 executes various functional applications and data processing of the terminal 1400 by executing software programs and modules stored in the memory 1440, and after the processor 1430 executes the program codes in the memory 1440, part or all of the processes in fig. 1 to 11 according to the embodiments of the present invention can be implemented.
Alternatively, the memory 1440 may mainly include a program storage area and a data storage area. Wherein, the storage program area can store an operating system, various application programs (such as communication application), various modules for WLAN connection, and the like; the storage data area may store data created according to the use of the terminal, and the like.
Further, the memory 1440 may include high speed random access memory, and may also include non-volatile memory, such as at least one magnetic disk storage device, flash memory device, or other volatile solid state storage device.
The input unit 1450 may be used to receive numeric or character information input by a user and to generate key signal inputs related to user settings and function control of the terminal 1400.
Alternatively, the input unit 1450 may include a touch panel 1451 and other input terminals 1452.
The touch panel 1451, also referred to as a touch screen, may collect touch operations of a user (for example, operations of the user on or near the touch panel 1451 using any suitable object or accessory such as a finger, a stylus pen, etc.) and drive a corresponding connection device according to a preset program. Alternatively, the touch panel 1451 may include two parts, i.e., a touch detection device and a touch controller. The touch detection device detects the touch direction of a user, detects a signal brought by touch operation and transmits the signal to the touch controller; the touch controller receives touch information from the touch sensing device, converts the touch information into touch point coordinates, and sends the touch point coordinates to the processor 1430, and can receive and execute commands sent from the processor 1430. In addition, the touch panel 1451 may be implemented in various types, such as resistive, capacitive, infrared, and surface acoustic wave.
Optionally, the other input terminals 1452 may include, but are not limited to, one or more of a physical keyboard, function keys (e.g., volume control keys, switch keys, etc.), a trackball, a mouse, a joystick, and the like.
The display unit 1460 may be used to display information input by a user or information provided to a user and various menus of the terminal 1400. The display unit 1460 is a display system of the terminal 1400, and is used for presenting an interface and implementing human-computer interaction.
The display unit 1460 may include a display panel 1461. Alternatively, the Display panel 1461 may be configured in the form of a Liquid Crystal Display (LCD), an Organic Light-emitting diode (OLED), or the like.
Further, the touch panel 1451 may cover the display panel 1461, and when the touch panel 1451 detects a touch operation on or near the touch panel, the touch operation is transmitted to the processor 1430 to determine the type of the touch event, and then the processor 1430 provides a corresponding visual output on the display panel 1461 according to the type of the touch event.
Although in fig. 14, the touch panel 1451 and the display 1461 are implemented as two separate components to implement the input and output functions of the terminal 1400, in some embodiments, the touch panel 1451 and the display 1461 may be integrated to implement the input and output functions of the terminal 1400.
The processor 1430 is a control center of the terminal 1400, connects various components using various interfaces and lines, performs various functions of the terminal 1400 and processes data by operating or executing software programs and/or modules stored in the memory 1440 and calling data stored in the memory 1440, thereby implementing various services based on the terminal.
Optionally, the processor 1430 may include one or more processing units. Optionally, the processor 1430 may integrate an application processor and a modem processor, wherein the application processor mainly handles operating systems, user interfaces, application programs, and the like, and the modem processor mainly handles wireless communications. It will be appreciated that the modem processor described above may not be integrated into the processor 1430.
The camera 1470 is configured to implement a shooting function of the terminal 1400 and shoot pictures or videos.
The terminal 1400 also includes a power supply 1420 (e.g., a battery) for powering the various components. Optionally, the power supply 1420 may be logically connected to the processor 1430 through a power management system, so as to implement functions of managing charging, discharging, power consumption, and the like through the power management system.
The microphone 14100 is used for collecting voice data input by a user.
Although not shown, the terminal 1400 may further include at least one sensor, an audio circuit, and the like, which are not described in detail herein.
As shown in fig. 15, a voice control apparatus according to an embodiment of the present invention includes:
the acquisition module 1501 is configured to acquire voice data input by a user and used for performing delay control on the terminal;
an identifying module 1502, configured to identify a trigger condition included in the voice data, and determine control information for controlling the terminal according to the voice data;
a storage module 1503, configured to store the corresponding relationship between the trigger condition and the control information in an operation set to be executed;
the executing module 1504, after determining that the trigger condition in the to-be-executed operation set is satisfied, acquires the control information corresponding to the satisfied trigger condition from the to-be-executed operation set, and executes the control operation corresponding to the control information.
Optionally, the identifying module 1502 is specifically configured to:
using the voice data as the control information;
the executing module 1504 is specifically configured to:
and identifying a control instruction contained in the control information, and executing a control operation corresponding to the control instruction.
Optionally, the identifying module 1502 is specifically configured to:
and the terminal identifies a control instruction contained in the voice data and takes the identified control instruction as the control information.
Optionally, after the storage module 1503 stores the corresponding relationship between the trigger condition and the control information in the operation set to be executed, before determining that the trigger condition is satisfied, the execution module 1504 is further configured to:
responding to a modification instruction triggered by the user, and modifying the trigger condition and/or the control information corresponding to the modification instruction in the operation set to be executed; or the like, or, alternatively,
and responding to a deleting instruction triggered by the user, and deleting the corresponding relation between the triggering condition corresponding to the deleting instruction in the operation set to be executed and the control information.
Optionally, the executing module 1504 is configured to determine that the trigger condition is met according to some or all of the following ways:
in the mode 1, for a trigger condition containing a time point, after the time point is reached, the trigger condition is determined to be met;
mode 2, for a trigger condition containing state information, when the trigger condition is in a state corresponding to the state information, determining that the trigger condition is met;
mode 3, for a trigger condition containing position information, when the trigger condition is located at a position corresponding to the position information, determining that the trigger condition is met;
mode 4, for the trigger condition containing the message type and the object, after receiving the message of the message type sent by the object, determining that the trigger condition is satisfied.
As will be appreciated by one skilled in the art, embodiments of the present application may be provided as a method, system, or computer program product. Accordingly, the present application may take the form of an entirely hardware embodiment, an entirely software embodiment or an embodiment combining software and hardware aspects. Furthermore, the present application may take the form of a computer program product embodied on one or more computer-usable storage media (including, but not limited to, disk storage, CD-ROM, optical storage, and the like) having computer-usable program code embodied therein.
The present application is described with reference to flowchart illustrations and/or block diagrams of methods, apparatus (systems), and computer program products according to the application. It will be understood that each flow and/or block of the flow diagrams and/or block diagrams, and combinations of flows and/or blocks in the flow diagrams and/or block diagrams, can be implemented by computer program instructions. These computer program instructions may be provided to a processor of a general purpose computer, special purpose computer, embedded processor, or other programmable data processing apparatus to produce a machine, such that the instructions, which execute via the processor of the computer or other programmable data processing apparatus, create means for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be stored in a computer-readable memory that can direct a computer or other programmable data processing apparatus to function in a particular manner, such that the instructions stored in the computer-readable memory produce an article of manufacture including instruction means which implement the function specified in the flowchart flow or flows and/or block diagram block or blocks.
These computer program instructions may also be loaded onto a computer or other programmable data processing apparatus to cause a series of operational steps to be performed on the computer or other programmable apparatus to produce a computer implemented process such that the instructions which execute on the computer or other programmable apparatus provide steps for implementing the functions specified in the flowchart flow or flows and/or block diagram block or blocks.
It will be apparent to those skilled in the art that various changes and modifications may be made in the present application without departing from the spirit and scope of the application. Thus, if such modifications and variations of the present application fall within the scope of the claims of the present application and their equivalents, the present application is intended to include such modifications and variations as well.

Claims (10)

1. A voice control terminal is characterized by comprising a microphone, a memory and a processor;
the microphone is used for collecting voice data which are input by a user and used for carrying out delay control on the terminal;
the memory storing program code that, when executed by the processor, causes the processor to perform the following:
recognizing a trigger condition contained in the voice data, and determining control information for controlling the terminal according to the voice data;
storing the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and after determining that the trigger condition is met in the operation set to be executed, acquiring control information corresponding to the met trigger condition from the operation set to be executed, and executing the control operation corresponding to the control information.
2. The terminal of claim 1, wherein the processor, when determining control information for controlling the terminal from the voice data, is specifically configured to:
using the voice data as the control information;
the processor, when executing the control operation corresponding to the control information, is specifically configured to:
and identifying a control instruction contained in the control information, and executing a control operation corresponding to the control instruction.
3. The terminal of claim 1, wherein the processor, when determining control information for controlling the terminal from the voice data, is specifically configured to:
and recognizing a control instruction contained in the voice data, and using the recognized control instruction as the control information.
4. The terminal according to any of claims 1 to 3, wherein after storing the correspondence between the trigger condition and the control information in the set of operations to be executed, before determining that the trigger condition is satisfied, the processor is further configured to:
responding to a modification instruction triggered by the user, and modifying the trigger condition and/or the control information corresponding to the modification instruction in the operation set to be executed; or
And responding to a deleting instruction triggered by the user, and deleting the corresponding relation between the triggering condition corresponding to the deleting instruction in the operation set to be executed and the control information.
5. The terminal of any of claims 1 to 3, wherein the processor is specifically configured to:
determining that the trigger condition is satisfied according to some or all of the following ways:
in the mode 1, for a trigger condition containing a time point, after the time point is reached, the trigger condition is determined to be met;
mode 2, for a trigger condition containing state information, when the trigger condition is in a state corresponding to the state information, determining that the trigger condition is met;
mode 3, for a trigger condition containing position information, when the trigger condition is located at a position corresponding to the position information, determining that the trigger condition is met;
mode 4, for the trigger condition containing the message type and the object, after receiving the message of the message type sent by the object, determining that the trigger condition is satisfied.
6. A method for voice control, the method comprising:
the method comprises the steps that a terminal collects voice data input by a user and used for carrying out delay control on the terminal through a microphone;
the terminal identifies a trigger condition contained in the voice data and determines control information for controlling the terminal according to the voice data;
the terminal stores the corresponding relation between the trigger condition and the control information into an operation set to be executed;
and after determining that the trigger condition is met in the operation set to be executed, the terminal acquires control information corresponding to the met trigger condition from the operation set to be executed and executes the control operation corresponding to the control information.
7. The method of claim 6, wherein the terminal determining control information for controlling the terminal based on the voice data comprises:
the terminal takes the voice data as the control information;
the terminal executes the control operation corresponding to the control information, and the control operation comprises the following steps:
and the terminal identifies the control instruction contained in the control information and executes the control operation corresponding to the control instruction.
8. The method of claim 6, wherein the terminal determining control information for controlling the terminal based on the voice data comprises:
and the terminal identifies a control instruction contained in the voice data and takes the identified control instruction as the control information.
9. The method according to any one of claims 6 to 8, wherein after the terminal stores the correspondence between the trigger condition and the control information in the operation set to be executed, before determining that the trigger condition is satisfied, the method further comprises:
the terminal responds to a modification instruction triggered by the user and modifies a trigger condition and/or control information corresponding to the modification instruction in the operation set to be executed; or
And the terminal responds to a deleting instruction triggered by the user and deletes the corresponding relation between the triggering condition corresponding to the deleting instruction in the operation set to be executed and the control information.
10. A computer storage medium having computer program instructions stored therein, which when run on a computer, cause the computer to perform the method of any one of claims 6-9.
CN201911007522.3A 2019-10-22 2019-10-22 Voice control method, terminal and computer storage medium Pending CN110706705A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201911007522.3A CN110706705A (en) 2019-10-22 2019-10-22 Voice control method, terminal and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201911007522.3A CN110706705A (en) 2019-10-22 2019-10-22 Voice control method, terminal and computer storage medium

Publications (1)

Publication Number Publication Date
CN110706705A true CN110706705A (en) 2020-01-17

Family

ID=69201052

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201911007522.3A Pending CN110706705A (en) 2019-10-22 2019-10-22 Voice control method, terminal and computer storage medium

Country Status (1)

Country Link
CN (1) CN110706705A (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114049894A (en) * 2022-01-11 2022-02-15 广州小鹏汽车科技有限公司 Voice interaction method and device, vehicle and storage medium
CN114067799A (en) * 2022-01-05 2022-02-18 广州小鹏汽车科技有限公司 Voice interaction method, terminal device, vehicle and storage medium

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070201639A1 (en) * 2006-02-14 2007-08-30 Samsung Electronics Co., Ltd. System and method for controlling voice detection of network terminal
CN102750126A (en) * 2012-06-27 2012-10-24 深圳Tcl新技术有限公司 Speech input method and terminal
CN108132768A (en) * 2016-12-01 2018-06-08 中兴通讯股份有限公司 The processing method of phonetic entry, terminal and network server
CN108281143A (en) * 2018-02-24 2018-07-13 姚诗晴 A kind of student's daily schedule intelligence management and control robot based on machine vision and interactive voice
CN109634550A (en) * 2018-11-20 2019-04-16 维沃移动通信有限公司 A kind of voice operating control method and terminal device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070201639A1 (en) * 2006-02-14 2007-08-30 Samsung Electronics Co., Ltd. System and method for controlling voice detection of network terminal
CN102750126A (en) * 2012-06-27 2012-10-24 深圳Tcl新技术有限公司 Speech input method and terminal
CN108132768A (en) * 2016-12-01 2018-06-08 中兴通讯股份有限公司 The processing method of phonetic entry, terminal and network server
CN108281143A (en) * 2018-02-24 2018-07-13 姚诗晴 A kind of student's daily schedule intelligence management and control robot based on machine vision and interactive voice
CN109634550A (en) * 2018-11-20 2019-04-16 维沃移动通信有限公司 A kind of voice operating control method and terminal device

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114067799A (en) * 2022-01-05 2022-02-18 广州小鹏汽车科技有限公司 Voice interaction method, terminal device, vehicle and storage medium
CN114049894A (en) * 2022-01-11 2022-02-15 广州小鹏汽车科技有限公司 Voice interaction method and device, vehicle and storage medium

Similar Documents

Publication Publication Date Title
CN107896272B (en) Call control method and device
CN106453511B (en) A kind of data back up method and equipment
CN108323237B (en) Terminal behavior triggering method and terminal
CN110837406B (en) Suspension ball view display method and terminal
CN110430592B (en) Network control method and related product
CN106446111B (en) Data migration method and terminal
CN106658753B (en) Data migration method and terminal equipment
WO2015024372A1 (en) Communication initiation method, apparatus, and mobile terminal
CN107317918B (en) Parameter setting method and related product
CN106445739B (en) Data migration method and terminal equipment
CN110753153A (en) Terminal control method, terminal and computer storage medium
CN110706705A (en) Voice control method, terminal and computer storage medium
CN106484563B (en) Data migration method and terminal equipment
CN107071151B (en) Control method for physical key corresponding to home key and mobile terminal
CN105227775A (en) A kind of voice incoming call processing method and device
WO2015024393A1 (en) Method, apparatus, and terminal for sending sms message
CN107040658B (en) Mobile terminal and method and device for controlling screen thereof
CN106850957B (en) Application notification management method and device and mobile terminal
CN109068000B (en) Sensor control method, mobile terminal, and computer-readable storage medium
CN109309755B (en) Mobile terminal and APP display method and device
CN106851023B (en) Method and equipment for quickly making call and mobile terminal
CN110602325B (en) Voice recommendation method and device for terminal
CN111741164B (en) Sound mode switching method and device and computer readable storage medium
CN111144845B (en) Mobile terminal meeting reminding method and device
CN104516803A (en) Electronic device information monitoring method and system and electronic device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication
RJ01 Rejection of invention patent application after publication

Application publication date: 20200117