CN114664296A - Voice instruction execution method and device, intelligent terminal and storage medium - Google Patents

Voice instruction execution method and device, intelligent terminal and storage medium Download PDF

Info

Publication number
CN114664296A
CN114664296A CN202011423206.7A CN202011423206A CN114664296A CN 114664296 A CN114664296 A CN 114664296A CN 202011423206 A CN202011423206 A CN 202011423206A CN 114664296 A CN114664296 A CN 114664296A
Authority
CN
China
Prior art keywords
voice
instruction
information
event type
operation event
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202011423206.7A
Other languages
Chinese (zh)
Inventor
王云华
陈晶泊
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shenzhen TCL New Technology Co Ltd
Original Assignee
Shenzhen TCL New Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shenzhen TCL New Technology Co Ltd filed Critical Shenzhen TCL New Technology Co Ltd
Priority to CN202011423206.7A priority Critical patent/CN114664296A/en
Publication of CN114664296A publication Critical patent/CN114664296A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/18Speech classification or search using natural language modelling
    • G10L15/1822Parsing for meaning understanding
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Artificial Intelligence (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses a voice instruction execution method, a voice instruction execution device, an intelligent terminal and a storage medium, wherein the method comprises the following steps: acquiring voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information; determining an operation instruction corresponding to the operation event type according to the operation event type; and executing the operation corresponding to the operation instruction according to the operation instruction. According to the method and the device, the corresponding operation event type is determined according to the voice instruction, and then the operation instruction is generated according to the operation event type so as to execute the operation instruction, so that the voice instruction can be executed by acquiring the corresponding operation event type, the voice instruction can be conveniently, quickly and accurately executed, the voice information does not need to be considered to be updated, and convenience is provided for a user.

Description

Voice instruction execution method and device, intelligent terminal and storage medium
Technical Field
The invention relates to the technical field of voice instruction execution, in particular to a voice instruction execution method, a voice instruction execution device, an intelligent terminal and a storage medium.
Background
With the increasingly vigorous natural language processing technology, the technology of voice interaction is more and more mature. The voice assistant is widely applied to various IOT (Internet of things) devices such as mobile phones, televisions and computers, and covers various fields such as movie and television, weather inquiry, device control, shopping and consumption. However, the fields related to voice interaction are various, and in the prior art, voice interaction is basically realized based on a voice assistant, but the recognition accuracy of the voice assistant for user intentions is not high, and in order to recognize new user intentions, a user needs to frequently update the voice assistant to adapt to new requirements, so that the operation is complicated, and the use of the user is influenced.
Thus, there is a need for improvements and enhancements in the art.
Disclosure of Invention
The present invention provides a method, an apparatus, an intelligent terminal and a storage medium for executing a voice command, which aim to solve the problems that the accuracy of recognizing a user intention by a voice assistant is not high and the user needs to frequently update the voice assistant in order to recognize a new user intention.
In order to solve the technical problems, the technical scheme adopted by the invention is as follows:
in a first aspect, the present invention provides a method for executing a voice instruction, wherein the method includes:
acquiring voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information;
determining an operation instruction corresponding to the operation event type according to the operation event type;
and executing the operation corresponding to the operation instruction according to the operation instruction.
In one implementation manner, the obtaining voice operation request information generated based on a voice instruction and determining an operation event type corresponding to the voice instruction according to the voice operation request information include
Analyzing the voice operation request information to obtain voice information corresponding to the voice instruction in the voice operation request information;
and converting the voice information into text information, and determining the operation event type corresponding to the voice instruction according to the text information.
In one implementation, the determining, according to the text information, an operation event type corresponding to the voice instruction includes:
analyzing the text information to obtain field information in the text information;
and determining the type of the operation event corresponding to the field information according to the field information.
In one implementation, the determining, according to the field information, an operation event type corresponding to the field information includes:
according to the field information, acquiring a candidate operation event type matched with the field information;
and acquiring the priority information of the candidate operation event types, and determining the operation event types corresponding to the field information according to the priority information.
In one implementation, the determining, according to the field information, an operation event type corresponding to the field information includes:
rewriting the field information into specified field information;
and generating a specified operation event type corresponding to the specified field information according to the specified field information, and taking the specified operation event type as the operation event type.
In one implementation, the determining, according to the operation event type, an operation instruction corresponding to the operation event type includes:
determining the name information of the operation event type according to the operation event type;
and determining the operation instruction corresponding to the name information according to the name information.
In one implementation, the determining, according to the name information, the operation instruction corresponding to the name information includes:
determining an instruction template corresponding to the name information according to the name information;
acquiring an application program name corresponding to the operation event type;
and filling the application program name into the instruction template to generate the operation instruction, wherein the operation instruction is used for operating the application program corresponding to the application program name.
In a second aspect, an embodiment of the present invention further provides a device for executing a voice instruction, where the device includes:
the voice instruction analysis unit is used for acquiring voice operation request information generated based on a voice instruction and determining an operation event type corresponding to the voice instruction according to the voice operation request information;
the operation instruction determining unit is used for determining an operation instruction corresponding to the operation event type according to the operation event type;
and the operation instruction execution unit is used for executing the operation corresponding to the operation instruction according to the operation instruction.
In a third aspect, an embodiment of the present invention further provides an intelligent terminal, where the intelligent terminal includes a memory, a processor, and a voice instruction execution program that is stored in the memory and is executable on the processor, and when the voice instruction execution program is executed by the processor, the steps of the voice instruction execution method according to any one of the above schemes are implemented.
In a fourth aspect, the embodiment of the present invention further provides a computer-readable storage medium, where a speech instruction execution program is stored thereon, and when the speech instruction execution program is executed by a processor, the steps of the speech instruction execution method in any one of the above schemes are implemented.
Has the advantages that: compared with the prior art, the invention provides a voice instruction execution method, which comprises the steps of firstly, obtaining voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information. And then determining an operation instruction corresponding to the operation event type according to the operation event type, and finally executing the operation corresponding to the operation instruction according to the operation instruction. According to the method and the device, the corresponding operation event type is determined according to the voice instruction, then the corresponding operation instruction is generated based on the operation event type, and the voice instruction is executed by executing the operation instruction, so that after the voice instruction is obtained, the operation event type only needs to be determined according to the voice instruction, and then the corresponding operation instruction is determined according to the operation event type.
Drawings
Fig. 1 is a flowchart of a specific implementation of a voice instruction execution method according to an embodiment of the present invention.
Fig. 2 is a flowchart of determining an operation event type in a voice instruction execution method according to an embodiment of the present invention.
Fig. 3 is a flowchart of determining an operation instruction in the voice instruction execution method according to the embodiment of the present invention.
Fig. 4 is a schematic block diagram of a video picture motion device according to an embodiment of the present invention.
Fig. 5 is a schematic block diagram of an internal structure of an intelligent terminal according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and effects of the present invention clearer and clearer, the present invention is further described in detail below with reference to the accompanying drawings and examples. It should be understood that the specific embodiments described herein are merely illustrative of the invention and are not intended to limit the invention.
With the increasingly vigorous natural language processing technology, the technology of voice interaction is more and more mature. The voice assistant is widely applied to various IOT (Internet of things) devices such as mobile phones, televisions and computers, and covers various fields such as movie and television, weather inquiry, device control, shopping and consumption. However, the fields related to voice interaction are various, and in the prior art, voice interaction is basically realized based on a voice assistant, but the recognition accuracy of the voice assistant for user intentions is not high, and in order to recognize new user intentions, a user needs to frequently update the voice assistant to adapt to new requirements, so that the operation is complicated, and the use of the user is influenced. For example, in the prior art, on an intelligent fitness apparatus in a fitness room, a user may often start a corresponding function on the intelligent fitness apparatus by sending a voice message, such as starting a certain fitness mode, but if the voice message spoken by the user cannot be recognized by the intelligent fitness apparatus, the intelligent fitness apparatus cannot respond to the intention of the user (i.e., the voice assistant cannot determine the intention of the user according to the voice message of the user), and thus the operation of the user is affected. In order to enable the intelligent fitness equipment to respond to the voice information of the user, the voice assistant on the intelligent fitness equipment needs to be upgraded, so that the voice assistant can recognize the voice information of the user, and the operation of the user is influenced.
Therefore, the embodiment provides a voice instruction execution method, and by using the method of the embodiment, the intention of the voice information can be accurately determined, so that accurate response to the voice information is realized, and whether the voice assistant needs to be upgraded or not does not need to be considered. In specific implementation, in this embodiment, first, voice operation request information generated based on a voice instruction is obtained, and an operation event type corresponding to the voice instruction is determined according to the voice operation request information. And then determining an operation instruction corresponding to the operation event type according to the operation event type, and finally executing the operation corresponding to the operation instruction according to the operation instruction. In this embodiment, the corresponding operation event type is determined according to the voice instruction, then the corresponding operation instruction is generated based on the operation event type, and the voice instruction is executed by executing the operation instruction, so that in this embodiment, only after the voice instruction is obtained, the operation event type needs to be determined according to the voice instruction, and then the corresponding operation finger is determined according to the operation event type.
For example, the method of the embodiment is applied to a gymnasium scene, and when the intelligent fitness equipment obtains the voice operation request information, the corresponding operation event type can be determined. Since the voice operation request information is generated based on the voice instruction, if the voice instruction of the user is "start jogging function", a request for "start jogging function" is included in the voice operation request information. The intelligent fitness equipment can determine that the event operation type is the event type for starting the running mode, so that the operation instruction can be determined according to the determined operation event type, the operation instruction is the instruction for starting the running function mode, and the intelligent fitness equipment can start the running function mode according to the determined operation instruction so as to meet the requirements of users.
Exemplary method
The voice instruction execution method of the embodiment can be applied to an intelligent terminal, and specifically as shown in fig. 1, the voice instruction execution method specifically includes the following steps:
step S100, voice operation request information generated based on a voice instruction is obtained, and an operation event type corresponding to the voice instruction is determined according to the voice operation request information.
In this embodiment, when a user wants to complete an operation, a voice instruction, which is voice information spoken by the user, is sent to the intelligent terminal, and then, after the intelligent terminal obtains the voice instruction, a voice operation request message is generated according to the voice instruction, where the voice operation request message is a request message for reflecting that the user wants to execute the voice instruction. For example, when the user wants to watch the swordsmen, the user sends a voice command of 'playing the swordsmen' to the smart television, and after receiving the voice command of 'playing the swordsmen', the smart television generates voice operation request information that the user wants to play the swordsmen. After the intelligent terminal receives the voice operation request information, the embodiment can determine the operation event type corresponding to the voice instruction according to the voice operation request information. The operation event type in this embodiment reflects what type of operation the user wants the intelligent terminal to perform. For example, in the above example, the operation event type corresponding to the voice instruction of "playing martial arts drama" is a play event, so that the smart television can determine that the smart television needs to complete the play operation.
In one implementation, as shown in fig. 2, the step S100 specifically includes the following steps:
step S101, analyzing the voice operation request information to obtain voice information corresponding to the voice instruction in the voice operation request information;
and S102, converting the voice information into text information, and determining an operation event type corresponding to the voice instruction according to the text information.
In specific implementation, the recording function is set in the intelligent terminal in this embodiment, and after the user outputs the voice information of "playing martial arts drama", the intelligent terminal can receive the voice information of "playing martial arts drama" by using the recording function, and the voice information is the voice instruction. After the intelligent terminal obtains the voice instruction of playing the swordsmen drama, corresponding voice operation request information can be generated, and the voice operation request information comprises the voice instruction of playing the swordsmen drama. Therefore, when the intelligent terminal in this embodiment determines the voice information according to the voice operation request information, the voice information- "play swordsmen" corresponding to the voice instruction can be obtained only by analyzing the voice operation request information. In order to determine the operation event type corresponding to the voice instruction in this embodiment, after the voice information is obtained, the voice information is converted into text information, and then the operation event type corresponding to the voice instruction is determined according to the text information.
In one implementation, the present embodiment may convert the voice information into the text information by using voice recognition or voice translation. After the text information is obtained, the present embodiment may analyze the text information, and then according to field information in the text information, the field information is some words or words in the text information. Since the text information is a sentence or words recognized by the speech information, and some words have no meaning, for example, when the recognized text information is "please play the hot-playing conutus". In the text information, only two field information, namely 'play' and 'conconututting' play a role in the execution process of the voice instruction, are available, and the intelligent terminal can determine the intention of the user as long as the two field information are obtained. The text message "please" and "hot" has no meaning to the execution process of the voice command. Therefore, when determining the field information according to the text information, the embodiment needs to screen some useless or unpractical fields in the text information, so as to achieve the purpose of more accurately determining the user intention.
After the field information is determined, the present embodiment may determine, according to the field information, an operation event type corresponding to the field information. In this embodiment, the field information is obtained from the text information, and the text information is obtained from the voice information corresponding to the voice command. Therefore, the field information may reflect the intention of the user, for example, the determined field information in the above example is "play" and "conconutct-conutct", and the corresponding user intention is the smart tv play-conutct. And the user intent may reflect the type of operational event. In an implementation manner, the embodiment may preset a mapping file, where the mapping file is provided with a corresponding relationship between field information and an operation event type. Therefore, after the field information is obtained, the field information can be matched with the mapping file to obtain the corresponding operation event type. For example, in the above example, the determined field information is "play" and "conconutu transmission", the present embodiment can determine that the operation event type matching with "play" is "play event" according to "play" in the field information, and further can determine that "movie drama" matching with "conutu transmission" is "movie drama" according to "conutu transmission" in the field information, and then combine the determined play event with the movie drama, that is, it can be determined that the final play event type is: and playing the movie and television play.
In one implementation, the mapping file in this embodiment may be set based on a user's historical usage record. For example, for an intelligent fitness apparatus, the functions used by the user are basically: running, rope skipping counting/timing, and playing of body-building teaching videos. Therefore, for the intelligent fitness equipment, the set mapping file can set the corresponding relation between the field information and the operation event type according to the previous history of the user. In addition, the present embodiment may further prioritize the correspondence relationship between each field information and the operation event type according to the use frequency of the user. If the voice instruction output by the user may include multiple intentions, the determined field information may also be multiple, the corresponding operation event types may also be multiple, and the user may have only one true intention. In order to better fulfill the intention of the user, the determined operation event types may be sorted according to the field information and the priority of the operation event types, and after a plurality of candidate operation event types are obtained, a higher priority may be used as a final operation event type based on the priority of the candidate operation event types. For example, when the determined field information includes "play", "conutting", "volume", and "turn up", the corresponding candidate operation event type is: playing movie and TV play and turning up volume. And then, according to the priorities of the two candidate operation event types, determining that the priority of playing the movie is higher than the priority of turning up the volume, so that the played movie is taken as the final operation event type so as to realize the real intention of the user. Of course, if the user does want to achieve multiple user intentions in the voice command, this embodiment may also sort the determined operation event types according to the priority levels according to the field information and the operation event types, and then sort according to each operation event type, so that the voice command is executed in the order of priority levels in the subsequent steps. For example, when the determined field information includes "play", "conutting", "volume", and "turn up", the corresponding operation event type is: playing movie and TV play and turning up volume. And then, according to the priorities of the two operation event types, the priority for playing the movie is determined to be higher than the priority for turning up the volume, so that the intelligent terminal can firstly execute the event for playing the movie and then execute the event for turning up the volume in the subsequent steps.
In addition, in order to satisfy more application scenarios, the present embodiment may further rewrite the field information, that is, customize the field information, so that the operation type event may be customized. In concrete implementation, when the field information is rewritten, the present embodiment can rewrite the field information obtained from the voice information to the specified field information, i.e., the intention information intended by the user. For example, in the above example, the field information obtained according to the voice information is "play" and "conututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututututut. In order to meet the requirement of the user for viewing, the present embodiment may perform adaptive rewriting according to the determined field information, that is, rewriting one or more of the field information into information associated with the original field information, that is, specifying the field information, then determining a specified operation event type according to the specified field information, and taking the specified operation event type as the operation event type. For example, in this embodiment, the "conutting transmission" may be rewritten into "exemplary transmission", because on the network, the associated entries of the two drams of "conutting transmission" and "exemplary transmission" are the most, so the field information is changed into the specified field information "exemplary transmission", the obtained operation event type is the broadcast drama, and thus, the normal viewing of the user can be satisfied.
And step S200, determining an operation instruction corresponding to the operation event type according to the operation event type.
In this embodiment, the determining the type of the operation event is to more accurately determine to execute the voice command. After the operation event type is obtained, the embodiment may determine, according to the operation event type, an operation instruction corresponding to the operation event type, where the operation instruction is used to complete a user intention corresponding to the operation event type, that is, a user intention realized by the operation instruction is a user intention in the voice instruction.
In one implementation, as shown in fig. 3, the step S200 specifically includes the following steps:
step S201, determining name information of the operation event type according to the operation event type;
and step S202, determining the operation instruction corresponding to the name information according to the name information.
In specific implementation, after the operation event type is determined, the name information corresponding to the operation event type may be obtained, and then a corresponding instruction template is called according to the name information, where the instruction template is a template for generating an instruction. Therefore, a plurality of instruction templates can be preset in the embodiment, so that the corresponding instruction template can be determined according to the name information. Since the operation event type in this embodiment is obtained from the voice command of the user, the user intention embodied in the embodiment is that the intelligent terminal is expected to execute the corresponding operation after receiving the voice command. Therefore, the operation instruction generated in this embodiment is also for controlling the intelligent terminal to execute the operation corresponding to the voice instruction. Therefore, in this embodiment, after obtaining the instruction template, the name of the application program corresponding to the operation event type is obtained, that is, the name of the application program to be operated by the voice instruction of the user is obtained. Then, the present embodiment fills the application name into the instruction template to generate the operation instruction, where the operation instruction is used to operate the application corresponding to the application name. For example, in the above example, when the determined operation event type is a play movie event, the obtained instruction template plays an instruction template of the movie, then the name information of the corresponding application program is obtained as a youku APP, and then the obtained instruction template is filled into the instruction template for playing the movie, so that an operation instruction can be generated, the youku APP can be opened through the operation instruction, and the movie-conkout can be played.
And step S300, executing the operation corresponding to the operation instruction according to the operation instruction.
After the operation instruction is obtained, the intelligent terminal may execute a corresponding operation according to the operation instruction, for example, in the above example, when the determined operation event type is a play movie event, the obtained instruction template plays an instruction template of the movie, then the obtained name information of the corresponding application program is a kuku APP, and then the obtained name information is filled into the instruction template for playing the movie, so as to generate the operation instruction, the kuku APP can be opened through the operation instruction, and the movie-screen upload is played, thereby satisfying the movie viewing requirement of the user. Of course, the voice instruction execution method of the embodiment may be applied to other fields, such as recommendation of fitness mode in an intelligent fitness apparatus applied in a fitness room.
In summary, in the present embodiment, first, voice operation request information generated based on a voice instruction is obtained, and an operation event type corresponding to the voice instruction is determined according to the voice operation request information. And then determining an operation instruction corresponding to the operation event type according to the operation event type, and finally executing the operation corresponding to the operation instruction according to the operation instruction. In this embodiment, the corresponding operation event type is determined according to the voice instruction, then the corresponding operation instruction is generated based on the operation event type, and the voice instruction is executed by executing the operation instruction, so that in this embodiment, only after the voice instruction is obtained, the operation event type needs to be determined according to the voice instruction, and then the corresponding operation finger is determined according to the operation event type.
Exemplary device
As shown in fig. 4, an embodiment of the present invention provides a video frame dynamic adjustment apparatus, including: a voice instruction analysis unit 10, an operation instruction determination unit 20, and an operation instruction execution unit 30. Specifically, the voice instruction analysis unit 10 is configured to acquire voice operation request information generated based on a voice instruction, and determine an operation event type corresponding to the voice instruction according to the voice operation request information. The operation instruction determining unit 20 is configured to determine, according to the operation event type, an operation instruction corresponding to the operation event type. The operation instruction execution unit 30 is configured to execute an operation corresponding to the operation instruction according to the operation instruction.
In one implementation, the voice instruction analysis unit 10 includes:
the voice information determining subunit is configured to analyze the voice operation request information to obtain voice information corresponding to the voice instruction in the voice operation request information;
and the operation event type determining subunit is used for converting the voice information into text information and determining the operation event type corresponding to the voice instruction according to the text information.
In one implementation, the operation instruction determination unit 20 includes:
the name information determining subunit is used for determining the name information of the operation event type according to the operation event type;
and the operation instruction determining subunit is used for determining the operation instruction corresponding to the name information according to the name information.
Based on the above embodiment, the present invention further provides an intelligent terminal, and a schematic block diagram thereof may be as shown in fig. 5. The intelligent terminal comprises a processor, a memory, a network interface, a display screen and a temperature sensor which are connected through a system bus. Wherein, the processor of the intelligent terminal is used for providing calculation and control capability. The memory of the intelligent terminal comprises a nonvolatile storage medium and an internal memory. The non-volatile storage medium stores an operating system and a computer program. The internal memory provides an environment for the operation of an operating system and computer programs in the non-volatile storage medium. The network interface of the intelligent terminal is used for being connected and communicated with an external terminal through a network. The computer program is executed by a processor to implement a method of voice instruction execution. The display screen of the intelligent terminal can be a liquid crystal display screen or an electronic ink display screen, and the temperature sensor of the intelligent terminal is arranged inside the intelligent terminal in advance and used for detecting the operating temperature of internal equipment.
It will be understood by those skilled in the art that the block diagram of fig. 5 is only a block diagram of a part of the structure related to the solution of the present invention, and does not constitute a limitation to the intelligent terminal to which the solution of the present invention is applied, and a specific intelligent terminal may include more or less components than those shown in the figure, or combine some components, or have different arrangements of components.
In one embodiment, an intelligent terminal is provided that includes a memory, and one or more programs, wherein the one or more programs are stored in the memory and configured to be executed by the one or more processors, the one or more programs including instructions for:
acquiring voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information;
determining an operation instruction corresponding to the operation event type according to the operation event type;
and executing the operation corresponding to the operation instruction according to the operation instruction.
It will be understood by those skilled in the art that all or part of the processes of the methods of the embodiments described above can be implemented by hardware instructions of a computer program, which can be stored in a non-volatile computer-readable storage medium, and when executed, can include the processes of the embodiments of the methods described above. Any reference to memory, storage, databases, or other media used in embodiments provided herein may include non-volatile and/or volatile memory. Non-volatile memory can include read-only memory (ROM), Programmable ROM (PROM), Electrically Programmable ROM (EPROM), Electrically Erasable Programmable ROM (EEPROM), or flash memory. Volatile memory can include Random Access Memory (RAM) or external cache memory. By way of illustration and not limitation, RAM is available in a variety of forms such as Static RAM (SRAM), Dynamic RAM (DRAM), Synchronous DRAM (SDRAM), Double Data Rate SDRAM (DDRSDRAM), Enhanced SDRAM (ESDRAM), Synchronous Link DRAM (SLDRAM), Rambus Direct RAM (RDRAM), direct bus dynamic RAM (DRDRAM), and memory bus dynamic RAM (RDRAM).
In summary, the present invention discloses a voice instruction execution method, apparatus, intelligent terminal and storage medium, the method comprising: acquiring voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information; determining an operation instruction corresponding to the operation event type according to the operation event type; and executing the operation corresponding to the operation instruction according to the operation instruction. The method and the device determine the corresponding operation event type according to the voice instruction, and then generate the operation instruction according to the operation event type so as to execute the operation instruction, thereby realizing that the operation of the voice instruction is completed by acquiring the corresponding operation event type, conveniently, quickly and accurately executing the voice instruction, and the updating of voice information is not required to be considered, thereby providing convenience for users.
Finally, it should be noted that: the above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; and such modifications or substitutions do not depart from the spirit and scope of the corresponding technical solutions of the embodiments of the present invention.

Claims (10)

1. A method for voice instruction execution, the method comprising:
acquiring voice operation request information generated based on a voice instruction, and determining an operation event type corresponding to the voice instruction according to the voice operation request information;
determining an operation instruction corresponding to the operation event type according to the operation event type;
and executing the operation corresponding to the operation instruction according to the operation instruction.
2. The method of claim 1, wherein the obtaining voice operation request information generated based on the voice instruction and determining the operation event type corresponding to the voice instruction according to the voice operation request information comprises
Analyzing the voice operation request information to obtain voice information corresponding to the voice instruction in the voice operation request information;
and converting the voice information into text information, and determining the operation event type corresponding to the voice instruction according to the text information.
3. The method according to claim 2, wherein the determining the operation event type corresponding to the voice instruction according to the text information comprises:
analyzing the text information to obtain field information in the text information;
and determining the type of the operation event corresponding to the field information according to the field information.
4. The method according to claim 3, wherein the determining the type of the operation event corresponding to the field information according to the field information comprises:
acquiring a candidate operation event type matched with the field information according to the field information;
and acquiring the priority information of the candidate operation event types, and determining the operation event types corresponding to the field information according to the priority information.
5. The method according to claim 3, wherein the determining the type of the operation event corresponding to the field information according to the field information comprises:
rewriting the field information into specified field information;
and generating a specified operation event type corresponding to the specified field information according to the specified field information, and taking the specified operation event type as the operation event type.
6. The method according to claim 1, wherein the determining the operation command corresponding to the operation event type according to the operation event type includes:
determining the name information of the operation event type according to the operation event type;
and determining the operation instruction corresponding to the name information according to the name information.
7. The method according to claim 6, wherein the determining the operation command corresponding to the name information according to the name information comprises:
determining an instruction template corresponding to the name information according to the name information;
acquiring an application program name corresponding to the operation event type;
and filling the application program name into the instruction template to generate the operation instruction, wherein the operation instruction is used for operating the application program corresponding to the application program name.
8. A voice instruction execution apparatus, comprising:
the voice instruction analysis unit is used for acquiring voice operation request information generated based on a voice instruction and determining an operation event type corresponding to the voice instruction according to the voice operation request information;
the operation instruction determining unit is used for determining an operation instruction corresponding to the operation event type according to the operation event type;
and the operation instruction execution unit is used for executing the operation corresponding to the operation instruction according to the operation instruction.
9. An intelligent terminal, characterized in that the intelligent terminal comprises a memory, a processor and a voice instruction execution program stored on the memory and operable on the processor, and when the voice instruction execution program is executed by the processor, the steps of the voice instruction execution method according to any one of claims 1 to 7 are realized.
10. A computer-readable storage medium, having stored thereon a program of execution of voice instructions, which when executed by a processor, performs the steps of the method of execution of voice instructions according to any one of claims 1 to 7.
CN202011423206.7A 2020-12-08 2020-12-08 Voice instruction execution method and device, intelligent terminal and storage medium Pending CN114664296A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011423206.7A CN114664296A (en) 2020-12-08 2020-12-08 Voice instruction execution method and device, intelligent terminal and storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011423206.7A CN114664296A (en) 2020-12-08 2020-12-08 Voice instruction execution method and device, intelligent terminal and storage medium

Publications (1)

Publication Number Publication Date
CN114664296A true CN114664296A (en) 2022-06-24

Family

ID=82024288

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011423206.7A Pending CN114664296A (en) 2020-12-08 2020-12-08 Voice instruction execution method and device, intelligent terminal and storage medium

Country Status (1)

Country Link
CN (1) CN114664296A (en)

Similar Documents

Publication Publication Date Title
CN111460150B (en) Classification model training method, classification method, device and storage medium
KR101777392B1 (en) Central server and method for processing of voice of user
US10846460B2 (en) Method and apparatus for launching application page, and electronic device
KR102451437B1 (en) Techniques for updating language understanding classifier models for digital personal assistants based on crowdsourcing
RU2710984C2 (en) Performing task without monitor in digital personal assistant
US8959536B2 (en) Method and system for providing applications to various devices
US10388277B1 (en) Allocation of local and remote resources for speech processing
US9622016B2 (en) Invisiblemask: a tangible mechanism to enhance mobile device smartness
US11024300B2 (en) Electronic device and control method therefor
CN107515870B (en) Searching method and device and searching device
CN113395538A (en) Sound effect rendering method and device, computer readable medium and electronic equipment
WO2018040106A1 (en) Method and apparatus for outputting search result, and electronic device
CN112035651B (en) Sentence completion method, sentence completion device and computer readable storage medium
CN114664296A (en) Voice instruction execution method and device, intelligent terminal and storage medium
US20080281852A1 (en) Method, apparatus and computer program product for providing an adaptive framework for a metadata-context switch
CN116757208A (en) Data processing method, device and equipment
WO2023093280A1 (en) Speech control method and apparatus, electronic device, and storage medium
US20170116174A1 (en) Electronic word identification techniques based on input context
CN109326291A (en) Television mode control method, TV and storage medium
CN112052377B (en) Resource recommendation method, device, server and storage medium
CN112770157A (en) Voice control method, device, equipment and medium for WEB front-end interface of television
CN113573132A (en) Multi-application screen splicing method and device based on voice realization and storage medium
CN112380871A (en) Semantic recognition method, apparatus, and medium
CN113345452A (en) Voice conversion method, training method, device and medium of voice conversion model
CN114724555A (en) Food replacement prompting method and device based on voice instruction, terminal equipment and storage medium

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination