CN108831469B - Voice command customizing method, device and equipment and computer storage medium - Google Patents

Voice command customizing method, device and equipment and computer storage medium Download PDF

Info

Publication number
CN108831469B
CN108831469B CN201810887444.XA CN201810887444A CN108831469B CN 108831469 B CN108831469 B CN 108831469B CN 201810887444 A CN201810887444 A CN 201810887444A CN 108831469 B CN108831469 B CN 108831469B
Authority
CN
China
Prior art keywords
voice command
voice
user
operation instruction
information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201810887444.XA
Other languages
Chinese (zh)
Other versions
CN108831469A (en
Inventor
韦泽光
张玉
陈琳婷
杨煜豪
程万里
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Gree Electric Appliances Inc of Zhuhai
Original Assignee
Gree Electric Appliances Inc of Zhuhai
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Gree Electric Appliances Inc of Zhuhai filed Critical Gree Electric Appliances Inc of Zhuhai
Priority to CN201810887444.XA priority Critical patent/CN108831469B/en
Publication of CN108831469A publication Critical patent/CN108831469A/en
Priority to PCT/CN2018/121040 priority patent/WO2020029500A1/en
Application granted granted Critical
Publication of CN108831469B publication Critical patent/CN108831469B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • User Interface Of Digital Computer (AREA)

Abstract

The invention discloses a voice command customization method, a voice command customization device, voice command customization equipment and a computer storage medium, which are used for realizing the customization of a voice control scheme and improving the use experience of a user. The method comprises the following steps: generating a first voice command according to the collected voice information; outputting first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment, and generating a first operation instruction for executing an operation step in the demonstration operation process based on the demonstration operation; and establishing an association relation between the first voice command and the first operation instruction, and storing the association relation so that the first operation instruction is executed when the voice command matched with the first voice command is received.

Description

Voice command customizing method, device and equipment and computer storage medium
Technical Field
The present invention relates to the field of computer technologies, and in particular, to a method, an apparatus, and a device for customizing a voice command and a computer storage medium.
Background
At present, smart home devices provide more convenience for life of people, and control methods for smart home devices are increasing, for example, a user may control smart home devices through voice or control smart home devices through an Application (APP) installed on a terminal. The mode of controlling through voice is usually that the association relationship between the voice library and the functions is predefined when the smart home device leaves the factory, and the user can perform corresponding voice control according to the instruction or the prompt.
Therefore, in the existing voice control scheme, both the voice and the function instruction set are predefined, but the use habits of different users have certain differences, and the predefined voice control scheme may not meet the use habits of the users, so that the use experience of the users when controlling the smart home devices through voice is obviously poor, and the predefined voice control scheme has limited functions, which may not meet the use requirements of the users, and further reduces the use experience of the users.
Disclosure of Invention
The embodiment of the invention provides a voice command customization method, a voice command customization device, a voice command customization equipment and a computer storage medium, which are used for realizing the customization of a voice control scheme and improving the use experience of a user.
In a first aspect, a method for customizing a voice command is provided, the method comprising:
generating a first voice command according to the collected voice information;
outputting first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment, and generating a first operation instruction for executing an operation step in the demonstration operation process based on the demonstration operation;
and establishing an association relation between the first voice command and the first operation instruction, and storing the association relation so that the first operation instruction is executed when the voice command matched with the first voice command is received.
In the embodiment of the invention, the voice command can be generated by collecting the voice information of the user, the corresponding operation instruction can be generated according to the demonstration operation of the user, and the voice command and the corresponding operation instruction are associated, so that the voice command is generated according to the voice information of the user, the speaking habit of the user can be further met, the operation instruction obtained based on the demonstration operation of the user can not be limited to the voice control instruction when the intelligent household equipment leaves a factory, and the application range is wider.
Optionally, the generating a first voice command according to the collected voice information includes:
analyzing according to the collected pieces of voice information to extract common features of the pieces of voice information;
generating the first voice command based on the common characteristic.
Optionally, the establishing an association relationship between the first voice command and the first operation instruction includes:
executing the first operation instruction;
determining whether a result of executing the first operation instruction meets the requirement of a user or not based on the received first feedback information;
and when the first feedback information shows that the result of executing the first operation instruction meets the requirement of the user, storing the first voice command after associating the first voice command with the first operation instruction, otherwise, prompting the user to perform demonstration operation again.
In the embodiment of the invention, after the operation instruction is generated, the operation instruction is executed once, and then the operation instruction is verified to confirm whether the function realized by the operation instruction meets the expected effect of a user.
Optionally, the method further includes:
carrying out semantic recognition on the collected voice information and outputting a semantic recognition result;
determining whether the semantic recognition result is the semantic expressed by the collected voice information based on the received second feedback information;
and if the result is yes, associating the first voice command with the semantic recognition result, otherwise, associating the received user-modified semantic with the first voice command, wherein the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same.
In the embodiment of the invention, the voice information input by the user can be subjected to semantic recognition, and the voice command generated according to the voice information is associated with the semantic recognition result, so that all instructions associated with the same semantic recognition result can realize the same function, and even if the user inputs dialects or other languages, the instructions can be supported, and the generalization capability of voice recognition is improved.
Optionally, the method further includes:
when the received second voice command is not successfully matched with any voice command in the stored voice command library, but the similarity between the received second voice command and at least one voice command in the voice command library is not less than a first preset threshold value, recording the second voice command;
prompting the user to input the voice command again;
and if the re-input voice command is successfully matched, updating the voice command successfully matched with the re-input voice command according to the second voice command.
In the embodiment of the invention, if the voice command input for the first time by the user is not recognized and the voice command input for the second time is successfully recognized, the voice command library can be updated according to the voice command input for the first time, so that the recognition capability of the voice command library is improved.
Optionally, the method further includes:
when the voice commands received for multiple times are not successfully matched with any voice command in the voice command library, outputting second prompt information for prompting the user whether to set the associated operation instruction;
and when the received third feedback information indicates that an operation instruction associated with the fourth voice command needs to be set, outputting the first prompt information.
In a second aspect, a voice command customizing apparatus is provided, including:
the generating unit is used for generating a first voice command according to the collected voice information;
the output unit is used for outputting first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment;
the generating unit is further used for generating a first operation instruction used for executing an operation step in the demonstration operation process based on the demonstration operation;
and the association unit is used for establishing an association relationship between the first voice command and the first operation instruction and storing the association relationship so as to execute the first operation instruction when the voice command matched with the first voice command is received.
Optionally, the generating unit is specifically configured to:
analyzing according to the collected pieces of voice information to extract common features of the pieces of voice information;
generating the first voice command based on the common characteristic.
Optionally, the association unit is specifically configured to:
executing the first operation instruction;
determining whether a result of executing the first operation instruction meets the requirement of a user or not based on the received first feedback information;
and when the first feedback information shows that the result of executing the first operation instruction meets the requirement of the user, storing the first voice command after associating the first voice command with the first operation instruction, otherwise, prompting the user to perform demonstration operation again.
Optionally, the apparatus further includes a semantic recognition unit, configured to:
carrying out semantic recognition on the collected voice information and outputting a semantic recognition result;
determining whether the semantic recognition result is the semantic expressed by the collected voice information based on the received second feedback information;
and if the result is yes, associating the first voice command with the semantic recognition result, otherwise, associating the received user-modified semantic with the first voice command, wherein the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same.
Optionally, the apparatus further includes an updating unit, configured to:
when the received second voice command is not successfully matched with any voice command in the stored voice command library, but the similarity between the received second voice command and at least one voice command in the voice command library is not less than a first preset threshold value, recording the second voice command;
prompting the user to input the voice command again;
and if the re-input voice command is successfully matched, updating the voice command successfully matched with the re-input voice command according to the second voice command.
Optionally, the output unit is further configured to:
when the voice commands received for multiple times are not successfully matched with any voice command in the voice command library, outputting second prompt information for prompting the user whether to set the associated operation instruction;
and when the received third feedback information indicates that an operation instruction associated with the fourth voice command needs to be set, outputting the first prompt information.
In a third aspect, there is provided a voice command customizing apparatus comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of the first aspect.
In a fourth aspect, a computer storage medium is provided,
the computer storage medium stores computer instructions which, when executed on a computer, cause the computer to perform the method of the first aspect.
Drawings
FIG. 1 is a flow chart illustrating a method for customizing a voice command according to an embodiment of the present invention;
FIG. 2 is a schematic flow chart illustrating a semantic recognition result associated with a voice command according to an embodiment of the present invention;
FIG. 3 is a flow chart of a voice control process according to an embodiment of the present invention;
FIG. 4 is a schematic structural diagram of a voice command customizing apparatus according to an embodiment of the present invention;
fig. 5 is a schematic structural diagram of a voice command customizing apparatus according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the embodiments of the present invention clearer, the technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention.
The technical background of the embodiments of the present invention is described below.
At present, voice and function instruction sets in a voice control scheme are predefined, but the use habits of different users have certain differences, and the predefined voice control scheme may not meet the use habits of the users, so that the use experience of the users when the users control the smart home devices through voice is obviously poor, and the functions included in the predefined voice control scheme are limited, which may not meet the use requirements of the users, and further reduces the use experience of the users.
In view of this, embodiments of the present invention provide a method, an apparatus, a device, and a computer storage medium for customizing a voice command, in which a voice command may be generated by collecting voice information of a user, and a corresponding operation instruction may be generated according to a demonstration operation of the user, and the two are associated with each other, so that the voice command may be generated according to the voice information of the user, and further may better conform to a speaking habit of the user, and an operation instruction obtained based on the demonstration operation of the user may not be limited to a voice control instruction when the smart home device leaves a factory, and the application range is wider.
The technical scheme provided by the embodiment of the invention is described below by combining the accompanying drawings.
Referring to fig. 1, an embodiment of the present invention provides a voice command customization method, which may be executed by a device (hereinafter, referred to as a device for short) provided with a voice control module, such as a smart home device or a terminal, where the smart home device may be, for example, a smart air conditioner, a smart gas stove, a smart television, or a smart refrigerator, and the terminal may be, for example, a mobile phone or a tablet computer (PAD), and the terminal may be a device that is installed with an APP for controlling the smart home device. The flow of the method is described below.
Step 101: and generating a first voice command according to the collected voice information.
In the embodiment of the invention, when a user wants to customize a new voice control scheme, the new voice control scheme can be newly established in the intelligent home equipment or APP, the intelligent home equipment or APP can provide a visual operation interface for the user, and the user operates step by step according to prompts on the operation interface to complete the customization process.
When a new voice control scheme is customized, a new voice command needs to be generated, so that a user can be prompted to input voice information, and after the user inputs the voice information, the voice information input by the user can be collected by the intelligent home equipment or the terminal provided with the APP. Specifically, a microphone is generally arranged in the terminal of the intelligent home equipment or the terminal of the installation APP, so that the voice information input by the user can be collected through the microphone by the terminal of the intelligent home equipment or the terminal of the installation APP.
After the voice information input by the user is collected, a first voice command can be generated according to the voice information input by the user.
In the embodiment of the invention, as the tone or content of each speaking of the user is more or less different, in order to improve the recognition rate of the generated voice command, the voice information input by the user can be collected for multiple times, and the voice information collected for multiple times is analyzed to extract the common characteristics of the voice information, so as to generate the first voice command according to the common characteristics. For example, the collected voice information input by the user for 3 times is respectively 'turn on the air conditioner', 'help me turn on the air conditioner' and 'please turn on the air conditioner', then the common characteristic input for 3 times is 'turn on the air conditioner', therefore, the section of voice of 'turn on the air conditioner' input by the user can be used as the common characteristic and used as the first voice command. Or after the voice information of the user is collected, vector features in the voice information are respectively extracted through a voice recognition model, the vector features extracted for multiple times are compared to obtain a common feature, and the first voice command is generated based on the common feature.
In practical application, after the voice information input by the user is collected, the user can be prompted to input the voice information next time. Specifically, the prompt may be performed through text information, or may also be performed through voice.
Step 102: and outputting the first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment.
In the embodiment of the invention, after the first voice command is generated, first prompt information can be output to prompt a user to perform demonstration operation. The first prompt message may be output in the form of text message, for example, a word "please perform a demonstration operation" may be displayed on the display unit; or, the first prompt information may also be output in a voice manner, for example, the "please perform the demonstration operation" may be output through a speaker included in the smart home device or the terminal in which the APP is installed; of course, the output may be performed by combining the above two modes.
Step 103: and generating a first operation instruction for executing an operation step in the demonstration operation process based on the demonstration operation.
In the embodiment of the invention, the demonstration operation performed by the user is used for demonstrating the operation steps required for realizing at least one function of the smart home equipment, so that after the demonstration operation is collected by the equipment, the operation steps included in the demonstration operation and the sequence relation among the operation steps can be obtained, and the first operation instruction is generated based on the operation steps and the sequence relation among the operation steps.
For the equipment, each operation step corresponds to one function or one function instruction of the intelligent household equipment. For example, if a user performs operations such as "turn on an air conditioner", "adjust an air outlet intensity", and "adjust a wind direction" in sequence during a demonstration operation, for a device, functional instructions corresponding to the operations and a sequential relationship between the functional instructions may be obtained, and then the operations are performed in sequence based on the generated first operation instruction. Of course, the function instruction may also be other possible function instructions, such as gear adjustment of the smart home device, or page jump of the APP, and the like.
In the specific implementation process, the sequence of step 101 and step 103 may also be switched, that is, step 103 is executed first, and then step 101 is executed, then step 102 is used to prompt the user to record the voice command.
Step 104: and executing the first operation instruction.
Step 105: and determining whether the result of executing the first operation instruction meets the requirement of the user.
In the embodiment of the present invention, after the first operation instruction is generated, the first operation instruction may be checked to check whether the first operation instruction can implement the function that the user wants to implement, so that the first operation instruction may be executed once, and then after the execution is completed, the prompt information is output to allow the user to confirm whether the user expects the first operation instruction. After the user performs feedback, the first feedback information may be received, and it is determined whether the result of the first operation instruction meets the requirement of the user based on the first feedback information.
Step 106: if the determination result in the step 105 is yes, establishing an association relationship between the first voice command and the first operation instruction, and storing the association relationship.
In the embodiment of the invention, if the first feedback information indicates that the result of executing the first operation instruction meets the requirement, the first voice command and the first operation instruction can be associated to generate a new voice control scheme, and the new voice control scheme is stored, so that in the using process of the intelligent home equipment, if the first voice command input by the user is received again, the first operation instruction associated with the first voice command can be searched, and the first operation instruction is executed to realize the function realized by the user demonstration operation.
Specifically, the association relationship between the voice command and the operation instruction may be stored in a storage unit included in the smart home device, so that the smart home device can smoothly complete voice control even in the absence of a network, or the association relationship between the voice command and the operation instruction may be stored in a server, so that the smart home device or the APP may obtain the operation instruction associated with the voice command input by the user from the server. For example, when the user controls the smart home device through the APP, the voice command input by the user may be sent to the server, and after the server matches the associated operation instruction, the operation instruction is sent to the smart home device, so that voice control is realized; or, the APP can also send the voice command input by the user to the server side, the server side feeds back the matched operation instruction to the APP, and the APP sends the operation instruction to the smart home device to realize voice control.
Step 107: and if the determination result in the step 105 is negative, prompting the user to perform the demonstration operation again.
In the embodiment of the invention, if the first feedback information indicates that the result of executing the first operation instruction is not in accordance with the requirement, the user can be prompted to perform the demonstration operation again, and then a new first operation instruction is generated based on the repeated demonstration operation.
Referring to fig. 2, after receiving the voice information input by the user, the method further includes the following steps:
step 201: and carrying out semantic recognition on the collected voice information and outputting a semantic recognition result.
In the embodiment of the invention, the geographic positions of the users are different, and the types of languages used in daily life are different, for example, Shanghai people may be more accustomed to speaking Shanghai dialect, and Sichuan people may be more accustomed to speaking Sichuan dialect, so that the users may input dialects when customizing a new voice control scheme, and the semantics corresponding to different dialects may be the same, so that the corresponding operation instructions are the same. Therefore, in the embodiment of the invention, after the voice information input by the user is received, the semantic recognition can be carried out on the voice information, and the semantic recognition result is output to the user, so that the user can confirm whether the semantic recognition result is correct. Generally, the semantic recognition result is generally described in a general language, for example, in mandarin. Specifically, when outputting, the semantic recognition result may be displayed on a display unit in a text form, or played through a speaker in a voice manner.
Step 202: and determining whether the semantic recognition result is the semantic expressed by the collected voice information.
In the embodiment of the invention, the user can feed back based on the output semantic recognition result, correspondingly, the equipment can receive the second feedback information input by the user and determine whether the semantic recognition result is the semantic expressed by the voice information input by the user based on the second feedback information.
Step 203: if the determination of step 202 is yes, the first voice command is associated with the semantic recognition result.
In the embodiment of the invention, if the semantic recognition result is the semantic expressed by the voice information input by the user, that is, the semantic recognition result is correct, the first voice command and the semantic recognition result can be associated, and the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same. In this way, even if the voice commands generated according to different dialects are different, if the semantic recognition results associated with the voice commands are the same, the corresponding operation instructions are also the same, so that the voice control can simultaneously support dialects, living spoken languages and the like.
Step 204: if the determination of step 202 is negative, the received user-modified semantics are associated with the first voice command.
In the embodiment of the invention, if the semantic recognition result is not the semantic expressed by the voice information input by the user, the user can be prompted to input correct semantic. The user can modify the original semantic recognition result or input correct semantic by himself, and then the device associates the received semantic modified by the user with the first voice command.
In the embodiment of the present invention, after the association relationship between the voice command and the operation instruction is stored, the association relationship may be applied to voice control. Please refer to fig. 3, which is a schematic flow chart of controlling the smart home device by voice.
Step 301: a second voice command input by the user is received.
Step 302: it is determined whether there is a voice command in the library of voice commands that matches the second voice command.
When the user wants to control the smart home device through voice, for example, the user needs to input a second voice command for a function that the user wants to implement, the user may input the second voice command, and accordingly, the device may receive the second voice command and match the second voice command with the voice command library to determine whether a voice command matching the second voice command exists in the voice command library. Specifically, when the second voice command is matched with the voice command library, all features included in the second voice command may be compared with features included in each voice command in the voice command library.
Step 303: if the determination result in step 302 is yes, the operation instruction associated with the voice command matching the second voice command is executed.
Specifically, when the device is an intelligent home device, the intelligent home device may directly execute an operation instruction associated with the voice command matched with the second voice command; when the device is a terminal for installing the APP, an operation instruction associated with the voice command matched with the second voice command may be sent to the smart home device, so that the smart home device executes the operation instruction.
Step 304: if the determination result in step 302 is negative, the user is prompted to re-input the voice command.
In the embodiment of the invention, if the second voice command input by the user fails to be matched with the voice command library, namely the second voice command is not recognized successfully, the equipment can be prompted to input the voice command again.
Specifically, over time, the user may forget the voice command that was input when the voice control scheme was customized, but remember the approximate content, and thus the second voice command may not be able to successfully match any of the voice commands in the library of voice commands, but the similarity of the second voice command to one or more of the voice commands in the library of voice commands is not less than the first preset threshold, and the device may temporarily save the second voice command. Of course, the device may also output the semantic recognition result associated with the one or more voice commands to the user, so that the user may confirm the voice command to be input, and execute the corresponding operation instruction after the user selects and confirms the voice command.
Step 305: it is determined whether there is a voice command in the voice command library that matches the re-entered voice command.
Step 306: if the determination of step 305 is yes, the voice command matching the re-entered voice command is updated according to the second voice command.
In the embodiment of the invention, if the matching of the voice command re-input by the user is successful and the similarity between the second voice command input for the first time and one or more voice commands in the voice command library is not less than the first preset threshold, which indicates that the semantics expressed by the second voice command input for the first time by the user and the voice command re-input by the user should be the same, the voice command successfully matched with the voice command re-input can be updated according to the second voice command, so that the voice command can be strengthened and has stronger compatibility. Specifically, the second voice command may be compared with the voice command successfully matched with the re-input voice command, and the common features of the two voice commands may be extracted and stored.
In the embodiment of the invention, if the voice commands input by the user for multiple times are not matched successfully, the voice control scheme associated with the voice command is not indicated, the second prompt information can be output, the second prompt information is used for prompting the user whether to set the operation instruction associated with the voice command, when the user feeds back the operation instruction required to be set, the first prompt information is output to prompt the user to perform operation demonstration, and after the user performs the demonstration operation, the newly generated operation instruction is associated with the common characteristic in the voice commands input for multiple times continuously, so as to obtain a new voice control scheme.
In summary, the voice command can be generated by collecting the voice information of the user, the corresponding operation instruction can be generated according to the demonstration operation of the user, and the voice command and the corresponding operation instruction are associated, so that the voice command is generated according to the voice information of the user, the speaking habit of the user can be further met, the operation instruction obtained based on the demonstration operation of the user can not be limited to the voice control instruction when the smart home device leaves the factory, and the application range is wider.
Referring to fig. 4, based on the same inventive concept, an embodiment of the present invention provides a voice command customizing apparatus, including:
a generating unit 401, configured to generate a first voice command according to the collected voice information;
the output unit 402 is configured to output first prompt information to instruct a user to input a demonstration operation for implementing at least one function of the smart home device;
a generating unit 401, configured to generate a first operation instruction for executing an operation step in the process of the demonstration operation based on the demonstration operation;
an association unit 403, configured to establish an association relationship between the first voice command and the first operation instruction, and store the association relationship, so that when a voice command matching the first voice command is received, the first operation instruction is executed.
Optionally, the generating unit 401 is specifically configured to:
analyzing according to the collected multiple pieces of voice information to extract common features of the multiple pieces of voice information;
a first voice command is generated based on the common characteristic.
Optionally, the associating unit 403 is specifically configured to:
executing a first operation instruction;
determining whether a result of executing the first operation instruction meets the requirement of the user or not based on the received first feedback information;
and when the first feedback information shows that the result of executing the first operation instruction meets the requirement of the user, storing the first voice command after associating the first voice command with the first operation instruction, otherwise, prompting the user to perform demonstration operation again.
Optionally, the apparatus further includes a semantic recognition unit 404, configured to:
carrying out semantic recognition on the collected voice information and outputting a semantic recognition result;
determining whether the semantic recognition result is the semantic expressed by the collected voice information based on the received second feedback information;
and if the result is yes, associating the first voice command with the semantic recognition result, otherwise, associating the received user-modified semantic with the first voice command, wherein the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same.
Optionally, the apparatus further includes an updating unit 405, configured to:
when the received second voice command is not successfully matched with any voice command in the stored voice command library, but the similarity between the received second voice command and at least one voice command in the voice command library is not less than a first preset threshold value, recording the second voice command;
prompting the user to input the voice command again;
and if the matching of the re-input voice command is successful, updating the voice command successfully matched with the re-input voice command according to the second voice command.
Optionally, the output unit 402 is further configured to:
when the voice commands received for multiple times are not successfully matched with any voice command in the voice command library, outputting second prompt information for prompting the user whether to set the associated operation instruction;
and outputting the first prompt information when the received third feedback information indicates that the operation instruction associated with the fourth voice command needs to be set.
The device may be configured to execute the method provided in the embodiment shown in fig. 1 to 3, and therefore, for functions and the like that can be realized by each functional module of the device, reference may be made to the description of the embodiment shown in fig. 1, which is not repeated here. Although shown together in fig. 4, semantic recognition section 404 and update section 405 are not essential functional sections and are therefore shown by broken lines.
Referring to fig. 5, based on the same inventive concept, an embodiment of the present invention provides a voice command customizing apparatus, which includes at least one processor 501, where the at least one processor 501 is configured to implement the steps of the voice command customizing method provided by the embodiment shown in fig. 1-3 when executing a computer program stored in a memory.
Optionally, at least one of the processors 501 may specifically include a Central Processing Unit (CPU), an Application Specific Integrated Circuit (ASIC), one or more integrated circuits for controlling program execution, a hardware circuit developed by using a Field Programmable Gate Array (FPGA), or a baseband processor.
Optionally, at least one processor 501 may include at least one processing core.
Optionally, the device further includes a memory 502, and the memory 502 may include a Read Only Memory (ROM), a Random Access Memory (RAM), and a disk memory. The memory 502 is used for storing data required by the at least one processor 501 during operation. The number of the memories 502 is one or more. The memory 502 is shown in fig. 5, but it should be understood that the memory 502 is not an optional functional block, and is shown in fig. 5 by a dotted line.
Based on the same inventive concept, embodiments of the present invention provide a computer-readable storage medium storing computer instructions that, when executed on a computer, cause the computer to perform the method shown in fig. 1-3.
In particular implementations, the computer-readable storage medium includes: various storage media capable of storing program codes, such as a Universal Serial Bus flash drive (USB), a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk, and an optical disk.
In the embodiments of the present invention, it should be understood that the disclosed apparatus and method may be implemented in other ways. For example, the above-described embodiments of the apparatus are merely illustrative, and for example, the described unit or division of units is only one division of logical functions, and there may be other divisions when actually implemented, for example, a plurality of units or components may be combined or may be integrated into another system, or some features may be omitted, or not executed. In addition, the shown or discussed mutual coupling or direct coupling or communication connection may be an indirect coupling or communication connection through some interfaces, devices or units, and may be in an electrical or other form.
The functional units in the embodiments of the present invention may be integrated into one processing unit, or each unit may be an independent physical module.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, all or part of the technical solutions of the embodiments of the present invention may be embodied in the form of a software product, which is stored in a storage medium and includes instructions for causing a computer device, such as a personal computer, a server, or a network device, or a processor (processor) to execute all or part of the steps of the methods according to the embodiments of the present invention. And the aforementioned storage medium includes: various media that can store program codes, such as a universal serial bus flash drive (usb flash drive), a removable hard disk, a ROM, a RAM, a magnetic disk, or an optical disk.
The above embodiments are only used to describe the technical solutions of the present application in detail, but the above embodiments are only used to help understanding the method of the embodiments of the present invention, and should not be construed as limiting the embodiments of the present invention. Variations or substitutions that may be readily apparent to one skilled in the art are intended to be included within the scope of the embodiments of the present invention.

Claims (12)

1. A method for customizing a voice command, comprising:
generating a first voice command according to the collected voice information;
outputting first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment, and generating a first operation instruction for executing an operation step in the demonstration operation process based on the demonstration operation;
establishing an association relation between the first voice command and the first operation instruction, and storing the association relation so that the first operation instruction is executed when a voice command matched with the first voice command is received;
when the received second voice command is not successfully matched with any voice command in the stored voice command library, but the similarity between the received second voice command and at least one voice command in the voice command library is not less than a first preset threshold value, recording the second voice command;
prompting the user to input the voice command again;
and if the re-input voice command is successfully matched, updating the voice command successfully matched with the re-input voice command according to the second voice command.
2. The method of claim 1, wherein generating a first voice command from the collected voice information comprises:
analyzing according to the collected pieces of voice information to extract common features of the pieces of voice information;
generating the first voice command based on the common characteristic.
3. The method of claim 1, wherein said establishing an association between the first voice command and the first operational instruction comprises:
executing the first operation instruction;
determining whether a result of executing the first operation instruction meets the requirement of a user or not based on the received first feedback information;
and when the first feedback information shows that the result of executing the first operation instruction meets the requirement of the user, storing the first voice command after associating the first voice command with the first operation instruction, otherwise, prompting the user to perform demonstration operation again.
4. The method of claim 1, wherein the method further comprises:
carrying out semantic recognition on the collected voice information and outputting a semantic recognition result;
determining whether the semantic recognition result is the semantic expressed by the collected voice information based on the received second feedback information;
and if the result is yes, associating the first voice command with the semantic recognition result, otherwise, associating the received user-modified semantic with the first voice command, wherein the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same.
5. The method of any of claims 1-4, wherein the method further comprises:
when the voice commands received for multiple times are not successfully matched with any voice command in the voice command library, outputting second prompt information for prompting the user whether to set the associated operation instruction;
and when the received third feedback information indicates that an operation instruction associated with a fourth voice command needs to be set, outputting the first prompt information.
6. A voice command customization apparatus, comprising:
the generating unit is used for generating a first voice command according to the collected voice information;
the output unit is used for outputting first prompt information to indicate a user to input demonstration operation for realizing at least one function of the intelligent household equipment;
the generating unit is further used for generating a first operation instruction used for executing an operation step in the demonstration operation process based on the demonstration operation;
the association unit is used for establishing an association relationship between the first voice command and the first operation instruction and storing the association relationship so as to execute the first operation instruction when a voice command matched with the first voice command is received;
the updating unit is used for recording the second voice command when the received second voice command is not successfully matched with any voice command in the stored voice command library, but the similarity between the received second voice command and at least one voice command in the voice command library is not less than a first preset threshold value; prompting the user to input the voice command again; and if the re-input voice command is successfully matched, updating the voice command successfully matched with the re-input voice command according to the second voice command.
7. The apparatus of claim 6, wherein the generating unit is specifically configured to:
analyzing according to the collected pieces of voice information to extract common features of the pieces of voice information;
generating the first voice command based on the common characteristic.
8. The apparatus as claimed in claim 6, wherein said association unit is specifically configured to:
executing the first operation instruction;
determining whether a result of executing the first operation instruction meets the requirement of a user or not based on the received first feedback information;
and when the first feedback information shows that the result of executing the first operation instruction meets the requirement of the user, storing the first voice command after associating the first voice command with the first operation instruction, otherwise, prompting the user to perform demonstration operation again.
9. The apparatus of claim 6, wherein the apparatus further comprises a semantic recognition unit to:
carrying out semantic recognition on the collected voice information and outputting a semantic recognition result;
determining whether the semantic recognition result is the semantic expressed by the collected voice information based on the received second feedback information;
and if the result is yes, associating the first voice command with the semantic recognition result, otherwise, associating the received user-modified semantic with the first voice command, wherein the operation instructions corresponding to all the voice commands associated with the same semantic recognition result are the same.
10. The apparatus of any of claims 6-9, wherein the output unit is further to:
when the voice commands received for multiple times are not successfully matched with any voice command in the voice command library, outputting second prompt information for prompting the user whether to set the associated operation instruction;
and when the received third feedback information indicates that an operation instruction associated with a fourth voice command needs to be set, outputting the first prompt information.
11. A voice command customization apparatus, comprising:
at least one processor; and
a memory communicatively coupled to the at least one processor; wherein the content of the first and second substances,
the memory stores instructions executable by the at least one processor to enable the at least one processor to perform the method of any one of claims 1 to 5.
12. A computer storage medium, characterized in that:
the computer storage medium stores computer instructions that, when executed on a computer, cause the computer to perform the method of any of claims 1-5.
CN201810887444.XA 2018-08-06 2018-08-06 Voice command customizing method, device and equipment and computer storage medium Active CN108831469B (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
CN201810887444.XA CN108831469B (en) 2018-08-06 2018-08-06 Voice command customizing method, device and equipment and computer storage medium
PCT/CN2018/121040 WO2020029500A1 (en) 2018-08-06 2018-12-14 Voice command customization method, device, apparatus, and computer storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810887444.XA CN108831469B (en) 2018-08-06 2018-08-06 Voice command customizing method, device and equipment and computer storage medium

Publications (2)

Publication Number Publication Date
CN108831469A CN108831469A (en) 2018-11-16
CN108831469B true CN108831469B (en) 2021-02-12

Family

ID=64153673

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810887444.XA Active CN108831469B (en) 2018-08-06 2018-08-06 Voice command customizing method, device and equipment and computer storage medium

Country Status (2)

Country Link
CN (1) CN108831469B (en)
WO (1) WO2020029500A1 (en)

Families Citing this family (15)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108831469B (en) * 2018-08-06 2021-02-12 珠海格力电器股份有限公司 Voice command customizing method, device and equipment and computer storage medium
CN109584875A (en) * 2018-12-24 2019-04-05 珠海格力电器股份有限公司 A kind of speech ciphering equipment control method, device, storage medium and speech ciphering equipment
CN109901707A (en) * 2018-12-27 2019-06-18 安徽语讯科技有限公司 A kind of learning-oriented operation module being configured in system
CN109871119A (en) * 2018-12-27 2019-06-11 安徽语讯科技有限公司 A kind of learning type intellectual voice operating method and system
US11170774B2 (en) * 2019-05-21 2021-11-09 Qualcomm Incorproated Virtual assistant device
CN110570867A (en) * 2019-09-12 2019-12-13 安信通科技(澳门)有限公司 Voice processing method and system for locally added corpus
CN110580904A (en) * 2019-09-29 2019-12-17 百度在线网络技术(北京)有限公司 Method and device for controlling small program through voice, electronic equipment and storage medium
CN110784384B (en) * 2019-10-16 2021-11-02 杭州九阳小家电有限公司 Household appliance voice technology generation method and intelligent household appliance
CN111785265A (en) * 2019-11-26 2020-10-16 北京沃东天骏信息技术有限公司 Intelligent sound box setting method and device, control method and device and intelligent sound box
CN111063353B (en) * 2019-12-31 2022-11-11 思必驰科技股份有限公司 Client processing method allowing user-defined voice interactive content and user terminal
CN111261158A (en) * 2020-01-15 2020-06-09 上海思依暄机器人科技股份有限公司 Function menu customization method, voice shortcut control method and robot
CN113160807A (en) * 2020-01-22 2021-07-23 广州汽车集团股份有限公司 Corpus updating method and system and voice control equipment
CN111179933A (en) * 2020-01-23 2020-05-19 珠海荣邦电子科技有限公司 Voice control method and device and intelligent terminal
CN114067792A (en) * 2020-08-07 2022-02-18 北京猎户星空科技有限公司 Control method and device of intelligent equipment
CN114246450B (en) * 2020-09-21 2024-02-06 佛山市顺德区美的电热电器制造有限公司 Information processing method, information processing device, cooking equipment and computer readable storage medium

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP5093963B2 (en) * 2000-09-08 2012-12-12 ニュアンス コミュニケーションズ オーストリア ゲーエムベーハー Speech recognition method with replacement command
US20030110040A1 (en) * 2001-12-07 2003-06-12 Creative Logic Solutions Inc. System and method for dynamically changing software programs by voice commands
CN101937693B (en) * 2010-08-17 2012-04-04 深圳市子栋科技有限公司 Video and audio playing method and system based on voice command
CN102842306B (en) * 2012-08-31 2016-05-04 深圳Tcl新技术有限公司 Sound control method and device, voice response method and device
CN103646646B (en) * 2013-11-27 2018-08-31 联想(北京)有限公司 A kind of sound control method and electronic equipment
CN103713905A (en) * 2013-12-29 2014-04-09 广州视源电子科技股份有限公司 User-defined method, device and system of operation steps
CN105845136A (en) * 2015-01-13 2016-08-10 中兴通讯股份有限公司 Voice control method and device, and terminal
CN105989841B (en) * 2015-02-17 2019-12-27 上海汽车集团股份有限公司 Vehicle-mounted voice control method and device
CN105931637A (en) * 2016-04-01 2016-09-07 金陵科技学院 User-defined instruction recognition speech photographing system
CN106484270A (en) * 2016-09-12 2017-03-08 深圳市金立通信设备有限公司 A kind of voice operating event adding method and terminal
CN108174030B (en) * 2017-12-26 2020-11-17 努比亚技术有限公司 Customized voice control implementation method, mobile terminal and readable storage medium
CN108831469B (en) * 2018-08-06 2021-02-12 珠海格力电器股份有限公司 Voice command customizing method, device and equipment and computer storage medium

Also Published As

Publication number Publication date
WO2020029500A1 (en) 2020-02-13
CN108831469A (en) 2018-11-16

Similar Documents

Publication Publication Date Title
CN108831469B (en) Voice command customizing method, device and equipment and computer storage medium
CN108683574B (en) Equipment control method, server and intelligent home system
KR102373905B1 (en) Shortened voice user interface for assistant applications
US11302302B2 (en) Method, apparatus, device and storage medium for switching voice role
US9953648B2 (en) Electronic device and method for controlling the same
CN109658932B (en) Equipment control method, device, equipment and medium
CN107644638B (en) Audio recognition method, device, terminal and computer readable storage medium
US10827067B2 (en) Text-to-speech apparatus and method, browser, and user terminal
KR102004872B1 (en) Electronic device, server and control methods thereof
US10708423B2 (en) Method and apparatus for processing voice information to determine emotion based on volume and pacing of the voice
CN108133707B (en) Content sharing method and system
KR102411619B1 (en) Electronic apparatus and the controlling method thereof
TW201440037A (en) Voice control method, mobile terminal apparatus and voice control system
CN103914234A (en) Interactive server, control method thereof, and interactive system
CN103903613A (en) Information processing method and electronic device
TW200813812A (en) Methods and apparatus for improved voice recognition and voice recognition systems
US20200265843A1 (en) Speech broadcast method, device and terminal
CN109215638B (en) Voice learning method and device, voice equipment and storage medium
CN110851221A (en) Smart home scene configuration method and device
CN111343028A (en) Distribution network control method and device
CN115327932A (en) Scene creation method and device, electronic equipment and storage medium
CN110619888B (en) AI voice rate adjusting method and device and electronic equipment
CN111933135A (en) Terminal control method and device, intelligent terminal and computer readable storage medium
CN113227962A (en) Electronic device and control method thereof
CN103426429A (en) Voice control method and voice control device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant