CN116844543A - Control method and system based on voice interaction - Google Patents

Control method and system based on voice interaction Download PDF

Info

Publication number
CN116844543A
CN116844543A CN202310748193.8A CN202310748193A CN116844543A CN 116844543 A CN116844543 A CN 116844543A CN 202310748193 A CN202310748193 A CN 202310748193A CN 116844543 A CN116844543 A CN 116844543A
Authority
CN
China
Prior art keywords
voice
information
module
wake
voice information
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202310748193.8A
Other languages
Chinese (zh)
Inventor
王峥嵘
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Individual
Original Assignee
Individual
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Individual filed Critical Individual
Priority to CN202310748193.8A priority Critical patent/CN116844543A/en
Publication of CN116844543A publication Critical patent/CN116844543A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L13/00Speech synthesis; Text to speech systems
    • G10L13/02Methods for producing synthetic speech; Speech synthesisers
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/225Feedback of the input speech
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02PCLIMATE CHANGE MITIGATION TECHNOLOGIES IN THE PRODUCTION OR PROCESSING OF GOODS
    • Y02P90/00Enabling technologies with a potential contribution to greenhouse gas [GHG] emissions mitigation
    • Y02P90/02Total factory control, e.g. smart factories, flexible manufacturing systems [FMS] or integrated manufacturing systems [IMS]

Abstract

The application discloses a control method and a control system based on voice interaction. The method comprises the steps that a voice interaction system is started through a wake-up signal, the voice interaction system monitors voice information in real time, the monitored voice information is converted into text information, the converted text information is analyzed, whether functional parameters of the text information corresponding to the voice information are complete or not is judged, and if the functional parameters are complete, corresponding operation is executed; if the operation is incomplete, prompting the user to reply the operation to be executed according to the missing function parameters, and therefore calling the system operation through voice in real time. By adopting the control method and the system, the application program can be called through voice at any time, different function operations can be selected at any time, or different execution parameters of the same function are selected, different requirements of users are met, eyes are not needed in the calling process, the system can be updated in real time, and the use convenience of the voice interaction control system is greatly improved.

Description

Control method and system based on voice interaction
The application relates to a division application of a control method and a system based on voice interaction, which is filed by Chinese patent office at the date of 2016 and 26 in the year of 08, and has the application number of CN201610736268.0, and the whole content of the division application is incorporated by reference.
Technical Field
The application relates to the field of intelligent control, in particular to a control method and system based on voice interaction.
Background
Voice interaction control refers to the invocation of an application program through voice interaction.
The existing voice interaction control method mainly starts an application program on the mobile device automatically according to a user instruction, but in the using process, the user needs to stare at a screen of the mobile device and click by hands. When a certain application program is opened through the voice interaction control method, a user is required to manually close the application program, and when other application programs are required to be called in the process of executing the application program, the voice interaction control method must be restarted to be called. For example, after the map program is opened by the voice interaction control method, the user is required to operate the map by himself to perform operations such as navigation or route query, and other application programs such as short messages cannot be continuously called by the voice interaction control method. This makes the existing voice interaction control method not highly convenient to use.
Disclosure of Invention
The application aims to provide a control method and a control system based on voice interaction, which realize the calling of application programs in a full voice interaction mode, and can realize the simultaneous calling and real-time control of a plurality of programs so as to solve the problem of low use convenience of the voice interaction control method and the voice interaction control system in the prior art.
In order to achieve the above object, the present application provides the following solutions:
a control method based on voice interaction, the method comprising:
acquiring a wake-up signal, wherein the wake-up signal is used for starting a voice interaction system to enter a voice interaction process;
starting the voice interaction system according to the wake-up signal;
acquiring voice information;
judging whether the voice information can be converted into text information or not to obtain a first judging result;
when the first judgment result is negative, returning to the step of acquiring the voice information;
when the first judging result is yes, analyzing the text information corresponding to the voice information, storing the analysis result, and judging whether the functional parameters of the text information are complete according to the analysis result to obtain a second judging result;
when the second judging result is yes, executing the operation corresponding to the character information function parameter, generating a confirmation message, synthesizing the voice information according to the confirmation message, and playing the synthesized voice information;
and when the second judging result is negative, generating a prompt message according to the missing functional parameters, synthesizing voice information according to the prompt message, playing the synthesized voice information, and returning to the step of acquiring the voice information.
Optionally, the starting the voice interaction system according to the wake-up signal specifically includes:
comparing the wake-up signal with wake-up words pre-stored in a system;
and after the comparison is successful, starting the voice interaction system.
Optionally, after analyzing the text information corresponding to the voice information and storing the analysis result, the method further includes:
and returning to the step of acquiring the wake-up signal.
Optionally, before the operation corresponding to the text information function parameter is executed, the method further includes:
when the function parameters of the text information are complete and more than one operation options are corresponding to the function parameters of the text information, generating a selection prompt problem, synthesizing the voice information according to the selection prompt problem, and playing the synthesized voice information, wherein the selection prompt problem is used for prompting a user to select the operation options to be executed.
Optionally, the generating the selection prompt question specifically includes: and generating a message for prompting the user to reply to the option numbers, wherein each option number corresponds to the content of different operation options.
Optionally, after generating the hint message according to the missing function parameters, the method further includes:
when the voice information includes complete functional parameters and the analysis result is incomplete,
prompting the user to supplement the semantic database; the semantic database stores functional parameters and operations corresponding to the functional parameters;
acquiring operation information input by the user;
and correspondingly adding the operation information input by the user to the semantic database, wherein the operation information comprises complete functional parameters.
A system based on voice interactions, the system comprising:
the voice wake-up module is used for acquiring a wake-up signal, and the wake-up signal is used for starting the system to enter a voice interaction process;
the voice interaction system starting module is used for starting the voice interaction system according to the wake-up signal;
the voice information acquisition module is used for acquiring voice information;
the voice information judging module is used for judging whether the voice information can be converted into text information or not to obtain a first judging result;
the voice analysis module is used for analyzing the text information corresponding to the voice information when the first judgment result is yes;
the storage module is used for storing an analysis result of the voice analysis module for analyzing the text information corresponding to the voice information;
the functional parameter judging module is used for judging whether the functional parameters of the text information are complete according to the analysis result to obtain a second judgment result;
the execution module is used for executing the operation corresponding to the character information function parameter when the second judgment result is yes;
the confirmation message generation module is used for generating a confirmation message;
the voice synthesis module is used for synthesizing voice information;
the voice playing module is used for playing the synthesized voice information;
and the prompt message generation module is used for generating a prompt message according to the missing functional parameters when the second judging result is negative.
Optionally, the voice interaction system starting module specifically includes:
the voice interaction system comparison module is used for comparing the wake-up signal with wake-up words prestored in the system;
and the voice interaction system starting module starts the voice interaction system after the comparison is successful.
Optionally, the system further includes a selection prompt question generation module, configured to generate a selection prompt question when the function parameters of the text message are complete and the operation options corresponding to the function parameters of the text message are more than one before the execution module executes the operation corresponding to the function parameters of the text message.
Optionally, the system further includes a semantic supplementing module, where the semantic supplementing module specifically includes:
the supplementary prompt generation module is used for prompting a user to supplement a semantic database when the voice information comprises complete functional parameters and the analysis result is incomplete after the prompt message generation module generates the prompt message according to the missing functional parameters; the semantic database stores functional parameters and operations corresponding to the functional parameters;
the input module is used for acquiring the operation information input by the user;
and the semantic database supplementing module is used for correspondingly adding the complete functional parameters and the operation information input by the user into the semantic database.
According to the specific embodiment provided by the application, the beneficial effects of the application are as follows:
the convenience in use of the voice interaction control method and the voice interaction control system is improved through the following aspects:
(1) All the operations are realized in a voice mode, and eyes are not needed to be used by hands;
(2) The operation is started at any time through voice, the operation of a new function can be started at any time through voice, the operation of different functions can be selected at any time, or different execution parameters of the same function are selected, so that different requirements of users are met;
(3) The semantic knowledge base is configured by user definition, so that semantic analysis can be flexibly expanded and adjusted, and the variety of functions of processing execution is increased.
(4) The operation is simple, and the application is convenient.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions of the prior art, the drawings that are needed in the embodiments will be briefly described below, it being obvious that the drawings in the following description are only some embodiments of the present application, and that other drawings may be obtained according to these drawings without inventive effort for a person skilled in the art.
FIG. 1 is a flowchart of an embodiment 1 of a control method based on voice interaction according to the present application;
FIG. 2 is a block diagram of a system based on voice interaction in accordance with the present application;
fig. 3 is a flowchart of an embodiment 2 of a control method based on voice interaction according to the present application.
Detailed Description
The following description of the embodiments of the present application will be made clearly and completely with reference to the accompanying drawings, in which it is apparent that the embodiments described are only some embodiments of the present application, but not all embodiments. All other embodiments, which can be made by those skilled in the art based on the embodiments of the application without making any inventive effort, are intended to be within the scope of the application.
In order that the above-recited objects, features and advantages of the present application will become more readily apparent, a more particular description of the application will be rendered by reference to the appended drawings and appended detailed description.
Fig. 1 is a flowchart of an embodiment 1 of a control method based on voice interaction according to the present application. As shown in fig. 1, embodiment 1 includes:
step 101: the method comprises the steps of obtaining a wake-up signal, wherein the wake-up signal is used for starting a voice interaction system to enter a voice interaction process, a user can start a new voice interaction process through voice wake-up control at any time, and when the user wakes up, the obtained sound data stream is compared with sounding of one or more pre-stored wake-up words. For example, the wake-up signal may be "turn on the voice interaction system" or "start black horse", and once the comparison is successful, wake-up is successful, start the voice interaction system, and enter the voice interaction process.
Step 102: acquiring voice information, and after starting a voice interaction system, continuously monitoring the voice information by the system;
step 103: judging whether the detected voice information can be converted into text information or not, and generating a first judging result; when the first determination result is yes, executing step 104; when the first judgment result is negative, returning to the step 102;
step 104: and analyzing the text information corresponding to the voice information, extracting the corresponding functional items to be operated and the parameter items corresponding to the functional items in the text information, storing the analysis result after each analysis, updating the current analysis result, and analyzing the new voice input according to the last stored state in the next analysis. The functional parameter extraction method can be used for extracting keywords or extracting regular expressions; for example, the text information "I want to go to Tiananmen" will be extracted the function item "navigation" and the parameter item "Tiananmen".
Step 105: judging whether the functional parameters are complete according to the extracted functional parameter items, obtaining a second judging result, and executing step 106 when the second judging result is negative; when the second determination result is yes, step 107 is executed;
step 106: and generating a prompt question according to the missing functional parameters, and returning to the step 102 to continuously acquire the voice information. For example, the analyzed voice information is "i want to go", the analysis result is a parameter of a destination lacking navigation, and the generated prompt question may be "please ask you what is you want to go? "please input your destination" is also possible.
Step 107: and executing corresponding operation according to the analyzed functional parameters. For example, the analysis result function item of "i want to go to the Tiananmen" is "navigation", the parameter item is "Tiananmen", and the map is turned on to implement the navigation function, and the destination is designated as Tiananmen. If the navigation route search results are multiple at this time, a selection prompt question is generated according to the navigation route to prompt the user to reply to the operation to be performed, for example, the route to the Tiananmen is 3, and the generated selection prompt question may be "please reply to the option number of the route to be selected: A. straight 500 m and then 1000 m to the destination; B. 300 meters of straight travel and 1000 meters of right-turning travel and 200 meters of left-turning travel reach the destination; C. 1000 meters for right turn and 500 meters for left turn to destination.
Step 108: a confirmation message is generated for confirming with the user that the corresponding operation has been performed. For example, performing an operational navigation to the Tiananmen, the generated confirmation message may be "navigate to Tiananmen for your open".
Fig. 2 is a system structure diagram based on voice interaction according to the present application. As shown in fig. 2, the system includes: a voice wake-up module 201, a voice interaction system starting module 202, a voice information obtaining module 203, a voice information judging module 204, a voice analyzing module 205, a storage module 206, a functional parameter judging module 207, a prompt message generating module 208, an executing module 209, a confirmation message generating module 2010, a voice synthesizing module 2011 and a voice playing module 2012.
The voice wake-up module 201 receives the wake-up signal, and compares the received wake-up signal with one or more wake-up times pre-stored in the system, for example, the wake-up signal may be "start the voice interaction system" or "start the black horse", and when the comparison is successful, the wake-up signal triggers the voice interaction system start-up module 202 to start the voice interaction system. After the voice interaction system is started, the voice information acquisition module 203 continuously listens to voice information and transmits the intercepted voice information to the voice information judgment module 204, the voice information judgment module 204 judges whether the transmitted voice information can be converted into text information, if so, the text information corresponding to the semantic information is transmitted to the voice analysis module 205 to analyze the text information, the voice analysis module 205 extracts corresponding functional items to be operated and parameter items corresponding to the functional items according to the text information, after each analysis, the analysis result is stored in the storage module 206, the current analysis result is updated, and the next analysis is performed on new voice input according to the last stored state. The functional parameter extraction method can be used for extracting keywords or extracting regular expressions; for example, the text information "I want to go to Tiananmen" will be extracted the function item "navigation" and the parameter item "Tiananmen".
The functional parameter judging module 207 judges whether the functional parameter of the text information is complete according to the analysis result, if the functional parameter is complete, the executing module 209 executes corresponding operation, for example, the analysis result functional item of "i want to go to the Tianmen" is "navigation", the parameter item is "Tianmen", the executing module 209 opens the map to implement the navigation function, designates the destination as the Tianmen, then generates a confirmation message through the confirmation message generating module 2010, for example, executes operation navigation to go to the Tianmen, the generated confirmation message may be "already open to navigate to the Tianmen" for you, then performs voice synthesis through the voice synthesizing module 2011, and performs voice playing through the voice playing module 2012. If the navigation route search results are multiple at this time, a selection prompt question is generated according to the navigation route to prompt the user to reply to the operation to be performed, for example, the route to the Tiananmen is 3, and the generated selection prompt question may be "please reply to the option number of the route to be selected: A. straight 500 m and then 1000 m to the destination; B. 300 meters of straight travel and 1000 meters of right-turning travel and 200 meters of left-turning travel reach the destination; C. the right turn is 1000 meters and then the left turn is 500 meters to reach the destination, the voice synthesis is carried out through the voice synthesis module 2011, and the voice playing is carried out through the voice playing module 2012; if it is determined that the functional parameter of the text information is incomplete, a prompt message is generated by the prompt message generating module 208 according to the missing functional parameter, for example, the analyzed voice information is "i want to go", the analysis result is a parameter of a destination where navigation is missing, and the generated prompt question may be "please ask you what is needed? "please reply to your destination", then the voice synthesis is performed by the voice synthesis module 2011, and the voice playing is performed by the voice playing module 2012.
Fig. 3 is a flowchart of an embodiment 2 of a control method based on voice interaction according to the present application. As shown in fig. 3, embodiment 2 includes:
step 301: the wake-up system wakes up the voice interaction system through the wake-up signal to enter the voice interaction process, sets the wake-up signal as 'start voice', and enters the voice interaction process when the system receives the 'start voice' signal;
step 302: acquiring voice information and continuously monitoring the voice information;
step 303: judging whether the intercepted voice information can be converted into text information, if so, executing step 304; discarding the voice which can not be recognized and converted into the text information, and re-intercepting the voice information. For example, bird sounds or wind sounds are detected, and cannot be converted into corresponding text information.
Step 304: for the voice information which can be converted into the text information, analyzing the corresponding text information, and extracting corresponding operation function items and parameter items;
step 305: comparing and calling the function items and parameter items stored in the system database in advance, judging whether the function parameters in the text information corresponding to the voice information are complete, if so, executing step 306, and if not, executing step 307;
step 306: and calling a corresponding module according to the analyzed function item and parameter item, and executing corresponding operation. For ambiguous operations, a selection question is further generated for selection by the user. For example, when the place name is repeated, "i want to go to Pujiang", because Zhejiang province jinhua city has Pujiang county, sichuan province has Pujiang county, and the two names are similar, the system prompts "please reply to the option number of destination: pujiang county is available in Jinhua city, zhejiang province, pujiang county is available in Yingdu city, sichuan province. For example, when two pieces of personal voice information are acquired simultaneously, "i want to go to the Tiananmen" and "call to the small" the system generates a selection question "please reply to the option number of the operation to be performed: and A, navigating to the Tiananmen, and B, calling to give the Ming.
Step 307: supplementing incomplete functional parameters, wherein the incomplete functional parameters comprise two types, namely a voice interaction form, a system carries out questioning according to the missing functional parameters, and a user replies the operation required to be carried out by voice; secondly, when the voice information comprises complete functional parameters and the analysis result is incomplete, the system prompts the user to supplement a semantic database, the semantic database stores the functional parameters and operations corresponding to the functional parameters, and the user can correspondingly increase the complete functional parameters in the voice information into the system database through manual input. In addition, the system of the application can automatically update the database in real time.
In the present specification, each embodiment is described in a progressive manner, and each embodiment is mainly described in a different point from other embodiments, and identical and similar parts between the embodiments are all enough to refer to each other. For the system disclosed in the embodiment, since it corresponds to the method disclosed in the embodiment, the description is relatively simple, and the relevant points refer to the description of the method section.
The principles and embodiments of the present application have been described herein with reference to specific examples, the description of which is intended only to assist in understanding the methods of the present application and the core ideas thereof; also, it is within the scope of the present application to be modified by those of ordinary skill in the art in light of the present teachings. In view of the foregoing, this description should not be construed as limiting the application.

Claims (10)

1. A control method based on voice interaction, the method comprising:
acquiring a wake-up signal, and starting a new voice interaction process through voice wake-up control at any time so as to realize full voice operation;
starting the voice interaction system according to the wake-up signal;
acquiring voice information;
judging whether the voice information can be converted into text information or not to obtain a first judging result;
when the first judgment result is yes, analyzing the text information corresponding to the voice information to extract the corresponding function item to be operated and the parameter item corresponding to the function item in the text information; judging whether the functional parameters of the text information are complete according to the analysis result to obtain a second judgment result;
when the second judging result is yes, executing the operation corresponding to the character information function parameter, generating a confirmation message, synthesizing the voice information according to the confirmation message, and playing the synthesized voice information;
and when the second judging result is negative, generating a prompt message according to the missing functional parameters, synthesizing voice information according to the prompt message, playing the synthesized voice information, and returning to the step of acquiring the voice information.
2. The method according to claim 1, wherein the starting the voice interaction system according to the wake-up signal specifically comprises:
comparing the wake-up signal with wake-up words pre-stored in a system;
and after the comparison is successful, starting the voice interaction system.
3. The method of claim 1, further comprising, prior to said performing the operation corresponding to the text message function parameter:
when the function parameters of the text information are complete and more than one operation options corresponding to the function parameters of the text information are available, generating a selection prompt problem, wherein the selection prompt problem is used for prompting a user to select operation options to be executed; the generating a selection hint question includes: and generating a message for prompting the user to reply to the option numbers, wherein each option number corresponds to the content of different operation options.
4. The method of claim 1, wherein after generating the selection hint question, further comprising:
and synthesizing the voice information according to the selection prompt problem, and playing the synthesized voice information.
5. The method of claim 1, wherein generating the hint message based on the missing function parameters further comprises:
when the voice information includes complete functional parameters and the analysis result is incomplete,
prompting the user to supplement the semantic database; the semantic database stores functional parameters and operations corresponding to the functional parameters;
acquiring operation information input by the user;
and correspondingly adding the operation information input by the user to the semantic database, wherein the operation information comprises complete functional parameters.
6. A system based on voice interaction, the system comprising:
the voice wake-up module is used for acquiring wake-up signals, and starting a new voice interaction process through voice wake-up control at any time so as to realize full voice operation;
the voice interaction system starting module is used for starting the voice interaction system according to the wake-up signal;
the voice information acquisition module is used for acquiring voice information;
the voice information judging module is used for judging whether the voice information can be converted into text information or not to obtain a first judging result;
the voice analysis module is used for analyzing the text information corresponding to the voice information when the first judgment result is yes so as to extract corresponding functional items to be operated and parameter items corresponding to the functional items in the text information;
the functional parameter judging module is used for judging whether the functional parameters of the text information are complete according to the analysis result to obtain a second judgment result;
the execution module is used for executing the operation corresponding to the character information function parameter when the second judgment result is yes;
the confirmation message generation module is used for generating a confirmation message;
the voice synthesis module is used for synthesizing voice information;
the voice playing module is used for playing the synthesized voice information;
and the prompt message generation module is used for generating a prompt message according to the missing functional parameters when the second judging result is negative, synthesizing voice information according to the prompt message, playing the synthesized voice information and returning to the step of acquiring the voice information.
7. The system of claim 6, wherein the voice interaction system activation module specifically comprises:
the voice interaction system comparison module is used for comparing the wake-up signal with wake-up words prestored in the system;
and the voice interaction system starting module starts the voice interaction system after the comparison is successful.
8. The system of claim 6, further comprising a semantic replenishment module, the semantic replenishment module comprising:
the supplementary prompt generation module is used for prompting a user to supplement a semantic database when the voice information comprises complete functional parameters and the analysis result is incomplete after the prompt message generation module generates the prompt message according to the missing functional parameters; the semantic database stores functional parameters and operations corresponding to the functional parameters;
the input module is used for acquiring the operation information input by the user;
and the semantic database supplementing module is used for correspondingly adding the complete functional parameters and the operation information input by the user into the semantic database.
9. An electronic device comprising a memory and a processor, characterized in that the processor executes the method according to any of claims 1 to 5 by invoking a computer program stored in the memory.
10. A computer-readable storage medium, on which a computer program is stored, characterized in that the computer program, when executed on a computer, causes the computer to perform the method of any one of claims 1 to 5.
CN202310748193.8A 2016-08-26 2016-08-26 Control method and system based on voice interaction Pending CN116844543A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202310748193.8A CN116844543A (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
CN201610736268.0A CN106205615B (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction
CN202310748193.8A CN116844543A (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
CN201610736268.0A Division CN106205615B (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction

Publications (1)

Publication Number Publication Date
CN116844543A true CN116844543A (en) 2023-10-03

Family

ID=57526871

Family Applications (2)

Application Number Title Priority Date Filing Date
CN202310748193.8A Pending CN116844543A (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction
CN201610736268.0A Active CN106205615B (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction

Family Applications After (1)

Application Number Title Priority Date Filing Date
CN201610736268.0A Active CN106205615B (en) 2016-08-26 2016-08-26 Control method and system based on voice interaction

Country Status (1)

Country Link
CN (2) CN116844543A (en)

Families Citing this family (16)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106782561A (en) * 2016-12-09 2017-05-31 深圳Tcl数字技术有限公司 Audio recognition method and system
CN106486122A (en) * 2016-12-26 2017-03-08 旗瀚科技有限公司 A kind of intelligent sound interacts robot
CN106782541A (en) * 2017-02-24 2017-05-31 太仓市同维电子有限公司 A kind of Design of Home Gateway method with speech identifying function
CN109215642A (en) * 2017-07-04 2019-01-15 阿里巴巴集团控股有限公司 Processing method, device and the electronic equipment of man-machine conversation
CN107220022A (en) * 2017-07-07 2017-09-29 上海思依暄机器人科技股份有限公司 A kind of method and apparatus for controlling to open navigation feature
CN107919130B (en) * 2017-11-06 2021-12-17 百度在线网络技术(北京)有限公司 Cloud-based voice processing method and device
CN107910005B (en) * 2017-11-16 2021-06-01 海信集团有限公司 Target service positioning method and device for interactive text
CN108132805B (en) * 2017-12-20 2022-01-04 深圳Tcl新技术有限公司 Voice interaction method and device and computer readable storage medium
CN108520748B (en) 2018-02-01 2020-03-03 百度在线网络技术(北京)有限公司 Intelligent device function guiding method and system
CN108320747A (en) * 2018-02-08 2018-07-24 广东美的厨房电器制造有限公司 Appliances equipment control method, equipment, terminal and computer readable storage medium
CN108965604A (en) * 2018-07-25 2018-12-07 安徽新华学院 A kind of mobile phone auxiliary operation system
CN109697981B (en) * 2019-01-02 2021-03-09 百度在线网络技术(北京)有限公司 Voice interaction method, device, equipment and storage medium
CN110033764A (en) * 2019-03-08 2019-07-19 中国科学院深圳先进技术研究院 Sound control method, device, system and the readable storage medium storing program for executing of unmanned plane
CN111692829B (en) * 2019-03-13 2022-02-01 青岛海尔电冰箱有限公司 Data interaction method and interaction system for refrigerator
CN111128168A (en) * 2019-12-30 2020-05-08 斑马网络技术有限公司 Voice control method, device and storage medium
CN111866883A (en) * 2020-06-09 2020-10-30 惠州高盛达科技有限公司 Router WIFI setting method and system based on voice module

Family Cites Families (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
DE19615693C1 (en) * 1996-04-19 1997-12-11 Siemens Ag Device and method for action determination
JPH1097281A (en) * 1996-09-19 1998-04-14 Sony Corp Speech recognition system and navigator
US6434524B1 (en) * 1998-09-09 2002-08-13 One Voice Technologies, Inc. Object interactive user interface using speech recognition and natural language processing
US6964023B2 (en) * 2001-02-05 2005-11-08 International Business Machines Corporation System and method for multi-modal focus detection, referential ambiguity resolution and mood classification using multi-modal input
US20050065779A1 (en) * 2001-03-29 2005-03-24 Gilad Odinak Comprehensive multiple feature telematics system
US7676026B1 (en) * 2005-03-08 2010-03-09 Baxtech Asia Pte Ltd Desktop telephony system
CN102469363A (en) * 2010-11-11 2012-05-23 Tcl集团股份有限公司 Television system with speech comment function and speech comment method
AU2014233517B2 (en) * 2013-03-15 2017-05-25 Apple Inc. Training an at least partial voice command system
CN103281223B (en) * 2013-05-15 2016-05-25 五邑大学 A kind of modern intelligent household security system
CN204390737U (en) * 2014-07-29 2015-06-10 科大讯飞股份有限公司 A kind of home voice disposal system
CN105739977A (en) * 2016-01-26 2016-07-06 北京云知声信息技术有限公司 Wakeup method and apparatus for voice interaction device

Also Published As

Publication number Publication date
CN106205615B (en) 2023-06-02
CN106205615A (en) 2016-12-07

Similar Documents

Publication Publication Date Title
CN106205615B (en) Control method and system based on voice interaction
CN109326289B (en) Wake-up-free voice interaction method, device, equipment and storage medium
US10847139B1 (en) Crowd sourced based training for natural language interface systems
US10832674B2 (en) Voice data processing method and electronic device supporting the same
US10170106B2 (en) Parameter collection and automatic dialog generation in dialog systems
CN106297789B (en) Personalized interaction method and system for intelligent robot
KR20200012933A (en) Shortened voice user interface for assistant applications
US10657959B2 (en) Information processing device, information processing method, and program
CN108737324B (en) Method and device for generating artificial intelligence service assembly and related equipment and system
CN104866275B (en) Method and device for acquiring image information
CN109996026B (en) Video special effect interaction method, device, equipment and medium based on wearable equipment
CN111949240A (en) Interaction method, storage medium, service program, and device
CN109782997B (en) Data processing method, device and storage medium
WO2017041377A1 (en) Method and device for generating calendar reminding information
CN110097884A (en) A kind of voice interactive method and device
CN113593595A (en) Voice noise reduction method and device based on artificial intelligence and electronic equipment
CN110310648A (en) Control method, device, mobile terminal and the readable storage medium storing program for executing of mobile terminal
CN112634896A (en) Operation method of application program on intelligent terminal and intelligent terminal
CN113409791A (en) Voice recognition processing method and device, electronic equipment and storage medium
CN111933135A (en) Terminal control method and device, intelligent terminal and computer readable storage medium
CN110660393A (en) Voice interaction method, device, equipment and storage medium
CN112286486B (en) Operation method of application program on intelligent terminal, intelligent terminal and storage medium
CN110910100A (en) Event reminding method, device, terminal, storage medium and system
CN106197394A (en) Air navigation aid and device
CN113223520B (en) Voice interaction method, system and platform for software operation live-action semantic understanding

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination