CN115775557A - Voice control method, control device and computer readable medium for user interface - Google Patents

Voice control method, control device and computer readable medium for user interface Download PDF

Info

Publication number
CN115775557A
CN115775557A CN202111038371.5A CN202111038371A CN115775557A CN 115775557 A CN115775557 A CN 115775557A CN 202111038371 A CN202111038371 A CN 202111038371A CN 115775557 A CN115775557 A CN 115775557A
Authority
CN
China
Prior art keywords
voice
voice control
control
word
function
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202111038371.5A
Other languages
Chinese (zh)
Inventor
许国华
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Shanghai Qwik Smart Technology Co Ltd
Original Assignee
Shanghai Qwik Smart Technology Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Shanghai Qwik Smart Technology Co Ltd filed Critical Shanghai Qwik Smart Technology Co Ltd
Priority to CN202111038371.5A priority Critical patent/CN115775557A/en
Publication of CN115775557A publication Critical patent/CN115775557A/en
Pending legal-status Critical Current

Links

Images

Landscapes

  • Selective Calling Equipment (AREA)

Abstract

The invention discloses a voice control method, a voice control device and a computer readable medium of a user interface. The voice control method comprises the following steps: loading the user interface, wherein the user interface comprises at least one voice control, and each voice control corresponds to at least one voice word and at least one function word; registering each voice control, and the at least one voice word and the at least one function word corresponding to the voice control to a registry; and determining a target voice control from the at least one voice control based on the received control instruction and executing the target voice control, wherein at least one voice word corresponding to the executed target voice control is related to the control instruction, and the executed target voice control is related to the function word. By implementing the voice control method, the voice control function and the control interface are tightly combined and do not need to be directly bound with the control, so that the development amount is reduced, and the use and learning cost is reduced.

Description

Voice control method, control device and computer readable medium for user interface
Technical Field
The present invention relates to the field of voice control, and in particular, to a voice control method and a voice control apparatus for a user interface, and a computer readable medium.
Background
At present, when a user drives a vehicle, the user sometimes needs to operate various control buttons of a vehicle or virtual buttons on a display screen of the vehicle to execute corresponding functions so as to deal with certain unexpected situations in some driving processes, such as functions of answering a call, starting a wiper and the like, or meet the requirements of the user on the environment in the vehicle, such as turning on an air conditioner, playing music and the like.
However, during driving, the user operates the button by a manual touch/press or other operation, which increases a certain driving risk. Therefore, many car-in-vehicle service providers provide the voice control function, but in the prior art, the voice control function is mostly realized through third-party application software. The voice control function system in the prior art is independent of the third-party application software, that is, the interface of the voice control function system and the third-party application software system is separated. The action executed by the voice control on the display interface of the application software is not controlled and executed by the application software system, and the control and execution party of the action is an independent voice control function system. Therefore, the coverage of the voice control function system on each interface loaded on the third-party application software is small, and the voice control function system cannot perform voice control on the contents on all the interfaces loaded on the application software.
In summary, in order to solve the above problems in the prior art, there is a need in the art for a voice control method and a voice control apparatus for a user interface, in which a provided voice control function is tightly combined with a control interface and does not need to be directly bound with a control, so that the technical effects of reducing the development amount and reducing the use and learning costs of the voice control function are achieved in the development stage of application software.
Disclosure of Invention
The following presents a simplified summary of one or more aspects in order to provide a basic understanding of such aspects. This summary is not an extensive overview of all contemplated aspects, and is intended to neither identify key or critical elements of all aspects nor delineate the scope of any or all aspects. Its sole purpose is to present some concepts of one or more aspects in a simplified form as a prelude to the more detailed description that is presented later.
An aspect of the present invention is to provide a voice control method, a voice control apparatus, and a computer readable medium for a user interface, which are advantageous in that the present invention enables a voice control function to be tightly combined with a control interface and simultaneously does not need to be directly bound with a control, thereby achieving the technical effects of reducing the development amount and reducing the use and learning costs of the voice control function in the development stage of application software.
An aspect of the present invention is to provide a voice control method, a voice control apparatus, and a computer readable medium for a user interface, which are advantageous in that each function word corresponding to a voice control in the present invention may correspond to a plurality of voice phrases, so that a user does not limit a voice instruction word for each voice control to a specific one, and the function word of the voice control is triggered as long as the user speaks one of the plurality of voice phrases corresponding to the function word.
An aspect of the present invention is to provide a voice control method, a voice control apparatus, and a computer readable medium for a user interface, and is advantageous in that a voice callback function is provided for a voice control corresponding to a plurality of groups of voice phrases, and the voice callback function can improve the operation efficiency of the voice control method provided by the present invention to a certain extent.
An aspect of the present invention is to provide a voice control method, a voice control apparatus, and a computer readable medium for a user interface, which are advantageous in that the present invention can provide a simulated click operation while performing voice control on the user interface, so that a user feels a confirmed click operation, thereby improving user experience of the user for the voice control of the interface.
An aspect of the present invention is to provide a voice control method, a voice control apparatus, and a computer readable medium for a user interface, and is advantageous in that the present invention can determine a voice word in a voice control instruction through semantic analysis, and based on the obtained voice word in the voice control instruction, find a voice phrase of a voice control corresponding to the voice word in a registry, and find a key value corresponding to the voice phrase (the key value corresponds to a control ID + a function word), so as to implement a corresponding control function through voice control of a voice control.
The advantages and features of the present invention will become more fully apparent from the following detailed description and appended claims, taken in conjunction with the accompanying drawings.
One aspect of the present invention provides a voice control method for a user interface, including the steps of: loading the user interface, wherein the user interface comprises at least one voice control, and each voice control corresponds to at least one voice word and at least one function word; registering each voice control, the at least one voice word and the at least one function word corresponding to the voice control to a registry; and determining a target voice control from the at least one voice control based on the received control instruction and executing the target voice control, wherein at least one voice word corresponding to the executed target voice control is related to the control instruction, and the executed target voice control is related to the function word.
According to the voice control method in one embodiment of the present invention, at least one voice word includes one or more groups of voice phrases, and each voice phrase corresponds to one functional word; registering each voice control and at least one voice word and at least one function word corresponding to the voice control to a registry comprises: correspondingly registering each voice control, each function word of the voice control and all voice words in the voice word group corresponding to the function word to the registry; based on the received control instruction, determining a target voice control from the at least one voice control and executing the steps of: determining the received voice word based on the control instruction, searching the target voice control and the target function word corresponding to the received voice word from the registry, and executing the function corresponding to the target voice control and the target function word.
According to the voice control method in one embodiment of the present invention, there are a plurality of voice controls, and at least some of the voice controls are provided with voice callbacks; the executing the function of the target voice control corresponding to the target function word comprises the following steps: determining whether the target voice control is provided with a voice callback; and responding to the target voice control and setting voice callback, taking the functional word of the target voice phrase as a parameter of the voice callback, and executing the voice callback to realize the corresponding control function; and responding to the condition that the target voice control is not provided with the voice callback, and directly implementing click operation processing.
Further, the voice control includes: the voice control of the multiple groups of the voice phrases and the voice control of the one group of the voice phrases, wherein the voice controls of the multiple groups of the voice phrases are provided with voice callbacks.
According to the voice control method in an embodiment of the present invention, the executing the function of the target voice control corresponding to the target function word further includes: after the voice callback is executed, selectively simulating a click operation according to the return value.
Further, the selectively simulating the clicking operations based on the return value comprises: in response to the return value being true, simulating a click operation; and responding to the return value being false, not executing the simulated click operation.
According to the voice control method in one embodiment of the invention, each voice control is provided with a unique control ID; registering each of the voice controls and the at least one voice word and the at least one function word corresponding to the voice control to a registry further comprises: splicing the control ID of each voice control with the corresponding function word of each voice phrase of the voice control to obtain a corresponding Key value; and registering each Key value of each voice control and each voice word in the voice word group corresponding to the corresponding functional word in the registry in a correlated manner; searching the target voice control and the target function word corresponding to the received voice word from the registry comprises: searching the Key value corresponding to the received voice word from the registry; and obtaining the target voice control and the target function word from the Key value.
Further, the control instruction comprises a voice control instruction, and the determining the received voice word based on the control instruction comprises: converting the voice control instruction into a text control instruction; and performing semantic analysis on the text control instruction to determine the received voice word.
Further, performing semantic analysis on the text control directive to determine the received speech word includes: semantic analysis is performed on the text control instructions based on a knowledge base of speech words to determine the received speech words.
Another aspect of the present invention provides a voice control apparatus for a user interface, including: a memory; and a processor. The processor is connected to the memory and configured to implement a method of controlling a user interface provided by an aspect of the invention.
Yet another aspect of the present invention provides a computer readable storage medium having computer instructions stored thereon. The computer instructions, when executed by a processor, implement a method for voice control of a user interface provided by an aspect of the present invention.
According to the voice control method, the voice control device and the computer readable medium of the user interface provided by the invention, the voice control function can be tightly combined with the control interface and simultaneously does not need to be directly bound with the control, so that the technical effects of reducing the development amount and reducing the use and learning cost of the voice control function in the development stage of application software are realized.
Drawings
The above features and advantages of the present disclosure will be better understood upon reading the detailed description of embodiments of the disclosure in conjunction with the following drawings. In the drawings, components are not necessarily drawn to scale, and components having similar relative characteristics or features may have the same or similar reference numerals.
FIG. 1 illustrates a flow diagram of a method for voice control of a user interface provided in accordance with some embodiments of the present invention;
FIG. 2 illustrates a flow diagram of a method for voice control of a user interface in some embodiments provided in accordance with an aspect of the present invention;
FIG. 3 illustrates a flow diagram of a method for voice control of a user interface in some embodiments provided in accordance with another aspect of the invention;
FIG. 4 illustrates a flow chart of voice control in a method of voice control of a user interface provided in accordance with some embodiments of the present invention;
FIG. 5 illustrates a flow diagram of a voice callback function determination in a method of voice control of a user interface provided according to some embodiments of the invention;
FIG. 6 illustrates a flow diagram of voice callback return determination in a method of voice control of a user interface provided in accordance with some embodiments of the invention; and
FIG. 7 illustrates a schematic diagram of a voice control device of a user interface provided in accordance with some embodiments of the present invention.
Reference numerals:
700: a control device for the user interface;
701: a memory;
702: a processor;
703: a bus;
704: a random access memory;
705: a cache memory;
706: a storage system;
707: a program module;
708: an external device;
709: a display;
710: an input/output (I/O) interface; and
711: a network adapter.
Detailed Description
The following description of the embodiments of the present invention is provided for illustrative purposes, and other advantages and effects of the present invention will become apparent to those skilled in the art from the present disclosure. While the invention will be described in connection with the preferred embodiments, there is no intent to limit the features of the invention to those embodiments. On the contrary, the invention has been described in connection with the embodiments for the purpose of covering alternatives or modifications as may be extended based on the claims of the invention. In the following description, numerous specific details are set forth in order to provide a thorough understanding of the present invention. The invention may be practiced without these particulars. Moreover, some of the specific details have been left out of the description in order to avoid obscuring or obscuring the focus of the present invention.
The invention is described in detail below with reference to the figures and the specific embodiments. It is noted that the aspects described below in connection with the figures and the specific embodiments are only exemplary and should not be construed as imposing any limitation on the scope of the present invention.
As described above, in the driving process, a user operates a button by a manual touch/press or the like, which increases a certain driving risk. Therefore, voice control functions are provided by many car machine service providers, but in the prior art, the voice control functions are mostly realized through third-party application software. The voice control function system in the prior art is independent of the third-party application software, that is, the interface of the voice control function system and the third-party application software system is separated. The action executed by the voice control on the display interface of the application software is not controlled and executed by the application software system, and the control and execution party of the action is an independent voice control function system. Therefore, the coverage of the voice control function system on each interface loaded on the third-party application software is small, and the voice control function system cannot perform voice control on the contents on all the interfaces loaded on the application software.
In order to solve the above problems in the prior art, the present invention provides a voice control method for a user interface, which can realize the technical effects of reducing the development amount and reducing the use and learning costs of the voice control function in the development stage of application software by tightly combining the provided voice control function with the control interface and simultaneously not directly binding the control.
In some non-limiting embodiments, the voice control method of the user interface provided by the aspect of the present invention may be implemented by the voice control apparatus of the user interface provided by the other aspect of the present invention. Specifically, the voice control device of the user interface has a memory and a processor. The memory includes, but is not limited to, the above-described computer-readable storage medium provided by another aspect of the present invention on which computer instructions are stored. The processor is connected to the memory and configured to execute the computer instructions stored in the memory to implement the voice control method of the user interface provided by an aspect of the present invention.
The working principle of the voice control device of the user interface will be described in conjunction with some embodiments of the control method of the user interface. It will be appreciated by those skilled in the art that the examples of the voice control method of the user interface are only non-limiting embodiments provided by the present invention, and are intended to clearly illustrate the main concepts of the present invention and provide specific solutions for the public to implement, rather than to limit the overall operation and functions of the voice control apparatus of the user interface. Similarly, the voice control device of the user interface is also only a non-limiting embodiment provided by the present invention, and does not limit the implementation subject of each step in the voice control method of the user interface.
The voice control method of the user interface is realized by developing the native control in the android system. Android (Android) is a mobile operating system based on the open source code of the Linux kernel (which does not contain GNU components) and other open source software. At present, the android system is also widely used in a vehicle machine.
The android official provides a large number of native controls for developers, and the developers can develop the native controls according to different personalized requirements on the basis of the self-contained basic functions of the android system, so that various functions of man-machine interaction under different characteristic requirements are realized. In the development stage of the application software, the voice control of the android system is developed by a development engineer, so that the voice control and the corresponding application software are closely combined, the voice control can cover all interfaces of the application software, the loaded interface of the application software is responded, and the voice control on the interface can directly click the control and execute the control function corresponding to the voice control according to the voice instruction of a user to the current interface.
Referring specifically to fig. 1, fig. 1 shows a flow diagram of a voice control method of a user interface provided according to some embodiments of the present invention. As shown in FIG. 1, in some embodiments of the invention, the steps of the method for voice control of a user interface include:
s100: loading a user interface, wherein the user interface comprises at least one voice control, and each voice control corresponds to at least one voice word and at least one function word.
In some embodiments of the present invention, the voice control method for a user interface provided by the present invention may be applied to various terminal devices, including any device capable of performing human-computer interaction, such as a smart phone, a vehicle-mounted system, and a tablet computer.
The user interface, which is typically loaded on a display screen of a human-computer interaction device, typically includes a plurality of speech controls. In computer programming, a control (or component, widget or control) is a graphical user interface element that displays an arrangement of information that can be changed by a user, such as a window or text box. The control definition features a single point of interaction for direct manipulation of given data. A widget is a basic visual building block contained in an application program that controls all data processed by the program and the interactive operations on that data.
In some embodiments of the present invention, a voice control is added to an application program by a development engineer, and a function action corresponding to the voice control is made, so as to implement a control function corresponding to the voice control. Each voice control on the current interface is associated with at least one voice word and at least one function word. The speech word refers to a word corresponding to a user's speech uttered, and the function word refers to a word representing a control function performed by the speech control.
Further, in each voice control, at least one voice word includes one or more groups of voice phrases, and each group of voice phrases corresponds to one function word.
For example, the currently loaded user interface is a music interface, and the music interface includes a voice control for music playing. The voice control for music playing corresponds to two functional words which are respectively played and paused, namely, the playing and the pausing of the music on the music interface are triggered by the same voice control. The playing function words correspond to a group of voice phrases which can be played, opened and the like respectively. The pause function word also corresponds to a group of voice phrases which can be pause, stop, close, etc. That is to say, when the user wants to let the music playing control on the current music interface execute the function of playing music, the user speaks any one of the speech words "play, open, and start" corresponding to the above playing function word, and the music playing control will execute the control function of playing music. Similarly, when the user wants to make the music playing control on the current music interface execute the music pausing function, the music playing control can execute the music pausing control function only by speaking any one of the speech words "pause, stop and close" corresponding to the pausing function word.
In this embodiment, each function word corresponding to each voice control is set to correspond to a plurality of voice phrases, so that the voice instruction word of each voice control is not limited to a specific one by the user, and the function word of the voice control is triggered as long as the user speaks one of the plurality of voice phrases corresponding to the function word.
The voice controls configured on the user interface, the corresponding functional words set by each voice control, the voice phrases corresponding to the functional words, and one or more voice words included in the voice phrases can be set by the developer, and the above settings are not limited to the embodiments mentioned in the invention.
In this embodiment, the voice control method for the user interface further includes:
s200: and registering each voice control, the at least one voice word and the at least one function word corresponding to the voice control to a registry.
Referring to fig. 2, in step S210, each voice control on the current interface, each function word of the voice control, and all the voice words in the voice phrases corresponding to the function word are registered in the registry correspondingly.
And continuing to take the music playing control part on the music interface and the corresponding functional words and voice phrases as an example, and correspondingly registering the music playing control part and the corresponding functional words and voice phrases in a registry. The registry is specifically shown in table 1 below:
Figure BDA0003248223010000091
TABLE 1
Still further, as can be seen in fig. 3, each voice control is provided with a unique control ID. In step S211 shown in fig. 3, the control ID of each voice control on the currently loaded page is spliced with the corresponding function word of each voice phrase of the voice control, so as to obtain a corresponding Key value. For example, the ID of the music playing control of the music interface is M1234, the ID (M1234) of the music playing control is spliced with the playing function word corresponding to the ID to obtain Key1, and the ID (M1234) of the music playing control is spliced with the pause function word corresponding to the ID to obtain Key2.
Then, step S211 further includes registering each Key value of each voice control in the registry in association with each voice word in the voice phrase corresponding to its corresponding function word. The registry containing key1, key2, using the above example, is shown in table 2 below:
Figure BDA0003248223010000092
TABLE 2
Referring to fig. 1, step S300 of executing the voice control method of the user interface is as follows:
s300: and determining a target voice control from the at least one voice control based on the received control instruction and executing the target voice control, wherein at least one voice word corresponding to the executed target voice control is related to the control instruction, and the mode of the executed target voice control is related to the function word.
In this embodiment, the control instruction mainly includes a voice control instruction, and the user interface is controlled by the voice control instruction.
Referring specifically to fig. 4, fig. 4 shows a voice control flowchart in a voice control method of a user interface provided according to some embodiments of the present invention. The voice control flow comprises the following steps:
s410: converting the voice control instruction into a text control instruction;
s420: semantic analysis is performed on the text control instructions to determine received speech words.
After the user speaks the voice control instruction, the voice control system converts the voice control instruction into a text control instruction, and then semantic analysis is performed on the text control instruction, so that a corresponding voice word in the user voice control instruction is obtained. Further, by comparing the corresponding voice word in the user voice control instruction with the voice word in the voice word knowledge base, if the voice word knowledge base has the same voice word as the received voice word, the received voice word can be determined to be the voice word required in the voice control method of the subsequent execution user interface.
For example, corresponding to a music interface, the voice control instruction spoken by the user is "please help me to open music", the voice control system converts the voice control instruction into a text control instruction, and performs semantic analysis on the text control instruction to obtain that the corresponding voice word in the user voice control instruction is "open". And searching whether the corresponding speech words in the music interface have the open speech words in the speech word knowledge base, and if the corresponding speech words in the music interface have the open speech words, determining that the open speech words are the speech words in the speech control instruction sent by the user. And if the corresponding speech words in the music interface do not have the open speech words, the system prompts the user to re-send the speech control instruction.
Referring to fig. 2, in step S310, the received voice word is determined based on the control command, and the target voice control and the target function word corresponding to the received voice word are searched from the registry. Further, referring to fig. 3, in step S311 shown in fig. 3, first, a Key value corresponding to a voice word received from the user voice control instruction is searched from the registry, and the target voice control and the target function word corresponding to the target voice control are found according to the Key value corresponding to the voice word.
For example, in the case of loading a music interface, after determining that the received speech word is "open" based on the user speech control instruction, the Key value corresponding to the speech word "open" is searched from the registry table 2. The Key value corresponding to the voice word "open" is Key1, and the Key1 can determine that the target voice control in the user voice control instruction is a music playing control and the target function word in the voice control instruction is playing.
And after the target voice control and the target function word are obtained through the Key value, executing the control function corresponding to the target voice control and the target function word.
In this embodiment, a voice word in the voice control instruction is determined through semantic analysis, and based on the obtained voice word in the voice control instruction, a voice phrase of the voice control corresponding to the voice word and a key value (the key value is equivalent to the control ID + the function word) corresponding to the voice phrase are found in the registry, so that the voice control on the interface can quickly respond to the voice control instruction of the user, and execute a corresponding control function.
Preferably, in an embodiment, there are a plurality of voice controls on the user interface, wherein at least part of the voice controls on the interface are provided with voice callbacks.
A callback function is a function passed as a parameter. The mechanism principle is as follows: (1) defining a callback function; (2) Providing a function implementation party, and registering a function pointer of a callback function to a caller during initialization; (3) When a specific event or condition occurs, the caller uses the function pointer to call the callback function to process the event.
In this embodiment, the caller is a main body of the voice control method of the entire user interface, and the callee is a function word corresponding to the target voice phrase. The caller does not need to define the callee, it only needs to know that there is one called function with a particular prototype and constraints. That is to say, the speech control method used in this embodiment does not need to limit the functional word corresponding to the target speech phrase to a specific word before executing the method, but can be executed in the speech control method as a parameter of the speech callback as long as the functional word corresponding to any target speech phrase satisfies the condition about the functional word.
In this embodiment, by setting the voice callback function on the voice control corresponding to the plurality of groups of voice phrases, the voice callback can improve the operation efficiency of the voice control method to a certain extent.
Referring specifically to fig. 5, fig. 5 is a flowchart illustrating a determination of a voice callback function in a voice control method of a user interface according to some embodiments of the present invention.
The step of judging the voice callback function as shown in fig. 5 includes: it is first determined whether the found target voice control has a voice callback. The types of the voice control comprise two types, namely, one voice control corresponds to a plurality of groups of voice phrases, and one voice control corresponds to one group of voice phrases. Preferably, in an embodiment, if one speech control corresponds to a speech control of a plurality of groups of speech word group classes, a speech callback is set for the speech control.
For example, the voice control for music playing corresponds to a playing function word and a pause function word, and each function word corresponds to a group of voice phrases, so that the voice control for music playing corresponds to two groups of voice phrases, and voice callback is set for the voice phrases. The voice control for indicating music switching only corresponds to one switching function word, and the switching function word only corresponds to one group of voice phrases related to switching meanings, such as switching, next head and the like. Therefore, the voice control for music switching only has one group of voice phrases corresponding to the voice control, and voice callback is not set for the voice control.
And when the judgment result is that the target voice control is not provided with the voice callback, directly implementing click operation processing. For example, in response to that the voice control for music switching is not provided with a voice callback, the voice control for music switching directly executes a click operation, and the voice control for music switching executes a function of switching music.
And when the judgment result shows that the target voice control is provided with the voice callback, taking the functional word of the target voice phrase as the parameter of the voice callback, and executing the voice callback so as to realize the corresponding control function. For example, in response to the voice control for music playing being provided with a voice callback, the functional word "play" of the target voice phrase is used as a parameter of the voice callback, so that the voice control for music playing executes a playing function.
Preferably, the click operation can be selectively simulated according to the return value of the voice callback. Turning to fig. 6, fig. 6 illustrates a flow chart of the voice callback return determination in the voice control method of the user interface according to some embodiments of the present invention.
In the embodiment shown in fig. 6, after the voice callback is executed, the return value is determined to be true or false. And if the obtained return value is true, executing the simulated click operation. In this case, a virtual mouse may be provided at the target voice control in the user voice control instruction on the user interface to simulate the effect of entity mouse clicking, so that the user may feel the confirmed clicking operation, and the user experience of the user on the voice control of the interface is improved. If the obtained return value is false, the simulated click operation is not executed.
In the embodiment of the invention, the simulated click operation is provided while the user interface is subjected to voice control, so that a user can feel the click operation of 'confirmation', and the user experience of the user on the voice control of the interface is improved.
Based on the above description, an aspect of the present invention provides a voice control method of a user interface. The voice control method can realize the technical effects of reducing the development amount and reducing the use and learning cost of the voice control function in the development stage of application software without directly binding with the control while the provided voice control function is tightly combined with the control interface.
While, for purposes of simplicity of explanation, the methodologies are shown and described as a series of acts, it is to be understood and appreciated that the methodologies are not limited by the order of acts, as some acts may, in accordance with one or more embodiments, occur in different orders and/or concurrently with other acts from that shown and described herein or not shown and described herein, as would be understood by one skilled in the art.
According to another aspect of the present invention, the present invention further provides a voice control apparatus 700 for a user interface. Referring to fig. 7, fig. 7 is a schematic diagram illustrating a voice control apparatus of a user interface provided according to some embodiments of the present invention.
As shown in fig. 7, in this embodiment, the voice control apparatus 700 of the user interface is represented in the form of a general-purpose computer device, and is used to implement the steps of the voice control method of the user interface described in any one of the above embodiments. For details, please refer to the description of the voice control method of the user interface above, which is not repeated herein.
The components of the user interface speech control apparatus 700 may include one or more memories 701, one or more processors 702, and a bus 703 that couples the various system components, including the memories 701 and the processors 702.
The bus 703 includes a data bus, an address bus, and a control bus. The product of the number of bits of the data bus and the operating frequency is proportional to the data transfer rate, the number of bits of the address bus determines the maximum addressable memory space, and the control bus (read/write) indicates the type of bus cycle and the time at which the present I/O operation is completed. The processor 702 is coupled to the memory 701 via the bus 703 and is configured to implement the voice control method of the user interface provided by any of the embodiments described above.
The processor 702 is a final execution unit for information processing and program execution, which is an operation and control core of the voice control apparatus 700. The operation of all software layers in the computer system will ultimately be mapped to the operation of the processor 702 by the instruction set. The processor 702 has the main functions of processing instructions, executing operations, controlling time and processing data.
The memory 701 is a storage device for storing programs and data in the computer. Memory 701 may include computer system readable media in the form of storage volatile memory. Such as Random Access Memory (RAM) 704 and/or cache memory 705.
A Random Access Memory (RAM) 704 is an internal memory that exchanges data directly with the processor 702. It can be read and written at any time (except for refreshing), and is fast, usually used as a temporary data storage medium for an operating system or other programs in operation, and the stored data will be lost when power is off. Cache memory (Cache) 705, which is a level one memory that exists between main memory and processor 702, is smaller in size but much faster than main memory, approaching the speed of processor 702.
It should be noted that, in the case where the voice control apparatus 700 includes a plurality of memories 701 and a plurality of processors 702, a distributed structure may be provided between the plurality of memories 701 and between the plurality of processors 702. For example, the system may include a memory 701 and a processor 702 respectively located at the local end and the backend cloud end, and the local end and the backend cloud end jointly implement the voice control method of the user interface. Furthermore, in the embodiment adopting the distributed structure, the specific implementation terminal may be adjusted according to the actual situation in each step, and the specific implementation scheme of each step in a specific terminal should not unduly limit the protection scope of the present invention.
The speech control apparatus 700 may further include other removable/non-removable, volatile/nonvolatile computer system storage media. In this embodiment, the storage system 706 may be used to read from and write to non-removable, nonvolatile magnetic media.
Memory 701 may also include at least one set of program modules 707. Program modules 707 may be stored in the memory 701. Program modules 707 include, but are not limited to, an operating system, one or more application programs, other program modules, and program data, each of which or some combination of which may comprise an implementation of a network environment. The program modules 707 generally perform the functions and/or methodologies of the described embodiments of the invention.
The voice control 700 may also communicate with one or more external devices 708. The external device 708 in this embodiment includes a microphone or the like for collecting user voice control instructions. The external device 708 further includes a display 709, and when implementing the voice control method of the user interface, a user can visually see each voice control on the user interface through the display 709 of the voice control apparatus 700 and control the voice control to execute a corresponding control function through a voice control instruction.
The voice-controlled apparatus 700 may also communicate with one or more devices that enable a user to interact with the voice-controlled apparatus 700 and/or with any devices (e.g., network cards, modems, etc.) that enable the voice-controlled apparatus 700 to communicate with one or more other computing devices. Such communication may occur through input/output (I/O) interfaces 710.
The voice control device 700 may also communicate with one or more networks (e.g., a Local Area Network (LAN), a Wide Area Network (WAN), and/or a public network such as the internet) via the network adapter 711. As shown in fig. 7, the network adapter 711 communicates with the other modules of the voice control apparatus 700 via the bus 703. It should be understood that although not shown in the figures, other hardware and/or software modules may be used in conjunction with the voice-controlled device 700, including but not limited to: microcode, device drivers, redundant processing units, external disk drive arrays, RAID systems, tape drives, and data backup storage systems, among others.
Another aspect of the present invention further provides a computer-readable storage medium storing a computer program, which when executed by a processor implements the steps of the voice control method of the user interface as described in any one of the above embodiments. For details, refer to the above description, and are not repeated herein. It is to be understood that the computer readable storage medium may be a system, which includes a plurality of computer readable storage sub-media, and the steps of the voice control method of the user interface described above are collectively implemented by the plurality of computer readable storage sub-media.
According to the voice control method, the voice control device and the computer readable storage medium of the user interface, provided by the invention, the voice control function can be tightly combined with the control interface and does not need to be directly bound with the control, so that the technical effects of reducing the development amount and reducing the use and learning cost of the voice control function in the development stage of application software are realized.
Those of skill in the art would understand that information, signals, and data may be represented using any of a variety of different technologies and techniques. For example, data, instructions, commands, information, signals, bits (bits), symbols, and chips that may be referenced throughout the above description may be represented by voltages, currents, electromagnetic waves, magnetic fields or particles, optical fields or particles, or any combination thereof.
Those of skill would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present invention.
The various illustrative logical modules, and circuits described in connection with the embodiments disclosed herein may be implemented or performed with a general purpose processor, a Digital Signal Processor (DSP), an Application Specific Integrated Circuit (ASIC), a Field Programmable Gate Array (FPGA) or other programmable logic device, discrete gate or transistor logic, discrete hardware components, or any combination thereof designed to perform the functions described herein. A general purpose processor may be a microprocessor, but in the alternative, the processor may be any conventional processor, controller, microcontroller, or state machine. A processor may also be implemented as a combination of computing devices, e.g., a combination of a DSP and a microprocessor, a plurality of microprocessors, one or more microprocessors in conjunction with a DSP core, or any other such configuration.
Although the processor 702 described in the above embodiments may be implemented by a combination of software and hardware. It is understood that the processor 702 may be implemented in software, hardware. For a hardware implementation, the processor 702 may be implemented within one or more Application Specific Integrated Circuits (ASICs), digital Signal Processors (DSPs), programmable Logic Devices (PLDs), field Programmable Gate Arrays (FPGAs), processors, controllers, micro-controllers, microprocessors, other electronic devices designed to perform the functions described herein, or a select combination thereof. For a software implementation, the processor 702 may be implemented by separate software modules running on a common chip, such as program modules (processes) and function modules (functions), each of which performs one or more of the functions and operations described herein.
The previous description of the disclosure is provided to enable any person skilled in the art to make or use the disclosure. Various modifications to the disclosure will be readily apparent to those skilled in the art, and the generic principles defined herein may be applied to other variations without departing from the spirit or scope of the disclosure. Thus, the disclosure is not intended to be limited to the examples and designs described herein but is to be accorded the widest scope consistent with the principles and novel features disclosed herein.

Claims (11)

1. A method for voice control of a user interface, the method comprising the steps of:
loading the user interface, the user interface including at least one voice control, each voice control corresponding to at least one voice word and to at least one function word;
registering each voice control and the at least one voice word and the at least one function word corresponding to the voice control to a registry; and
and determining a target voice control from the at least one voice control based on the received control instruction, and executing the target voice control, wherein at least one voice word corresponding to the executed target voice control is related to the control instruction, and the executed target voice control is related to the function word.
2. The voice control method according to claim 1, wherein the at least one voice word comprises one or more groups of voice phrases, each of the voice phrases corresponds to one of the function words,
registering each voice control and the at least one voice word and the at least one function word corresponding to the voice control to a registry comprises: correspondingly registering each voice control with all the voice words in the voice phrases corresponding to the function words and the function words of the voice control to the registry,
the determining and executing the target voice control from the at least one voice control based on the received control instruction comprises: and determining the received voice word based on the control instruction, searching the target voice control and the target function word corresponding to the received voice word from the registry, and executing the function corresponding to the target voice control and the target function word.
3. The voice control method of claim 2, said voice control being a plurality of, at least some of said voice controls being provided with voice callbacks,
the executing the function of the target voice control corresponding to the target function word comprises:
determining whether the target voice control is provided with a voice callback; and
taking the functional words of the target voice phrases as the parameters of the voice callback in response to the target voice control being provided with the voice callback, executing the voice callback to realize the corresponding control function,
and responding to the condition that the target voice control is not provided with the voice callback, and directly implementing click operation processing.
4. The voice control method of claim 3, the voice control comprising: the voice control of the plurality of groups of voice phrases and the voice control of the group of voice phrases, wherein the voice controls of the plurality of groups of voice phrases are provided with voice callbacks.
5. The voice control method of claim 3, the performing the function of the target voice control corresponding to the target function word further comprising:
and selectively simulating click operation according to the return value after the voice callback is executed.
6. The voice control method of claim 5, the selectively simulating a click operation according to the return value comprising:
in response to the return value being true, simulating a click operation;
and responding to the return value being false, not executing the simulated click operation.
7. The voice-control method of claim 2, each of the voice controls being provided with a unique control ID,
registering each of the voice controls and the at least one voice word and the at least one function word corresponding to the voice control to a registry further comprises: splicing the control ID of each voice control with the corresponding function word of each voice phrase of the voice control to obtain a corresponding Key value; and registering each said Key value of each said voice control in association with each said voice word in said voice phrase corresponding to the corresponding said function word in said registry,
the searching for the target voice control and the target function word corresponding to the received voice word from the registry comprises: searching the Key value corresponding to the received voice word from the registry; and obtaining the target voice control and the target function word from the Key value.
8. The voice control method of claim 2, the control instruction comprising a voice control instruction, the determining the received voice word based on the control instruction comprising:
converting the voice control instruction into a text control instruction; and
performing semantic analysis on the text control instructions to determine the received speech words.
9. The speech control method of claim 8, the performing semantic analysis on the text control instruction to determine the received speech word comprising:
performing semantic analysis on the text control instructions based on a speech word knowledge base to determine the received speech words.
10. A voice control apparatus of a user interface, comprising:
a memory; and
a processor configured to execute a computer program on the memory to implement the method of any one of claims 1-9.
11. A computer-readable medium having stored thereon computer-executable instructions that, when executed by a processor, implement the method of any one of claims 1-9.
CN202111038371.5A 2021-09-06 2021-09-06 Voice control method, control device and computer readable medium for user interface Pending CN115775557A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111038371.5A CN115775557A (en) 2021-09-06 2021-09-06 Voice control method, control device and computer readable medium for user interface

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111038371.5A CN115775557A (en) 2021-09-06 2021-09-06 Voice control method, control device and computer readable medium for user interface

Publications (1)

Publication Number Publication Date
CN115775557A true CN115775557A (en) 2023-03-10

Family

ID=85387971

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111038371.5A Pending CN115775557A (en) 2021-09-06 2021-09-06 Voice control method, control device and computer readable medium for user interface

Country Status (1)

Country Link
CN (1) CN115775557A (en)

Similar Documents

Publication Publication Date Title
CN108829235B (en) Voice data processing method and electronic device supporting the same
US11080016B2 (en) Speech-based interaction with a display window
US10621972B2 (en) Method and device extracting acoustic feature based on convolution neural network and terminal device
US20180374476A1 (en) System and device for selecting speech recognition model
KR20180117485A (en) Electronic device for processing user utterance and method for operation thereof
JP2013143151A (en) Method and device for performing user function by using voice recognition
JP7158217B2 (en) Speech recognition method, device and server
CN108845736A (en) Exchange method and system for vehicle-mounted voice system
AU2010254812A1 (en) Contextual voice commands
CN107967055A (en) A kind of man-machine interaction method, terminal and computer-readable medium
WO2021110133A1 (en) Control operation method and electronic device
CN106257410B (en) Method, electronic device and apparatus for multi-mode disambiguation of voice-assisted inputs
CN110727410A (en) Man-machine interaction method, terminal and computer readable storage medium
JP2002116796A (en) Voice processor and method for voice processing and storage medium
US20200265843A1 (en) Speech broadcast method, device and terminal
EP3745253B1 (en) Voice control methods and apparatuses for electronic device, computer devices, and storage media
CN112652302B (en) Voice control method, device, terminal and storage medium
CN112286485B (en) Method and device for controlling application through voice, electronic equipment and storage medium
US20210166687A1 (en) Terminal device, server and controlling method thereof
CN103177724A (en) Method, device and terminal for text operating controlled by voice
KR20210088464A (en) Method and apparatus for processing voice interaction, electronic equipment, storage medium, and computer program product
CN108052212A (en) A kind of method, terminal and computer-readable medium for inputting word
EP3745252B1 (en) Voice control method and apparatus of electronic device, computer device and storage medium
CN115775557A (en) Voice control method, control device and computer readable medium for user interface
US9047059B2 (en) Controlling a voice site using non-standard haptic commands

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination