CN113270096A - Voice response method and device, electronic equipment and computer readable storage medium - Google Patents

Voice response method and device, electronic equipment and computer readable storage medium Download PDF

Info

Publication number
CN113270096A
CN113270096A CN202110524931.1A CN202110524931A CN113270096A CN 113270096 A CN113270096 A CN 113270096A CN 202110524931 A CN202110524931 A CN 202110524931A CN 113270096 A CN113270096 A CN 113270096A
Authority
CN
China
Prior art keywords
operation instruction
information
voice information
voice
user
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110524931.1A
Other languages
Chinese (zh)
Inventor
钟煌辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Qianhai Qijian Technology Shenzhen Co ltd
Original Assignee
Qianhai Qijian Technology Shenzhen Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Qianhai Qijian Technology Shenzhen Co ltd filed Critical Qianhai Qijian Technology Shenzhen Co ltd
Priority to CN202110524931.1A priority Critical patent/CN113270096A/en
Publication of CN113270096A publication Critical patent/CN113270096A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/226Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics
    • G10L2015/227Procedures used during a speech recognition process, e.g. man-machine dialogue using non-speech characteristics of the speaker; Human-factor methodology

Abstract

The invention relates to a voice response method, a voice response device, electronic equipment and a computer readable storage medium.

Description

Voice response method and device, electronic equipment and computer readable storage medium
Technical Field
The application belongs to the field of voice control, and particularly relates to a voice response method and device, electronic equipment and a computer readable storage medium.
Background
With the development of human-computer interaction technology, more and more terminals or Application programs (APPs) installed in the terminals can recognize and respond to voice commands sent by users, so as to interact with the users.
However, in the current human-computer voice interaction application scenario, when a user has a requirement for controlling the terminal device through a voice instruction, the user is required to send a complete voice instruction, which consumes time and affects the user experience.
Disclosure of Invention
In view of the above, an object of the present application is to provide a voice response method, apparatus, electronic device and computer-readable storage medium, which can save time and improve user experience.
The embodiment of the application is realized as follows:
in a first aspect, an embodiment of the present application provides a voice response method, where the method includes: acquiring voice information; judging whether the voice information has missing information or not; if so, determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information according to the historical data of the user; responding to the corresponding operation instruction.
In the embodiment of the application, the voice information with the missing information generated by the user can be matched with the corresponding operation instruction, so that the failure of the voice control function can be avoided, and the user experience can be improved. In addition, when the user has the requirement of voice control, the user does not need to dictate complete voice information, and the user experience can be improved from the perspective.
With reference to the embodiment of the first aspect, in a possible implementation manner, the determining whether the voice information has missing information includes: searching whether an operation instruction corresponding to the voice information exists in a mapping relation stored in advance; if yes, determining that the voice information has no missing information; otherwise, determining that the voice information has missing information.
With reference to the embodiment of the first aspect, in a possible implementation manner, the method is applied to an in-vehicle system, where the in-vehicle system includes a display device, and before the responding to the corresponding operation instruction, the method further includes: generating description information for determining whether to execute the corresponding operation instruction; displaying the description information to a user through the display equipment;
correspondingly, the responding to the corresponding operation instruction includes: responding to the corresponding operation instruction after determining that the reverse instruction is not acquired; wherein the reverse instruction is used for characterizing not to execute the corresponding operation instruction. In this embodiment, the user may control the determined execution authority of the corresponding operation instruction, so as to avoid a misoperation.
With reference to the embodiment of the first aspect, in a possible implementation manner, the determining that a backward instruction is not acquired includes: and when user feedback information representing that the corresponding operation instruction is confirmed to be executed is acquired within a preset time period, or when any user feedback information aiming at the corresponding operation instruction is not acquired within the preset time period, determining that the reverse instruction is not acquired.
With reference to the embodiment of the first aspect, in a possible implementation manner, before the determining that the operation instruction with the highest matching degree with the voice information is the operation instruction corresponding to the voice information, the method further includes: and acquiring historical data corresponding to the account information through the account information logged in by the user. In this embodiment, the user can synchronize the historical data on different terminal devices, and then it can be ensured that the user performs voice control on different terminal devices through the voice information with missing information, thereby improving the user experience.
With reference to the embodiment of the first aspect, in a possible implementation manner, the history data further includes a use frequency corresponding to the various operation instructions, and the determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information includes: and when determining that a plurality of operation instructions with the highest matching degree exist, determining the operation instruction with the highest use frequency in the plurality of operation instructions with the highest matching degree as the corresponding operation instruction.
With reference to the embodiment of the first aspect, in a possible implementation manner, the history data further includes habit data of the user, and the responding to the corresponding operation instruction includes: responding to the corresponding operation instruction by combining the habit data. In this embodiment, the operation instruction is responded by combining the habit data of the user, so that more humanized service can be provided for the user, and the user experience can be improved.
In a second aspect, an embodiment of the present application provides a voice response apparatus, including: the device comprises an acquisition module, a judgment module, a determination module and a response module.
The acquisition module is used for acquiring voice information;
the judging module is used for judging whether the voice information has missing information or not;
the determining module is used for determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information according to the historical data of the user when the judging module judges that the voice information is the voice information;
and the response module is used for responding to the corresponding operation instruction.
With reference to the second aspect, in a possible implementation manner, the determining module is configured to: searching whether an operation instruction corresponding to the voice information exists in a mapping relation stored in advance; if yes, determining that the voice information has no missing information; otherwise, determining that the voice information has missing information.
With reference to the embodiment of the second aspect, in a possible implementation manner, the method is applied to an on-board system, where the on-board system includes a display device, and the apparatus further includes a display module, configured to generate description information used for determining whether to execute the corresponding operation instruction; displaying the description information to a user through the display equipment;
correspondingly, the response module is configured to respond to the corresponding operation instruction after determining that the reverse instruction is not obtained; wherein the reverse instruction is used for characterizing not to execute the corresponding operation instruction.
With reference to the second aspect embodiment, in a possible implementation manner, the response module is configured to determine that the backward instruction is not acquired when user feedback information representing that it is confirmed to execute the corresponding operation instruction is acquired within a preset time period, or when any user feedback information for the corresponding operation instruction is not acquired within the preset time period.
With reference to the second aspect, in a possible implementation manner, the obtaining module is further configured to obtain, through account information logged in by the user, historical data corresponding to the account information.
With reference to the second aspect, in a possible implementation manner, the history data further includes usage frequencies corresponding to the various operation instructions, and the determining module is configured to determine, when it is determined that there are a plurality of operation instructions with the highest matching degree, an operation instruction with the highest usage frequency among the plurality of operation instructions with the highest matching degree as the corresponding operation instruction.
With reference to the second aspect, in a possible implementation manner, the history data further includes habit data of the user, and the response module is configured to respond to the corresponding operation instruction in combination with the habit data.
In a fourth aspect, an embodiment of the present application further provides an electronic device, including: the device comprises a voice acquisition unit, a memory and a processor, wherein the processor is respectively connected with the memory and the voice acquisition unit; the memory is used for storing programs; the processor calls a program stored in the memory to perform the method of the first aspect embodiment and/or any possible implementation manner of the first aspect embodiment.
In a fourth aspect, the present application further provides a non-transitory computer-readable storage medium (hereinafter, referred to as a computer-readable storage medium), on which a computer program is stored, where the computer program is executed by a computer to perform the method in the foregoing first aspect and/or any possible implementation manner of the first aspect.
Additional features and advantages of the application will be set forth in the description which follows, and in part will be obvious from the description, or may be learned by the practice of the embodiments of the application. The objectives and other advantages of the application may be realized and attained by the structure particularly pointed out in the written description and drawings.
Drawings
In order to more clearly illustrate the embodiments of the present application or the technical solutions in the prior art, the drawings needed to be used in the embodiments will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present application, and it is obvious for those skilled in the art to obtain other drawings without creative efforts. The foregoing and other objects, features and advantages of the application will be apparent from the accompanying drawings. Like reference numerals refer to like parts throughout the drawings. The drawings are not intended to be to scale as practical, emphasis instead being placed upon illustrating the subject matter of the present application.
Fig. 1 shows a flowchart of a voice response method according to an embodiment of the present application.
Fig. 2 shows a block diagram of a voice response apparatus according to an embodiment of the present application.
Fig. 3 shows a schematic structural diagram of an electronic device provided in an embodiment of the present application.
Icon: 100-an electronic device; 110-a processor; 120-a memory; 130-a voice acquisition unit; 400-a voice response means; 410-an obtaining module; 420-a judgment module; 430-a determination module; 440-response module.
Detailed Description
The technical solutions in the embodiments of the present application will be described below with reference to the drawings in the embodiments of the present application.
It should be noted that: like reference numbers and letters refer to like items in the following figures, and thus, once an item is defined in one figure, it need not be further defined and explained in subsequent figures. Meanwhile, relational terms such as "first," "second," and the like may be used solely in the description herein to distinguish one entity or action from another entity or action without necessarily requiring or implying any actual such relationship or order between such entities or actions. Also, the terms "comprises," "comprising," or any other variation thereof, are intended to cover a non-exclusive inclusion, such that a process, method, article, or apparatus that comprises a list of elements does not include only those elements but may include other elements not expressly listed or inherent to such process, method, article, or apparatus. Without further limitation, an element defined by the phrase "comprising an … …" does not exclude the presence of other identical elements in a process, method, article, or apparatus that comprises the element.
Further, the term "and/or" in the present application is only one kind of association relationship describing the associated object, and means that three kinds of relationships may exist, for example, a and/or B may mean: a exists alone, A and B exist simultaneously, and B exists alone.
In addition, the defects of the voice control type products appearing in the prior art (in the current human-computer voice interaction application scenario, whenever a user needs to control a terminal device through a voice instruction, the user is required to issue a complete voice instruction, and the use experience of the user is further affected) are the results obtained after the applicant has practiced and studied carefully, and therefore, the discovery process of the above defects and the solutions proposed in the embodiments of the present application to the above defects in the following description should be considered as contributions of the inventor to the present application.
In order to solve the above problem, embodiments of the present application provide a voice response method, an apparatus, an electronic device, and a computer-readable storage medium, which can reduce the number of repeated voice contents required by a user during voice control, and improve user experience.
The technology can be realized by adopting corresponding software, hardware and a combination of software and hardware. The following describes embodiments of the present application in detail.
First, referring to fig. 1, an embodiment of the present application provides a voice response method. The method can be applied to the terminal device, the Application program (APP) installed in the terminal device, and the applet embedded in some Application programs, such as the applet in the WeChat.
The steps involved will be described below with reference to fig. 1.
Step S110: and acquiring voice information.
In the embodiment of the application, when a user has a requirement for controlling the terminal device, the corresponding operation instruction can be triggered by dictating voice information, and the corresponding operation instruction can also be triggered by controlling the physical key and the virtual key, so that the terminal device can respond to the operation instruction to realize the control of the terminal device.
Correspondingly, the terminal device may be internally provided with or externally connected with a voice obtaining unit for obtaining the voice information provided by the user. The voice acquiring unit may be a microphone, a sound pickup, or the like.
Certainly, in some embodiments, the terminal device triggers the corresponding operation instruction through the acquired voice information only on the premise that it is confirmed that the voice control function of the terminal device is in the awake state, otherwise, the default voice control function is closed.
The user can switch the voice control function to the wake-up state by controlling the physical key and the virtual key, or can switch the voice control function to the wake-up state by a specific wake-up password (for example, "little X classmates").
In other embodiments, the terminal device may also trigger a corresponding operation instruction through the acquired voice information in any case.
Step S120: and judging whether the voice information has missing information.
As mentioned above, in the embodiment of the present application, the corresponding operation instruction may be triggered by voice information. In order to achieve this effect, it is necessary to establish a mapping relationship between various operation instructions and the corresponding standard voice information in advance, so that after the voice information sent by the user is acquired, the standard voice information is used as a medium, and the operation instructions corresponding to the voice information are matched through the mapping relationship.
For example, a mapping relation is established between standard voice information 'turn on air conditioner' and an instruction A in advance, and when the terminal equipment responds to the instruction A, the air conditioner can be turned on; and establishing a mapping relation between standard voice information 'open navigation' and the instruction B, and opening the navigation when the terminal equipment responds to the instruction B. Correspondingly, when the user dictates the voice message "turn on the air conditioner", the command a can be triggered.
Under the premise, in the conventional scheme, when a user has a requirement for triggering a specific operation instruction by sending out voice information, the user needs to dictate complete standard voice information corresponding to the specific operation instruction every time. The complete standard voice information needs to cover all the voice contents included in the standard voice information corresponding to the specific operation instruction.
If the speech information dictated by the user does not cover all speech contents included in the standard speech information corresponding to the specific operation instruction, it is indicated that the speech information dictated by the user has missing information, and at this time, the terminal device cannot accurately recognize the operation instruction which is expected to be triggered by the user.
For example, in the above example, if the acquired voice information is "on", the terminal device cannot confirm whether the operation instruction that the user desires to trigger is instruction a or instruction B, and further the voice control function is disabled, which affects user experience.
In order to solve the above problem, in the embodiment of the present application, when the voice information is acquired, it may be determined whether the voice information has missing information.
In some embodiments, when the missing information determination is performed, whether an operation instruction corresponding to the voice information exists may be searched for in a mapping relationship saved in advance. If the voice information exists, determining that the voice information does not have missing information, otherwise, indicating that the voice information has missing information.
When the missing information is judged, optionally, if the judgment result represents that the voice information has no missing information, the terminal device can directly match the corresponding operation instruction according to the voice information and the mapping relation and respond.
Optionally, if the determination result indicates that the voice information has missing information, the corresponding operation instruction may be matched for the voice information according to the flow shown in the subsequent step S130, so as to avoid the failure of the voice control function, thereby improving the user experience.
Step S130: if so, determining the operation instruction with the highest matching degree with the voice information according to the historical data of the user.
The historical data of the user comprises various operation instructions triggered by the user and operation details of the various operation instructions.
It should be noted that the operation instruction in the history data may be triggered by the user through the voice message or may be triggered by the user through the physical key or the virtual key.
In addition, it is worth pointing out that the historical data can be stored locally in advance, and can also be obtained from the cloud in real time.
In some embodiments, the terminal device is an in-vehicle device, that is, the voice response scheme according to the embodiment of the present application is applied to an in-vehicle system. In this embodiment, since the same user may drive different vehicles (for example, the user may own multiple vehicles, or the user may rent a shared vehicle, etc.), in order to ensure the synchronization of the history data, the user may log in account information corresponding to the user to obtain history data generated in the past.
In the embodiment of the present application, when the determination result indicates that the voice information has missing information, the missing voice information may be matched to a corresponding operation instruction according to the historical data of the user, and the process will be described below.
Optionally, when the voice information with missing information is acquired, the standard voice information corresponding to each operation information included in the history data may be acquired first, and then the voice information with missing information is matched with each standard voice information, so as to obtain a matching result including a plurality of matching degrees (similarities).
Because the currently acquired voice information has missing information, the standard voice information with the matching degree of 100% cannot be matched with the currently acquired voice information. In this case, the operation command corresponding to the standard speech information corresponding to the highest matching degree value may be determined as the operation command corresponding to the speech information having the missing information, from among the plurality of matching degrees included in the matching result.
Still taking the mapping relationship described above as an example, when the obtained voice information is "open air", when the "open air" is respectively matched with the "open air conditioner" and the "open navigation", the matching degree of the "open air conditioner" is greater than the matching degree of the "open navigation", and at this time, the operation instruction a corresponding to the standard voice information "open air conditioner" corresponding to the highest matching degree is determined as the operation instruction corresponding to the voice information "open air".
In addition, in some application scenarios, when matching the speech information with the standard speech information, the matching result may represent that there are multiple standard speech information corresponding to the highest matching degree values. For example, the acquired voice information is "on", and when the "on" is respectively matched with the "air conditioner on" and the "navigation on", the matching degree of the "air conditioner on" is equal to the matching degree of the "navigation on".
In order to avoid the situation that the operation instruction corresponding to the voice information cannot be determined under the similar scenario, in some embodiments, the operation details of each operation instruction included in the history data may include a frequency of use corresponding to each operation instruction. In the above scenario, the operation command with the highest frequency of use may be determined as the operation command corresponding to the voice information from among the plurality of operation commands with the highest matching degree.
Step S140: responding to the corresponding operation instruction.
After the operation instruction corresponding to the voice information is determined, the terminal equipment can respond to the operation instruction, so that a corresponding function is triggered.
For example, in the above example, if the determined corresponding operation instruction is instruction a, the terminal device may respond to instruction a and start the air conditioner.
For another example, if the determined corresponding operation instruction is used to characterize and close the current application program, the terminal device may search for the application program currently running in the background and close the application program.
Further, in some embodiments, the historical data may also include habit data of the user. Wherein the habit data may include, but is not limited to, at least one of: the air conditioner temperature value using data of the user, the frequently listened song data of the user, the frequently used navigation data of the user and the like.
For example, when the habit data includes air conditioner temperature value usage data of the user, and when the corresponding operation instruction represents that the air conditioner is turned on, the air conditioner can be turned on by combining the air conditioner temperature value usage data of the user, and the air conditioner value is adjusted to a temperature value with the highest user frequency.
For example, when the habit data includes the frequently listened song data of the user, and when the corresponding operation instruction represents that the song is played, the player can be opened and the song with the highest use frequency can be played in combination with the frequently listened song data of the user.
In addition, if a song is already played and the song with the highest use frequency is played, if the corresponding operation instruction needs to be responded at this time and the corresponding operation instruction represents the song to be played, the currently played song can be switched to the song with the second highest use frequency.
For example, when the habit data includes the common navigation data of the user, and when the corresponding operation instruction represents that the navigation is opened, the navigation can be opened by combining the common navigation data of the user, and the route with the highest use frequency and taking the current positioning information as the starting point/the ending point in the common navigation data is displayed by combining the current positioning information.
Furthermore, in some embodiments, the habit data can be further subdivided according to different time periods, for example, divided into working day habit data, weekend habit data, early peak habit data, late peak habit data, average peak habit data, and the like.
Accordingly, in this embodiment, when responding to the corresponding operation instruction, the corresponding habit data can be determined according to the current time, and then the operation instruction is responded by combining the corresponding habit data.
For example, the current time is the weekend, the habit data comprises the frequently listened song data of the user on the weekend, and when the corresponding operation instruction represents that the song is played, the player can be turned on in combination with the frequently listened song data of the user on the weekend, and the song with the highest use frequency in the weekend time period is played.
In addition, it is mentioned in the foregoing that the voice response scheme according to the embodiment of the present application may be applied to a vehicle-mounted system. Optionally, the in-vehicle system may further include a presentation device, such as a voice player or a display screen.
In some embodiments, after the corresponding operation instruction is determined, description information for determining whether to execute the corresponding operation instruction may be further generated, and the description information is displayed through the display device.
Wherein, when the display device is a voice player, the display device displaying the description information may be voice playing: whether to perform the XX operation. For example, when the corresponding operation instruction represents turning on the air conditioner, the content of the voice playing may be: whether or not to turn on the air conditioner.
When the display device is a display screen, the display device may display the description information by text: whether to perform the XX operation. For example, when the corresponding operation instruction represents turning on the air conditioner, the displayed content may be: whether or not to turn on the air conditioner.
In addition, in some embodiments, when the description information is presented through the display screen, the related keywords may also be highlighted, for example, "air conditioner" is highlighted when "whether to turn on air conditioner" is displayed.
In this embodiment, the in-vehicle system may respond to the corresponding operation instruction after determining that the reverse instruction is not acquired. The reverse instruction is used for representing that the corresponding operation instruction is not executed, and is triggered by a user according to the description information, and the triggering mode can be voice triggering or key triggering.
Specifically, the vehicle-mounted system may determine that the reverse instruction is not acquired when the vehicle-mounted system acquires user feedback within a preset time period after the description information is displayed and represents and confirms user feedback information of the corresponding operation instruction. Through the strategy, the user can control the execution authority of the corresponding operation instruction, and further misoperation of the vehicle-mounted system can be avoided.
In addition, when the vehicle-mounted system does not acquire any user feedback information for the corresponding operation instruction within a preset time period after the description information is displayed, the representation user defaults to need to execute the corresponding operation instruction, and at this time, it can also be determined that the reverse instruction is not acquired.
According to the voice response method provided by the embodiment of the application, after the voice information is acquired, whether the voice information has missing information or not is judged, and when the voice information has missing information, the corresponding operation instruction can be matched for the voice information with the missing information through various operation instructions which are included in historical data of a user and are triggered by the user in the past, so that the failure of a voice control function can be avoided, and the user experience is improved.
As shown in fig. 2, an embodiment of the present application further provides a voice response apparatus 400, where the voice response apparatus 400 may include: an acquisition module 410, a determination module 420, a determination module 430, and a response module 440.
An obtaining module 410, configured to obtain voice information;
a determining module 420, configured to determine whether the voice information has missing information;
a determining module 430, configured to determine, according to historical data of a user, that an operation instruction with a highest matching degree with the voice information is an operation instruction corresponding to the voice information when the determining module determines that the operation instruction is yes;
a response module 440, configured to respond to the corresponding operation instruction.
In a possible implementation manner, the determining module 420 is configured to: searching whether an operation instruction corresponding to the voice information exists in a mapping relation stored in advance; if yes, determining that the voice information has no missing information; otherwise, determining that the voice information has missing information.
In a possible implementation manner, the method is applied to an on-board system, the on-board system comprises a display device, the device further comprises a display module, and the display module is used for generating description information used for determining whether to execute the corresponding operation instruction; displaying the description information to a user through the display equipment;
correspondingly, the response module 440 is configured to respond to the corresponding operation instruction after determining that the reverse instruction is not obtained; wherein the reverse instruction is used for characterizing not to execute the corresponding operation instruction.
In a possible implementation manner, the response module 440 is configured to determine that the reverse instruction is not acquired when user feedback information representing that it is confirmed to execute the corresponding operation instruction is acquired within a preset time period, or when any user feedback information for the corresponding operation instruction is not acquired within the preset time period.
In a possible implementation manner, the obtaining module 410 is further configured to obtain, through account information logged in by the user, historical data corresponding to the account information.
In a possible implementation manner, the historical data further includes usage frequencies corresponding to the various operation instructions, and the determining module 430 is configured to determine, when it is determined that there are a plurality of operation instructions with the highest matching degree, an operation instruction with the highest usage frequency among the plurality of operation instructions with the highest matching degree as the corresponding operation instruction.
In a possible implementation manner, the historical data further includes habit data of the user, and the response module 440 is configured to respond to the corresponding operation instruction by combining the habit data.
The voice response apparatus 400 provided in the embodiment of the present application has the same implementation principle and the same technical effect as those of the foregoing method embodiments, and for the sake of brief description, reference may be made to the corresponding contents in the foregoing method embodiments for the parts of the apparatus embodiments that are not mentioned.
In addition, an embodiment of the present application further provides a computer-readable storage medium, where a computer program is stored on the computer-readable storage medium, and when the computer program is executed by a computer, the steps included in the voice response method are executed.
In addition, referring to fig. 3, an embodiment of the present application further provides an electronic device 100 for implementing the voice response method and apparatus in the embodiment of the present application.
The electronic device 100 may include a terminal device, a home appliance device, a vehicle such as an electric vehicle, and the like.
Among them, the electronic device 100 may include: processor 110, memory 120, voice acquisition unit 130.
It should be noted that the components and structure of electronic device 100 shown in FIG. 3 are exemplary only, and not limiting, and electronic device 100 may have other components and structures as desired. For example, in some cases, electronic device 100 may also include a display screen for displaying some interface information for confirmation by the user.
The processor 110, the memory 120, the speech acquisition unit 130, and other components that may be present in the electronic device 100 are electrically connected to each other, directly or indirectly, to enable the transfer or interaction of data. For example, the processor 110, the memory 120, the voice capturing unit 130, and other components that may be present may be electrically connected to each other via one or more communication buses or signal lines.
The memory 120 is used for storing programs, such as programs corresponding to the voice response methods or voice response devices. Optionally, when the voice response device is stored in the memory 120, the voice response device includes at least one software function module that can be stored in the memory 120 in the form of software or firmware (firmware).
Alternatively, the software function module included in the voice response device may also be solidified in an Operating System (OS) of the electronic device 100.
The processor 110 is used to execute executable modules stored in the memory 120, such as software functional modules or computer programs included in the voice response apparatus. When the processor 110 receives the execution instruction, it may execute the computer program, for example, to perform: acquiring voice information; judging whether the voice information has missing information or not; if so, determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information according to the historical data of the user; the historical data comprises various operation instructions triggered by the user in the past; responding to the corresponding operation instruction.
Of course, the method disclosed in any of the embodiments of the present application can be applied to the processor 110, or implemented by the processor 110.
In summary, according to the voice response method, the voice response device, the electronic apparatus, and the computer-readable storage medium provided in the embodiments of the present invention, after the voice information is obtained, it is first determined whether the voice information has missing information, and when the voice information has missing information, the voice information having the missing information is matched with a corresponding operation instruction through various operation instructions that are included in the historical data of the user and are triggered by the user in the past, so that a failure of a voice control function can be avoided, and thus user experience is improved.
It should be noted that, in the present specification, the embodiments are all described in a progressive manner, each embodiment focuses on differences from other embodiments, and the same and similar parts among the embodiments may be referred to each other.
In the embodiments provided in the present application, it should be understood that the disclosed apparatus and method can be implemented in other ways. The apparatus embodiments described above are merely illustrative, and for example, the flowchart and block diagrams in the figures illustrate the architecture, functionality, and operation of possible implementations of apparatus, methods and computer program products according to various embodiments of the present application. In this regard, each block in the flowchart or block diagrams may represent a module, segment, or portion of code, which comprises one or more executable instructions for implementing the specified logical function(s). It should also be noted that, in some alternative implementations, the functions noted in the block may occur out of the order noted in the figures. For example, two blocks shown in succession may, in fact, be executed substantially concurrently, or the blocks may sometimes be executed in the reverse order, depending upon the functionality involved. It will also be noted that each block of the block diagrams and/or flowchart illustration, and combinations of blocks in the block diagrams and/or flowchart illustration, can be implemented by special purpose hardware-based systems which perform the specified functions or acts, or combinations of special purpose hardware and computer instructions.
In addition, functional modules in the embodiments of the present application may be integrated together to form an independent part, or each module may exist separately, or two or more modules may be integrated to form an independent part.
The functions, if implemented in the form of software functional modules and sold or used as a stand-alone product, may be stored in a computer-readable storage medium. Based on such understanding, the technical solution of the present application or portions thereof that substantially contribute to the prior art may be embodied in the form of a software product stored in a storage medium and including instructions for causing a computer device (which may be a personal computer, a notebook computer, a server, or a network device) to execute all or part of the steps of the method according to the embodiments of the present application. And the aforementioned storage medium includes: a U-disk, a removable hard disk, a Read-Only Memory (ROM), a Random Access Memory (RAM), a magnetic disk or an optical disk, and other various media capable of storing program codes.
The above description is only for the specific embodiments of the present application, but the scope of the present application is not limited thereto, and any person skilled in the art can easily conceive of the changes or substitutions within the technical scope of the present application, and shall be covered by the scope of the present application.

Claims (10)

1. A voice response method, characterized in that the method comprises:
acquiring voice information;
judging whether the voice information has missing information or not;
if so, determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information according to the historical data of the user; the historical data comprises various operation instructions triggered by the user in the past;
responding to the corresponding operation instruction.
2. The method of claim 1, wherein the determining whether the voice information has missing information comprises:
searching whether an operation instruction corresponding to the voice information exists in a mapping relation stored in advance;
if yes, determining that the voice information has no missing information;
otherwise, determining that the voice information has missing information.
3. The method of claim 1, wherein the method is applied to an in-vehicle system, the in-vehicle system comprising a display device, and prior to the responding to the corresponding operation instruction, the method further comprises:
generating description information for determining whether to execute the corresponding operation instruction;
displaying the description information to a user through the display equipment;
correspondingly, the responding to the corresponding operation instruction includes:
responding to the corresponding operation instruction after determining that the reverse instruction is not acquired;
wherein the reverse instruction is used for characterizing not to execute the corresponding operation instruction.
4. The method of claim 3, wherein the determining that a backward direction instruction is not fetched comprises:
and when user feedback information representing that the corresponding operation instruction is confirmed to be executed is acquired within a preset time period, or when any user feedback information aiming at the corresponding operation instruction is not acquired within the preset time period, determining that the reverse instruction is not acquired.
5. The method according to claim 1, wherein before the determining that the operation instruction with the highest matching degree with the voice information is the operation instruction corresponding to the voice information, the method further comprises:
and acquiring historical data corresponding to the account information through the account information logged in by the user.
6. The method according to any one of claims 1 to 5, wherein the history data further includes a usage frequency corresponding to the various operation instructions, and the determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information includes:
and when determining that a plurality of operation instructions with the highest matching degree exist, determining the operation instruction with the highest use frequency in the plurality of operation instructions with the highest matching degree as the corresponding operation instruction.
7. The method of claim 6, wherein the historical data further comprises habit data of the user, and wherein the responding to the corresponding operation instruction comprises:
responding to the corresponding operation instruction by combining the habit data.
8. A voice response apparatus, characterized in that the apparatus comprises:
the acquisition module is used for acquiring voice information;
the judging module is used for judging whether the voice information has missing information or not;
the determining module is used for determining the operation instruction with the highest matching degree with the voice information as the operation instruction corresponding to the voice information according to the historical data of the user when the judging module judges that the voice information is the voice information; the historical data comprises various operation instructions triggered by the user in the past;
and the response module is used for responding to the corresponding operation instruction.
9. An electronic device, comprising: the device comprises a voice acquisition unit, a memory and a processor, wherein the processor is respectively connected with the memory and the voice acquisition unit;
the memory is used for storing programs;
the processor calls a program stored in the memory to perform the method of any of claims 1-6.
10. A computer-readable storage medium, on which a computer program is stored which, when executed by a computer, performs the method of any one of claims 1-6.
CN202110524931.1A 2021-05-13 2021-05-13 Voice response method and device, electronic equipment and computer readable storage medium Pending CN113270096A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110524931.1A CN113270096A (en) 2021-05-13 2021-05-13 Voice response method and device, electronic equipment and computer readable storage medium

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110524931.1A CN113270096A (en) 2021-05-13 2021-05-13 Voice response method and device, electronic equipment and computer readable storage medium

Publications (1)

Publication Number Publication Date
CN113270096A true CN113270096A (en) 2021-08-17

Family

ID=77230749

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110524931.1A Pending CN113270096A (en) 2021-05-13 2021-05-13 Voice response method and device, electronic equipment and computer readable storage medium

Country Status (1)

Country Link
CN (1) CN113270096A (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024002298A1 (en) * 2022-07-01 2024-01-04 华为技术有限公司 Voice instruction processing method, apparatus and system, and storage medium

Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN106782526A (en) * 2016-12-12 2017-05-31 深圳Tcl数字技术有限公司 Sound control method and device
WO2017161829A1 (en) * 2016-03-25 2017-09-28 乐视控股(北京)有限公司 Voice signal information processing method and device
CN107908116A (en) * 2017-10-20 2018-04-13 深圳市艾特智能科技有限公司 Sound control method, intelligent domestic system, storage medium and computer equipment
CN108922529A (en) * 2018-07-04 2018-11-30 深圳万发创新进出口贸易有限公司 A kind of vehicle-mounted voice control system
CN109724215A (en) * 2018-06-27 2019-05-07 平安科技(深圳)有限公司 Air conditioning control method, air conditioning control device, air-conditioning equipment and storage medium
CN110019354A (en) * 2017-09-20 2019-07-16 杭州海康机器人技术有限公司 Control instruction generation method generates system, electronic equipment and storage medium
CN110019740A (en) * 2018-05-23 2019-07-16 京东方科技集团股份有限公司 Exchange method, car-mounted terminal, server and the storage medium of car-mounted terminal
CN110602307A (en) * 2018-06-12 2019-12-20 范世汶 Data processing method, device and equipment
CN111182390A (en) * 2019-12-27 2020-05-19 拉克诺德(深圳)科技有限公司 Volume data processing method and device, computer equipment and storage medium
CN111862974A (en) * 2020-07-15 2020-10-30 广州三星通信技术研究有限公司 Control method of intelligent equipment and intelligent equipment
CN112309384A (en) * 2019-08-28 2021-02-02 北京字节跳动网络技术有限公司 Voice recognition method, device, electronic equipment and medium

Patent Citations (11)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2017161829A1 (en) * 2016-03-25 2017-09-28 乐视控股(北京)有限公司 Voice signal information processing method and device
CN106782526A (en) * 2016-12-12 2017-05-31 深圳Tcl数字技术有限公司 Sound control method and device
CN110019354A (en) * 2017-09-20 2019-07-16 杭州海康机器人技术有限公司 Control instruction generation method generates system, electronic equipment and storage medium
CN107908116A (en) * 2017-10-20 2018-04-13 深圳市艾特智能科技有限公司 Sound control method, intelligent domestic system, storage medium and computer equipment
CN110019740A (en) * 2018-05-23 2019-07-16 京东方科技集团股份有限公司 Exchange method, car-mounted terminal, server and the storage medium of car-mounted terminal
CN110602307A (en) * 2018-06-12 2019-12-20 范世汶 Data processing method, device and equipment
CN109724215A (en) * 2018-06-27 2019-05-07 平安科技(深圳)有限公司 Air conditioning control method, air conditioning control device, air-conditioning equipment and storage medium
CN108922529A (en) * 2018-07-04 2018-11-30 深圳万发创新进出口贸易有限公司 A kind of vehicle-mounted voice control system
CN112309384A (en) * 2019-08-28 2021-02-02 北京字节跳动网络技术有限公司 Voice recognition method, device, electronic equipment and medium
CN111182390A (en) * 2019-12-27 2020-05-19 拉克诺德(深圳)科技有限公司 Volume data processing method and device, computer equipment and storage medium
CN111862974A (en) * 2020-07-15 2020-10-30 广州三星通信技术研究有限公司 Control method of intelligent equipment and intelligent equipment

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2024002298A1 (en) * 2022-07-01 2024-01-04 华为技术有限公司 Voice instruction processing method, apparatus and system, and storage medium

Similar Documents

Publication Publication Date Title
US11676601B2 (en) Voice assistant tracking and activation
US8494439B2 (en) Application state and activity transfer between devices
CN108595919B (en) Software activation method, device, system and storage medium
CN104520866A (en) Privacy protection method and terminal equipment
CN112017646A (en) Voice processing method and device and computer storage medium
EP3751795B1 (en) Group chat initiating method on basis of shared document, device and apparatus therefor, and storage medium
CN112135274B (en) Bluetooth scanning method, device, electronic equipment and readable storage medium
CN111427339B (en) Vehicle remote control method, device and server
CN110097884B (en) Voice interaction method and device
CN111131422A (en) Push message prompting method, vehicle-mounted terminal and computer storage medium
EP3435215B1 (en) Method, device, storage medium and mobile terminal for recognizing an off-screen gesture
CN113672748A (en) Multimedia information playing method and device
CN114844680A (en) Vehicle control method, device, equipment and readable storage medium
CN110949404B (en) Warning method and device, central control equipment, storage medium and system
CN113270096A (en) Voice response method and device, electronic equipment and computer readable storage medium
CN111752169A (en) Vehicle-mounted terminal control method, device and system
WO2023246390A1 (en) Data processing method and apparatus
CN112491690A (en) Method for transmitting voice information, mobile terminal, computer storage medium and system
CN110139259B (en) Bluetooth control method and system
RU2631269C2 (en) Method, device and system for telephone interaction
CN112526890A (en) Intelligent household control method and device and computer readable storage medium
CN114095488B (en) Information processing method, device, client and storage medium
CN115657991B (en) Display screen control method of intelligent automobile, domain controller and related device
CN113066493B (en) Equipment control method and system and first electronic equipment
KR20190115361A (en) Electronic device for managing application associated with a key of external electronic device and the method for the same

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination