CN114120987A - Voice awakening method, electronic equipment and chip system - Google Patents

Voice awakening method, electronic equipment and chip system Download PDF

Info

Publication number
CN114120987A
CN114120987A CN202010800181.1A CN202010800181A CN114120987A CN 114120987 A CN114120987 A CN 114120987A CN 202010800181 A CN202010800181 A CN 202010800181A CN 114120987 A CN114120987 A CN 114120987A
Authority
CN
China
Prior art keywords
equipment
voice information
value
wake
awakening
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010800181.1A
Other languages
Chinese (zh)
Other versions
CN114120987B (en
Inventor
牛庆建
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Honor Device Co Ltd
Original Assignee
Honor Device Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Honor Device Co Ltd filed Critical Honor Device Co Ltd
Priority to CN202010800181.1A priority Critical patent/CN114120987B/en
Publication of CN114120987A publication Critical patent/CN114120987A/en
Application granted granted Critical
Publication of CN114120987B publication Critical patent/CN114120987B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L17/00Speaker identification or verification
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Abstract

The embodiment of the application provides a voice awakening method, electronic equipment and a chip system, relates to the technical field of multi-equipment interaction, and can solve the problems that in a scene of multiple equipment, multiple pieces of electronic equipment can be awakened by voice information of a user, and the multiple pieces of electronic equipment can respond to the voice information of the user. The method comprises the following steps: under the condition that the voice information monitored by the first equipment is matched with the awakening voiceprint of the first equipment, the first equipment sends the competitive value of the first equipment on the voice information to the second equipment and monitors the information sent by the second equipment, wherein the competitive value of the first equipment on the voice information is the supporting capacity of the first equipment on the awakening operation corresponding to the voice information; in response to the monitored information sent by the second device, the first device determines whether the first device is a wake-up device based on the information sent by the second device; and if the first equipment is awakening equipment, the first equipment executes awakening operation corresponding to the voice information.

Description

Voice awakening method, electronic equipment and chip system
Technical Field
The embodiment of the application relates to the field of multi-device interaction, in particular to a voice awakening method for multiple devices, electronic equipment and a chip system.
Background
With the increasing degree of intelligence of electronic devices, more and more electronic devices support the human-computer interaction function, and voice wake-up is very important as an entrance for human-computer interaction between users and the electronic devices.
Usually, the voice wake-up words of electronic devices of the same brand are the same, and even the voice wake-up words of electronic devices of different brands may be the same. When a plurality of electronic devices with the same voice wake-up word exist in an application scene, for example, when the voice wake-up words corresponding to the wake-up instructions of the plurality of electronic devices in a home environment are all "small E", a user speaks the voice wake-up word "small E" corresponding to the wake-up instruction, and possibly all the plurality of electronic devices can be woken up in the scene; when the user continues to speak the voice wake-up word "play music" corresponding to the action command, it is possible that all of the electronic devices in the scene respond to the voice wake-up word.
Disclosure of Invention
The embodiment of the application provides a voice awakening method, electronic equipment and a chip system, and solves the problems that in a scene with multiple pieces of equipment, the multiple pieces of electronic equipment are awakened by voice information of a user, and the multiple pieces of electronic equipment respond to the voice information of the user.
In order to achieve the purpose, the technical scheme is as follows:
in a first aspect, an embodiment of the present application provides a voice wake-up method, including:
responding to the monitored voice information, and matching the voice information with the awakening voiceprint of the first equipment by the first equipment; if the voice information is matched with the awakening voiceprint of the first device, the first device sends a bidding value of the first device to the voice information to the second device and monitors information sent by the second device, wherein the bidding value of the first device to the voice information is the support capability of the first device to the awakening operation corresponding to the voice information; in response to the monitored information sent by the second device, the first device determines whether the first device is a wake-up device based on the information sent by the second device; and if the first equipment is awakening equipment, the first equipment executes awakening operation corresponding to the voice information.
In the embodiment of the application, the bidding value of the electronic equipment for the voice information is set for each electronic equipment, the awakening equipment is determined through the bidding value interaction process between the electronic equipment (between the first equipment and the second equipment), and the awakening operation corresponding to the voice information is executed through the determined awakening equipment, so that the problems that in the application scene of a plurality of electronic equipment, the electronic equipment can be awakened by the voice information of a user, and the electronic equipment can respond to the voice information of the user are solved.
In a possible implementation manner of the first aspect, the second device is an electronic device located in the same network as the first device, and the information sent by the second device includes a bid value of the second device for the voice information;
correspondingly, the determining, by the first device, whether the first device is a wake-up device based on the information sent by the second device includes: if at least one bidding value of the second equipment on the voice information is higher than the bidding value of the first equipment on the voice information in the bidding values of the second equipment on the voice information received by the first equipment within the first preset time, determining that the first equipment is not awakening equipment; and if the bidding values of the second equipment to the voice information received by the first equipment within the first preset time are all lower than the bidding values of the first equipment to the voice information, determining that the first equipment is awakening equipment.
In a possible implementation manner of the first aspect, the determining, by the first device, whether the first device is a wake-up device based on the information sent by the second device further includes:
if the bid value of the second equipment for the voice information received by the first equipment within the first preset time is not higher than the bid value of the first equipment for the voice information and is equal to the bid value of the first equipment for the voice information, marking the second equipment with the bid value equal to the bid value of the first equipment for the voice information as third equipment, and acquiring state information of the third equipment; the first device determines whether the first device is a wake-up device or not based on state information of the first device and state information of the third device, wherein the state information includes at least one of the following state parameters: the total time of the starting-up state, the position of the equipment, the residual electric quantity and whether the equipment is in the use state within the second preset time.
In the embodiment of the application, the first device serves as an electronic device in a multi-device application scene, and after receiving the competitive value of the second device, the first device determines whether the first device is a wake-up device or not by analyzing various relations between the competitive value of the second device and the competitive value of the first device; and the process of analyzing the various relationships between the bid value of the second device and the bid value of the first device to determine whether the first device is a wake-up device is based on the principle that the electronic device with the highest bid value is taken as the wake-up device and the electronic device with the non-highest bid value is not taken as the wake-up device. Therefore, the determined awakening equipment is the best equipment for executing the awakening operation corresponding to the voice information.
In a possible implementation manner of the first aspect, if the bid value of the second device for the voice information received by the first device within the first preset time is lower than the bid value of the first device for the voice information, determining that the first device is a wake-up device includes:
if the voice information bidding value of the second equipment received by the first equipment within the first preset time is lower than the voice information bidding value of the first equipment, the first equipment determines whether the voice information bidding value of the first equipment is higher than a preset value; and if the bidding value of the first equipment on the voice information is higher than the preset value, determining that the first equipment is awakening equipment.
In the embodiment of the application, the condition that the awakening operation corresponding to the voice information of the user exceeds the execution capacity of the electronic equipment with the highest bidding value is analyzed, so that the electronic equipment determined as the awakening equipment has the capacity of executing the awakening operation corresponding to the voice information.
In a possible implementation manner of the first aspect, after the determining, by the first device, whether the bid value of the first device for the voice information is higher than a preset value, the method further includes:
if the bidding value of the first equipment to the voice information is lower than or equal to a preset value, the first equipment determines whether a fourth equipment exists in a network where the first equipment is located, wherein the fourth equipment is electronic equipment, voice print information of which is matched with the voice information, in second equipment, the first equipment does not monitor the bidding value within a first preset time; if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice information to the fourth device so as to indicate the fourth device to execute wakeup operation corresponding to the voice information under the condition that the bid value of the fourth device for the voice information is higher than the preset value.
In the embodiment of the application, in an application scenario, when the electronic device that monitors the voice information of the user has no execution capability for the wake-up operation corresponding to the voice information, and the electronic device that does not monitor the voice information of the user due to a long distance has an execution capability for the wake-up operation of the voice information, it is described how to execute the wake-up operation of the voice information on the electronic device that has the execution capability for the wake-up operation of the voice information by not monitoring the voice information of the user; therefore, the application scene of the embodiment of the application is enlarged, and the experience degree of the user is improved.
In a possible implementation manner of the first aspect, the second device is an electronic device located in the same network as the first device, and the method further includes:
if the first equipment does not receive the information sent by the second equipment within the first preset time, the first equipment determines whether the bidding value of the first equipment on the voice information is higher than a preset value or not; if the bidding value of the first equipment on the voice information is higher than the preset value, the first equipment determines that the first equipment is awakening equipment; if the bidding value of the first equipment to the voice information is lower than or equal to a preset value, the first equipment determines whether fourth equipment exists in a network where the first equipment is located, wherein the fourth equipment is electronic equipment, voice print information of which is matched with the voice information, in second equipment, the first equipment does not monitor the bidding value within first preset time; if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice information to the fourth device to indicate that the fourth device executes wakeup operation corresponding to the voice information under the condition that the bid value of the fourth device for the voice information is higher than the preset value.
The implementation mode enlarges the application scene of the embodiment of the application, and when only one electronic device capable of monitoring the voice information of the user is available and the bid value of the electronic device for the voice information is smaller than the preset value, the process of comparing the bid values of the electronic devices does not exist. In this case, however, it is necessary to find whether there is an electronic apparatus having an ability to perform a wake-up operation on voice information because the user's voice information is not heard at a long distance, as described in the above implementation. And executing the awakening operation corresponding to the voice information through the electronic equipment. Therefore, the application scene of the embodiment of the application is expanded, and the experience degree of a user is improved.
In a possible implementation manner of the first aspect, before the determining, by the first device, whether a fourth device exists in a network in which the first device is located, the method further includes:
the first device acquires a wakeup voiceprint of a second device, wherein the first device does not monitor the bid value within a first preset time.
In a possible implementation manner of the first aspect, the obtaining, by the first device, a wakeup voiceprint of a second device for which the first device does not monitor the bid value within a first preset time includes:
when the first equipment is preset voiceprint summarizing equipment, the first equipment receives awakening voiceprints of other electronic equipment except the first equipment in a network where the first equipment is located; the first device generates a voiceprint characteristic summary based on the awakening voiceprint of the first device and the received awakening voiceprints of other electronic devices, and sends the voiceprint characteristic summary to the other electronic devices; the method comprises the steps that a first device obtains a wakeup voiceprint of a second device, wherein the first device does not monitor the competitive value within a first preset time, from a voiceprint characteristic collection;
when the first device is not the preset voiceprint collecting device, the first device sends the awakening voiceprint of the first device to the preset voiceprint collecting device so as to instruct the voiceprint collecting device to generate voiceprint characteristic collection of each electronic device in the network where the voiceprint collecting device is located; the first equipment receives voiceprint characteristic summary sent by the voiceprint summary equipment; the first device obtains the awakening voiceprint of the second device, which does not monitor the competitive value within the first preset time, from the voiceprint characteristic collection.
In this embodiment of the application, in order to ensure that the first device does not have the capability of performing the wakeup operation corresponding to the voice information, the first device needs to acquire the wakeup voiceprint of each electronic device in advance to discover whether the fourth device exists or not, in order to discover the fourth device that does not monitor the voice information of the user and matches the wakeup voiceprint with the voice information. And earlier guarantee is provided for the implementation of the embodiment of the application.
In a possible implementation manner of the first aspect, the second device is a preset decision device, and the decision device is configured to receive bid values of the voice information sent by each electronic device in a network where the decision device is located, send a first message to an electronic device with a highest bid value, and send a second message to an electronic device with a non-highest bid value;
correspondingly, the determining, by the first device, whether the first device is a wake-up device based on the information sent by the second device includes: if the information sent by the second device and received by the first device is a first message, the first device determines that the first device is a wake-up device; and if the information sent by the second device and received by the first device is the second message, the first device determines that the first device is not the awakening device.
In the embodiment of the application, another bidding process of a plurality of electronic devices is also provided, namely, a decision device is preset, and the plurality of electronic devices all send respective bidding values to the decision device; and the decision-making equipment determines the electronic equipment with the highest bidding value as the awakening equipment by comparing the received bidding values of the plurality of electronic equipment. The process of determining the awakening device is described from another perspective by a mode of setting the decision device, so that the flexibility of the voice awakening method provided by the embodiment of the application in implementation is improved.
In a possible implementation manner of the first aspect, the sending, by the first device, the bid value for the voice information by the first device to the second device includes:
the first device acquires an awakening mode corresponding to the voice information, wherein the awakening mode comprises: waking up a scene; the method comprises the steps that first equipment obtains a bidding value of the first equipment in an awakening mode, and the bidding value is used as the bidding value of the first equipment for voice information; the first device sends the bid value of the first device for the voice information to the second device.
In the embodiment of the application, the voice information can be classified into a plurality of awakening scenes according to the semantics of the voice information, and the competitive value of the electronic equipment for the voice information is determined according to the competitive value of the electronic equipment in different awakening scenes, so that the problem that the electronic equipment can acquire the competitive value of the voice information only when a user awakens the electronic equipment by using specific voice information is solved.
In a possible implementation manner of the first aspect, the obtaining, by the first device, a bid value of the first device in an awake mode includes:
the method comprises the steps that a first device obtains a preset competitive value of the first device in an awakening mode;
or the first device obtains the device parameters of the first device and the dependency of the wake-up mode on the device parameters; the first device generates a bid value of the first device in the wake mode based on the device parameters of the first device and the dependency of the wake mode on the device parameters.
The embodiment of the application provides two processes for obtaining the bid value of the first device for the voice information, which can be preset, and also can calculate the bid value of each electronic device in different wake-up modes according to the device parameters of each electronic device and the dependence of different wake-up modes on the electronic device. The preset mode is simple and convenient, and the efficiency is high. The method is more easily adapted to different application scenes in a calculation mode and is more accurate. The flexibility of the voice awakening method provided by the embodiment of the application in implementation is improved through two modes.
In a possible implementation manner of the first aspect, the wake mode further includes: a wake-up location; the first device acquiring the wake-up mode corresponding to the voice information comprises: the first equipment determines a wake-up place corresponding to the voice information based on a network to which the first equipment is currently connected.
In the embodiment of the present application, in order to meet different application scenarios, a wake-up location may be further added as a factor in the wake-up mode. Therefore, the voice awakening method provided by the embodiment of the application can adapt to more scenes in implementation.
In a second aspect, an embodiment of the present application provides an electronic device, including:
the voiceprint matching unit is used for responding to the monitored voice information and matching the voice information with the awakening voiceprint of the electronic equipment;
the information interaction unit is used for sending a bidding value of the electronic equipment to the voice information to the second equipment and monitoring the information sent by the second equipment if the voice information is matched with the awakening voiceprint of the electronic equipment, wherein the bidding value of the electronic equipment to the voice information is the support capability of the electronic equipment to the awakening operation corresponding to the voice information;
the bidding unit is used for responding to the monitored information sent by the second equipment and determining whether the electronic equipment is awakening equipment or not based on the information sent by the second equipment;
and the awakening unit is used for executing awakening operation corresponding to the voice information if the electronic equipment is awakening equipment.
In a third aspect, an electronic device is provided, comprising a processor for executing a computer program stored in a memory, to implement the method of any one of the first aspect of the present application.
In a fourth aspect, a chip system is provided, which includes a processor coupled to a memory, the processor executing a computer program stored in the memory to implement the method of any one of the first aspect of the present application.
In a fifth aspect, there is provided a computer readable storage medium storing a computer program which, when executed by one or more processors, performs the method of any one of the first aspects of the present application.
In a sixth aspect, embodiments of the present application provide a computer program product, which when run on a device, causes the device to perform any one of the methods of the first aspect.
It is understood that the beneficial effects of the second to sixth aspects can be found in the related descriptions of the first aspect, and are not described herein again.
Drawings
Fig. 1 is a schematic view of an application scenario of a voice wake-up method according to an embodiment of the present application;
fig. 2 is a schematic diagram of a hardware structure of an electronic device that executes a voice wake-up method according to an embodiment of the present disclosure;
fig. 3 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
FIG. 4 is a schematic diagram of an interface for setting bid values of different wake modes for different electronic devices according to an embodiment of the present disclosure;
FIG. 5 is an illustration of another interface for setting bid values for different wake modes for different electronic devices according to an embodiment of the disclosure;
fig. 6 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
fig. 7 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
FIG. 8 is a schematic diagram of an application framework based on the voice wake-up method provided in FIG. 7;
fig. 9 is a schematic view of an application scenario of the voice wake-up method according to the embodiment shown in fig. 6;
fig. 10 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
fig. 11 is a schematic view of an application scenario of the voice wake-up method according to the embodiment shown in fig. 10;
fig. 12 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
fig. 13 is a schematic view of an application scenario of the voice wake-up method according to the embodiment shown in fig. 12;
fig. 14 is an interaction diagram illustrating that in the voice wake method provided in the embodiment shown in fig. 12, the electronic device obtains wake voiceprints of other electronic devices;
fig. 15 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application;
fig. 16 is a schematic view of an application scenario of the voice wake-up method according to the embodiment shown in fig. 15;
fig. 17 is a schematic block diagram of functional architecture modules of an electronic device that executes a voice wake-up method according to an embodiment of the present application.
Detailed Description
In the following description, for purposes of explanation and not limitation, specific details are set forth such as particular system structures, techniques, etc. in order to provide a thorough understanding of the embodiments of the present application. It will be apparent, however, to one skilled in the art that the present application may be practiced in other embodiments that depart from these specific details.
It will be understood that the terms "comprises" and/or "comprising," when used in this specification and the appended claims, specify the presence of stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or groups thereof.
It should also be understood that in the embodiments of the present application, "one or more" means one, two, or more than two; "and/or" describes the association relationship of the associated objects, indicating that three relationships may exist; for example, a and/or B, may represent: a alone, both A and B, and B alone, where A, B may be singular or plural. The character "/" generally indicates that the former and latter associated objects are in an "or" relationship.
Furthermore, in the description of the present application and the appended claims, the terms "first," "second," "third," and the like are used for distinguishing between descriptions and not necessarily for describing or implying relative importance.
Reference throughout this specification to "one embodiment" or "some embodiments," or the like, means that a particular feature, structure, or characteristic described in connection with the embodiment is included in one or more embodiments of the present application. Thus, appearances of the phrases "in one embodiment," "in some embodiments," "in other embodiments," or the like, in various places throughout this specification are not necessarily all referring to the same embodiment, but rather "one or more but not all embodiments" unless specifically stated otherwise. The terms "comprising," "including," "having," and variations thereof mean "including, but not limited to," unless expressly specified otherwise.
The voice wake-up method provided by the embodiment of the application can be applied to a scene with multiple devices, as shown in fig. 1, fig. 1 is an application scene of the voice wake-up method provided by the embodiment of the application. In this application scenario, there are a plurality of electronic devices, which are respectively: the mobile phone comprises a mobile phone 1 of a user A, a mobile phone 2 of a user B, an intelligent screen, an intelligent sound box and a wireless router. The wireless router converts the broadband network signal into a wireless signal and provides a wireless network. Cell-phone 1, cell-phone 2, wisdom screen and smart sound box all connect in the wireless network that wireless router provided. The user A inputs voiceprint information of the user A into the mobile phone 1, the intelligent screen and the intelligent sound box in advance to serve as awakening voiceprints of the mobile phone 1, the intelligent screen and the intelligent sound box, and the user B inputs voiceprint information of the user B into the mobile phone 2 and the intelligent screen in advance to serve as awakening voiceprints of the mobile phone 2 and the intelligent screen. The awakening voiceprint of the intelligent screen comprises voiceprint information of a user A and voiceprint information of a user B.
User a utters voice information in the application scenario, e.g., "little E, put a song to me". The mobile phone 1, the mobile phone 2, the smart screen and the smart sound box in the application scene can monitor the voice information sent by the user A, and the mobile phone 1, the mobile phone 2, the smart screen and the smart sound box respectively match the monitored voice information with respective awakening voiceprints.
It should be noted that, at this time, the matching operation performed by the electronic device does not determine whether the wake-up voiceprint of the electronic device and the monitored voice information are semantically matched, but determines whether the wake-up voiceprint of the electronic device and the monitored voice information are timbre matched, that is, determines whether the user who sends the current voice information has the capability of waking up the electronic device.
As before, the awakening voiceprints of the mobile phone 1, the smart screen and the smart speaker are voiceprint information of the user a, that is, the awakening voiceprints of the mobile phone 1, the smart screen and the smart speaker are matched with the currently monitored voice information, then the mobile phone 1, the smart screen and the smart speaker respectively determine respective competitive values of the voice information, through data interaction of the respective competitive values between the mobile phone 1, the smart screen and the smart speaker, one of the mobile phone 1, the smart screen and the smart speaker is determined as awakening equipment, and the awakening equipment executes awakening operation corresponding to the voice information. The bidding value of each electronic device in the application scene on the voice information can be preset.
The embodiment of the application provides a voice awakening method which can be applied to electronic equipment. The electronic device may be: the mobile phone includes an electronic device such as a mobile phone, a tablet computer, a smart screen, a wearable device, an on-board device, a smart sound box, a smart screen, an Augmented Reality (AR)/Virtual Reality (VR) device, a notebook computer, an ultra-mobile personal computer (UMPC), a netbook, and a Personal Digital Assistant (PDA). The embodiment of the present application does not limit the specific type of the electronic device.
Fig. 2 shows a schematic structural diagram of an electronic device. The electronic device 200 may include a processor 210, an external memory interface 220, an internal memory 221, a Universal Serial Bus (USB) interface 230, a charging management module 240, a power management module 241, a battery 242, an antenna 1, an antenna 2, a mobile communication module 250, a wireless communication module 260, an audio module 270, a speaker 270A, a receiver 270B, a microphone 270C, an earphone interface 270D, a sensor module 280, keys 290, a motor 291, an indicator 292, a camera 293, a display 294, and a Subscriber Identification Module (SIM) card interface 295, among others. The sensor module 280 may include a pressure sensor 280A, a gyroscope sensor 280B, an air pressure sensor 280C, a magnetic sensor 280D, an acceleration sensor 280E, a distance sensor 280F, a proximity light sensor 280G, a fingerprint sensor 280H, a temperature sensor 280J, a touch sensor 280K, an ambient light sensor 280L, a bone conduction sensor 280M, and the like.
It is to be understood that the illustrated structure of the embodiment of the present application does not specifically limit the electronic device 200. In other embodiments of the present application, the electronic device 200 may include more or fewer components than illustrated, or some components may be combined, some components may be separated, or a different arrangement of components may be used. The illustrated components may be implemented in hardware, software, or a combination of software and hardware.
Processor 210 may include one or more processing units, such as: the processor 210 may include an Application Processor (AP), a modem processor, a Graphics Processing Unit (GPU), an Image Signal Processor (ISP), a controller, a memory, a video codec, a Digital Signal Processor (DSP), a baseband processor, and/or a neural-Network Processing Unit (NPU), etc. The different processing units may be separate devices or may be integrated into one or more processors. For example, the processor 210 is configured to execute the voice wakeup method in the embodiment of the present application, for example, the following steps 301 to 304.
The controller may be, among other things, a neural center and a command center of the electronic device 200. The controller can generate an operation control signal according to the instruction operation code and the timing signal to complete the control of instruction fetching and instruction execution.
A memory may also be provided in processor 210 for storing instructions and data. In some embodiments, the memory in the processor 210 is a cache memory. The memory may hold instructions or data that have just been used or recycled by the processor 210. If the processor 210 needs to reuse the instruction or data, it may be called directly from memory. Avoiding repeated accesses reduces the latency of the processor 210, thereby increasing the efficiency of the system.
In some embodiments, processor 210 may include one or more interfaces. The interface may include an integrated circuit (I2C) interface, an integrated circuit built-in audio (I2S) interface, a Pulse Code Modulation (PCM) interface, a universal asynchronous receiver/transmitter (UART) interface, a Mobile Industry Processor Interface (MIPI), a general-purpose input/output (GPIO) interface, a Subscriber Identity Module (SIM) interface, and/or a Universal Serial Bus (USB) interface, etc.
The I2C interface is a bi-directional synchronous serial bus that includes a serial data line (SDA) and a Serial Clock Line (SCL). In some embodiments, processor 210 may include multiple sets of I2C buses. The processor 210 may be coupled to the touch sensor 280K, the charger, the flash, the camera 293, etc. through different I2C bus interfaces. For example: the processor 210 may be coupled to the touch sensor 280K via an I2C interface, such that the processor 210 and the touch sensor 280K communicate via an I2C bus interface to implement the touch function of the electronic device 200.
The I2S interface may be used for audio communication. In some embodiments, processor 210 may include multiple sets of I2S buses. Processor 210 may be coupled to audio module 270 via an I2S bus to enable communication between processor 210 and audio module 270. In some embodiments, audio module 270 may communicate audio signals to wireless communication module 260 through an I2S interface.
The PCM interface may also be used for audio communication, sampling, quantizing and encoding analog signals. In some embodiments, audio module 270 and wireless communication module 260 may be coupled by a PCM bus interface.
In some embodiments, the audio module 270 may also transmit audio signals to the wireless communication module 260 through the PCM interface, so as to implement a function of answering a call through a bluetooth headset. Both the I2S interface and the PCM interface may be used for audio communication.
The UART interface is a universal serial data bus used for asynchronous communications. The bus may be a bidirectional communication bus. It converts the data to be transmitted between serial communication and parallel communication.
In some embodiments, a UART interface is generally used to connect the processor 210 with the wireless communication module 260. For example: the processor 210 communicates with the bluetooth module in the wireless communication module 260 through the UART interface to implement the bluetooth function. In some embodiments, the audio module 270 may communicate the audio signal to the wireless communication module 260 through a UART interface.
The MIPI interface may be used to connect the processor 210 with peripheral devices such as the display screen 294, the camera 293, and the like. The MIPI interface includes a Camera Serial Interface (CSI), a Display Serial Interface (DSI), and the like. In some embodiments, processor 210 and camera 293 communicate via a CSI interface to implement the capture functionality of electronic device 200. The processor 210 and the display screen 294 communicate through the DSI interface to implement a display function of the electronic device 200.
The GPIO interface may be configured by software. The GPIO interface may be configured as a control signal and may also be configured as a data signal. In some embodiments, a GPIO interface may be used to connect processor 210 with camera 293, display 294, wireless communication module 260, audio module 270, sensor module 280, and the like. The GPIO interface may also be configured as an I2C interface, an I2S interface, a UART interface, a MIPI interface, and the like.
The USB interface 230 is an interface conforming to the USB standard specification, and may specifically be a Mini USB interface, a Micro USB interface, a USB Type C interface, or the like. The USB interface 230 may be used to connect a charger to charge the electronic device 200, and may also be used to transmit data between the electronic device 200 and a peripheral device.
It should be understood that the interfacing relationship between the modules illustrated in the embodiments of the present application is only an illustration, and does not limit the structure of the electronic device 200. In other embodiments of the present application, the electronic device 200 may also adopt different interface connection manners or a combination of multiple interface connection manners in the above embodiments.
The charge management module 240 is configured to receive a charging input from a charger. The charger may be a wireless charger or a wired charger. In some wired charging embodiments, the charging management module 240 may receive charging input from a wired charger via the USB interface 230. In some wireless charging embodiments, the charging management module 240 may receive a wireless charging input through a wireless charging coil of the electronic device 200. The charging management module 240 may also supply power to the electronic device through the power management module 241 while charging the battery 242.
The power management module 241 is used to connect the battery 242, the charging management module 240 and the processor 210. The power management module 241 receives input from the battery 242 and/or the charging management module 240, and provides power to the processor 210, the internal memory 221, the external memory, the display 294, the camera 293, and the wireless communication module 260. The power management module 241 may also be used to monitor parameters such as battery capacity, battery cycle number, battery state of health (leakage, impedance), etc.
In some other embodiments, the power management module 241 may also be disposed in the processor 210. In other embodiments, the power management module 241 and the charging management module 240 may be disposed in the same device.
The wireless communication function of the electronic device 200 may be implemented by the antenna 1, the antenna 2, the mobile communication module 250, the wireless communication module 260, a modem processor, a baseband processor, and the like.
The antennas 1 and 2 are used for transmitting and receiving electromagnetic wave signals. Each antenna in the electronic device 200 may be used to cover a single or multiple communication bands. Different antennas can also be multiplexed to improve the utilization of the antennas. For example: the antenna 1 may be multiplexed as a diversity antenna of a wireless local area network. In other embodiments, the antenna may be used in conjunction with a tuning switch.
The mobile communication module 250 may provide a solution including 2G/3G/4G/5G wireless communication applied to the electronic device 200. The mobile communication module 250 may include at least one filter, a switch, a power amplifier, a Low Noise Amplifier (LNA), and the like. The mobile communication module 250 may receive the electromagnetic wave from the antenna 1, filter, amplify, etc. the received electromagnetic wave, and transmit the electromagnetic wave to the modem processor for demodulation. The mobile communication module 250 may also amplify the signal modulated by the modem processor, and convert the signal into electromagnetic wave through the antenna 1 to radiate the electromagnetic wave.
In some embodiments, at least some of the functional modules of the mobile communication module 250 may be disposed in the processor 210. In some embodiments, at least some of the functional modules of the mobile communication module 250 may be disposed in the same device as at least some of the modules of the processor 210.
The modem processor may include a modulator and a demodulator. The modulator is used for modulating a low-frequency baseband signal to be transmitted into a medium-high frequency signal. The demodulator is used for demodulating the received electromagnetic wave signal into a low-frequency baseband signal. The demodulator then transmits the demodulated low frequency baseband signal to a baseband processor for processing. The low frequency baseband signal is processed by the baseband processor and then passed to the application processor. The application processor outputs audio signals through an audio device (not limited to speaker 270A, receiver 270B, etc.) or displays images or video through display screen 294. In some embodiments, the modem processor may be a stand-alone device. In other embodiments, the modem processor may be provided in the same device as the mobile communication module 250 or other functional modules, independent of the processor 210.
The wireless communication module 260 may provide a solution for wireless communication applied to the electronic device 200, including Wireless Local Area Networks (WLANs) (e.g., wireless fidelity (Wi-Fi) networks), Bluetooth (BT), Global Navigation Satellite System (GNSS), Frequency Modulation (FM), Near Field Communication (NFC), Infrared (IR), and the like. The wireless communication module 260 may be one or more devices integrating at least one communication processing module. The wireless communication module 260 receives electromagnetic waves via the antenna 2, performs frequency modulation and filtering processing on electromagnetic wave signals, and transmits the processed signals to the processor 210. The wireless communication module 260 may also receive a signal to be transmitted from the processor 210, frequency-modulate and amplify the signal, and convert the signal into electromagnetic waves via the antenna 2 to radiate the electromagnetic waves.
In some embodiments, antenna 1 of electronic device 200 is coupled to mobile communication module 250 and antenna 2 is coupled to wireless communication module 260, such that electronic device 200 may communicate with networks and other devices via wireless communication techniques. The wireless communication technology may include global system for mobile communications (GSM), General Packet Radio Service (GPRS), code division multiple access (code division multiple access, CDMA), Wideband Code Division Multiple Access (WCDMA), time-division code division multiple access (time-division code division multiple access, TD-SCDMA), Long Term Evolution (LTE), BT, GNSS, WLAN, NFC, FM, and/or IR technologies, etc. GNSS may include Global Positioning System (GPS), global navigation satellite system (GLONASS), beidou satellite navigation system (BDS), quasi-zenith satellite system (QZSS), and/or Satellite Based Augmentation System (SBAS).
The electronic device 200 implements display functions via the GPU, the display screen 294, and the application processor. The GPU is a microprocessor for image processing, and is connected to the display screen 294 and an application processor. The GPU is used to perform mathematical and geometric calculations for graphics rendering. Processor 210 may include one or more GPUs that execute program instructions to generate or alter display information.
The display screen 294 is used to display images, video, and the like. The display screen 294 includes a display panel. The display panel may adopt a Liquid Crystal Display (LCD), an organic light-emitting diode (OLED), an active-matrix organic light-emitting diode (active-matrix organic light-emitting diode, AMOLED), a flexible light-emitting diode (FLED), a miniature, a Micro-oeld, a quantum dot light-emitting diode (QLED), and the like. In some embodiments, the electronic device 200 may include 1 or N display screens 294, N being a positive integer greater than 1.
The electronic device 200 may implement a shooting function through the ISP, the camera 293, the video codec, the GPU, the display screen 294, and the application processor.
The ISP is used to process the data fed back by the camera 293. For example, when a photo is taken, the shutter is opened, light is transmitted to the camera photosensitive element through the lens, the optical signal is converted into an electrical signal, and the camera photosensitive element transmits the electrical signal to the ISP for processing and converting into an image visible to the naked eye. The ISP can also carry out algorithm optimization on the noise, brightness and skin color of the image. The ISP can also optimize parameters such as exposure, color temperature and the like of a shooting scene. In some embodiments, the ISP may be provided in camera 293.
The camera 293 is used to capture still images or video. The object generates an optical image through the lens and projects the optical image to the photosensitive element. The photosensitive element may be a Charge Coupled Device (CCD) or a complementary metal-oxide-semiconductor (CMOS) phototransistor. The light sensing element converts the optical signal into an electrical signal, which is then passed to the ISP where it is converted into a digital image signal. The ISP outputs the digital image signal to the DSP for processing. The DSP converts the digital image signal into image signal in standard RGB, YUV and other formats. In some embodiments, electronic device 200 may include 1 or N cameras 293, N being a positive integer greater than 1.
The digital signal processor is used for processing digital signals, and can process digital image signals and other digital signals.
Video codecs are used to compress or decompress digital video. The electronic device 200 may support one or more video codecs. In this way, the electronic device 200 may play or record video in a variety of encoding formats, such as: moving Picture Experts Group (MPEG) 1, MPEG2, MPEG3, MPEG4, and the like.
The NPU is a neural-network (NN) computing processor, which processes input information quickly by referring to a biological neural network structure, for example, by referring to a transfer mode between neurons of a human brain, and can also learn by itself continuously. Applications such as intelligent recognition of the electronic device 200 can be realized by the NPU, for example: image recognition, face recognition, speech recognition, text understanding, and the like.
In the embodiment of the present application, the NPU or other processor may be configured to perform operations such as face detection, face tracking, face feature extraction, and image clustering on a face image in a video stored in the electronic device 200; the operations such as face detection, face feature extraction and the like are performed on the face images in the pictures stored in the electronic device 200, and the pictures stored in the electronic device 200 are clustered according to the face features of the pictures and the clustering results of the face images in the video.
The external memory interface 220 may be used to connect an external memory card, such as a Micro SD card, to extend the memory capability of the electronic device 200. The external memory card communicates with the processor 210 through the external memory interface 220 to implement a data storage function. For example, files such as music, video, etc. are saved in an external memory card.
Internal memory 221 may be used to store computer-executable program code, which includes instructions. The processor 210 executes various functional applications of the electronic device 200 and data processing by executing instructions stored in the internal memory 221. The internal memory 221 may include a program storage area and a data storage area. The storage program area may store an operating system, and an application program (such as a sound playing function, an image playing function, and the like) required by at least one function. The stored data area may store data created during use of the electronic device 200 (e.g., a bid, a wake-up voiceprint, etc.).
In addition, the internal memory 221 may include a high-speed random access memory, and may further include a nonvolatile memory, such as at least one magnetic disk storage device, a flash memory device, a universal flash memory (UFS), and the like.
Electronic device 200 may implement audio functions via audio module 270, speaker 270A, receiver 270B, microphone 270C, headset interface 270D, and an application processor, among other things. Such as music playing, recording, etc.
Audio module 270 is used to convert digital audio signals to analog audio signals for output and also to convert analog audio inputs to digital audio signals. Audio module 270 may also be used to encode and decode audio signals. In some embodiments, the audio module 270 may be disposed in the processor 210, or some functional modules of the audio module 270 may be disposed in the processor 210.
The speaker 270A, also called a "horn", is used to convert an audio electrical signal into an acoustic signal. The electronic apparatus 200 can listen to music through the speaker 270A or listen to a handsfree call.
The receiver 270B, also called "earpiece", is used to convert the electrical audio signal into an acoustic signal. When the electronic apparatus 200 receives a phone call or voice information, it is possible to receive a voice by placing the receiver 270B close to the ear of the person.
The microphone 270C, also referred to as a "microphone," is used to convert acoustic signals into electrical signals. When making a call or sending voice information, the user can input a voice signal to the microphone 270C by speaking the user's mouth near the microphone 270C. The electronic device 200 may be provided with at least one microphone 270C. In other embodiments, the electronic device 200 may be provided with two microphones 270C to implement a noise reduction function in addition to listening to voice information. In other embodiments, the electronic device 200 may further include three, four or more microphones 270C to collect sound signals, reduce noise, identify sound sources, implement directional recording functions, and so on. For example, the microphone 270C may be used to collect voice information related to embodiments of the present application.
The headphone interface 270D is used to connect wired headphones. The headset interface 270D may be the USB interface 230, or may be an open mobile electronic device platform (OMTP) standard interface of 3.5mm, or a Cellular Telecommunications Industry Association (CTIA) standard interface.
The pressure sensor 280A is used to sense a pressure signal, which can be converted into an electrical signal. In some embodiments, the pressure sensor 280A may be disposed on the display screen 294. The pressure sensor 280A can be of a wide variety of types, such as a resistive pressure sensor, an inductive pressure sensor, a capacitive pressure sensor, and the like. The capacitive pressure sensor may be a sensor comprising at least two parallel plates having an electrically conductive material. When a force acts on the pressure sensor 280A, the capacitance between the electrodes changes. The electronic device 200 determines the intensity of the pressure from the change in capacitance. When a touch operation is applied to the display screen 294, the electronic apparatus 200 detects the intensity of the touch operation based on the pressure sensor 280A. The electronic apparatus 200 may also calculate the position of the touch based on the detection signal of the pressure sensor 280A.
In some embodiments, the touch operations that are applied to the same touch position but different touch operation intensities may correspond to different operation instructions. For example: and when the touch operation with the touch operation intensity smaller than the first pressure threshold value acts on the short message application icon, executing an instruction for viewing the short message. And when the touch operation with the touch operation intensity larger than or equal to the first pressure threshold value acts on the short message application icon, executing an instruction of newly building the short message.
The gyro sensor 280B may be used to determine the motion pose of the electronic device 200. In some embodiments, the angular velocity of the electronic device 200 about three axes (i.e., x, y, and z axes) may be determined by the gyroscope sensor 280B. The gyro sensor 280B may be used for photographing anti-shake. For example, when the shutter is pressed, the gyro sensor 280B detects a shake angle of the electronic device 200, calculates a distance to be compensated for by the lens module according to the shake angle, and allows the lens to counteract the shake of the electronic device 200 through a reverse movement, thereby achieving anti-shake. The gyro sensor 280B may also be used for navigation, somatosensory gaming scenes.
The air pressure sensor 280C is used to measure air pressure. In some embodiments, the electronic device 200 calculates altitude, aiding in positioning and navigation from barometric pressure values measured by the barometric pressure sensor 280C.
The magnetic sensor 280D includes a hall sensor. The electronic device 200 may detect the opening and closing of the flip holster using the magnetic sensor 280D. In some embodiments, when the electronic device 200 is a flip, the electronic device 200 may detect the opening and closing of the flip according to the magnetic sensor 280D. And then according to the opening and closing state of the leather sheath or the opening and closing state of the flip cover, the automatic unlocking of the flip cover is set.
The acceleration sensor 280E may detect the magnitude of acceleration of the electronic device 200 in various directions (typically three axes). The magnitude and direction of gravity can be detected when the electronic device 200 is stationary. The method can also be used for recognizing the posture of the electronic equipment, and is applied to horizontal and vertical screen switching, pedometers and other applications.
A distance sensor 280F for measuring distance. The electronic device 200 may measure the distance by infrared or laser. In some embodiments, taking a picture of a scene, the electronic device 200 may utilize the distance sensor 280F to range for fast focus.
The proximity light sensor 280G may include, for example, a Light Emitting Diode (LED) and a light detector, such as a photodiode. The light emitting diode may be an infrared light emitting diode. The electronic apparatus 200 emits infrared light to the outside through the light emitting diode. The electronic device 200 detects infrared reflected light from a nearby object using a photodiode. When sufficient reflected light is detected, it can be determined that there is an object near the electronic apparatus 200. When insufficient reflected light is detected, the electronic device 200 may determine that there are no objects near the electronic device 200. The electronic device 200 can utilize the proximity light sensor 280G to detect that the user holds the electronic device 200 close to the ear for talking, so as to automatically turn off the screen to save power. The proximity light sensor 280G may also be used in a holster mode, a pocket mode automatically unlocks and locks the screen.
The ambient light sensor 280L is used to sense the ambient light level. The electronic device 200 may adaptively adjust the brightness of the display screen 294 according to the perceived ambient light level. The ambient light sensor 280L may also be used to automatically adjust the white balance when taking a picture. The ambient light sensor 280L may also cooperate with the proximity light sensor 280G to detect whether the electronic device 200 is in a pocket to prevent inadvertent contact.
The fingerprint sensor 280H is used to collect a fingerprint. The electronic device 200 can utilize the collected fingerprint characteristics to unlock the fingerprint, access the application lock, photograph the fingerprint, answer an incoming call with the fingerprint, and the like.
The temperature sensor 280J is used to detect temperature. In some embodiments, the electronic device 200 implements a temperature processing strategy using the temperature detected by the temperature sensor 280J. For example, when the temperature reported by the temperature sensor 280J exceeds the threshold, the electronic device 200 performs a reduction in performance of a processor located near the temperature sensor 280J, so as to reduce power consumption and implement thermal protection. In other embodiments, the electronic device 200 heats the battery 242 when the temperature is below another threshold, so as to avoid abnormal shutdown of the electronic device 200 due to low temperature. In other embodiments, when the temperature is below a further threshold, the electronic device 200 performs a boost on the output voltage of the battery 242 to avoid an abnormal shutdown due to low temperature.
The touch sensor 280K is also referred to as a "touch panel". The touch sensor 280K may be disposed on the display screen 294, and the touch sensor 280K and the display screen 294 form a touch screen, which is also called a "touch screen". The touch sensor 280K is used to detect a touch operation applied thereto or thereabout. The touch sensor can communicate the detected touch operation to the application processor to determine the touch event type. Visual output related to touch operations may be provided through the display screen 294. In other embodiments, the touch sensor 280K can be disposed on a surface of the electronic device 200 at a different location than the display screen 294.
The bone conduction sensor 280M may acquire a vibration signal. In some embodiments, bone conduction sensor 280M may obtain a vibration signal of a vibrating bone mass of a human vocal part. The bone conduction sensor 280M may also contact the pulse of the human body to receive the blood pressure pulsation signal.
In some embodiments, bone conduction sensor 280M may also be disposed in a headset, integrated into a bone conduction headset. The audio module 270 may analyze a voice signal based on the vibration signal of the bone mass vibrated by the sound part acquired by the bone conduction sensor 280M, so as to implement a voice function. The application processor can analyze heart rate information based on the blood pressure beating signal acquired by the bone conduction sensor 280M, so as to realize the heart rate detection function.
The keys 290 include a power-on key, a volume key, etc. The keys 290 may be mechanical keys. Or may be touch keys. The electronic apparatus 200 may receive a key input, and generate a key signal input related to user setting and function control of the electronic apparatus 200.
The motor 291 may generate a vibration cue. The motor 291 can be used for both incoming call vibration indication and touch vibration feedback. For example, touch operations applied to different applications (e.g., photographing, audio playing, etc.) may correspond to different vibration feedback effects. The motor 291 may also respond to different vibration feedback effects for touch operations on different areas of the display 294. Different application scenes (such as time reminding, receiving information, alarm clock, game and the like) can also correspond to different vibration feedback effects. The touch vibration feedback effect may also support customization.
Indicator 292 may be an indicator light that may be used to indicate a state of charge, a change in charge, or may be used to indicate a message, missed call, notification, etc.
The SIM card interface 295 is used to connect a SIM card. The SIM card can be attached to and detached from the electronic apparatus 200 by being inserted into the SIM card interface 295 or being pulled out from the SIM card interface 295. The electronic device 200 may support 1 or N SIM card interfaces, N being a positive integer greater than 1. The SIM card interface 295 may support a Nano SIM card, a Micro SIM card, a SIM card, etc. Multiple cards can be inserted into the same SIM card interface 295 at the same time. The types of the plurality of cards may be the same or different. The SIM card interface 295 may also be compatible with different types of SIM cards. The SIM card interface 295 may also be compatible with external memory cards. The electronic device 200 interacts with the network through the SIM card to implement functions such as communication and data communication. In some embodiments, the electronic device 200 employs esims, namely: an embedded SIM card. The eSIM card can be embedded in the electronic device 200 and cannot be separated from the electronic device 200.
The embodiment of the present application does not particularly limit the specific structure of the main execution body of the voice wakeup method, as long as the program recorded with the code of the voice wakeup method provided by the embodiment of the present application can be run to perform communication according to the voice wakeup method provided by the embodiment of the present application. For example, an execution main body of the voice wake-up method provided in the embodiment of the present application may be a functional module capable of calling a program and executing the program in the electronic device, or a communication device, such as a chip, applied in the electronic device.
Referring to fig. 3, fig. 3 is a schematic flowchart of a voice wake-up method according to an embodiment of the present application, where as shown in the figure, the method includes:
step 301, in response to the monitored voice information, the first device matches the voice information with a wake-up voiceprint of the first device.
The embodiment of the application is applied to an application scenario of multiple devices, where multiple electronic devices exist in the application scenario of the multiple devices, and any one of the electronic devices is used as an execution subject in the embodiment of the application for description. In addition, for convenience of description, the electronic apparatus described as the execution subject is referred to as a first apparatus.
For example, in the application scenario shown in fig. 1, if the wireless router does not have a wake-up voiceprint, and the wireless router does not have a competitive value for voice information, that is, the wireless router does not play a role of a wake-up device but a role of a wireless network providing communication between other electronic devices, the wireless router providing the wireless network is not a first device, however, the wireless router providing the wireless network may play a role of a second device in a network where the first device is located. Reference will be made in detail to the description of the embodiment shown in fig. 13.
The mobile phone 1, the mobile phone 2, the smart screen and the smart speaker in the application scenario shown in fig. 1 may be referred to as a first device. The smart sound box in the mobile phone 1, the mobile phone 2, the smart screen and the smart sound box serves as a first device for example, after a user sends voice information, the smart sound box monitors the voice information sent by the user, and the voice information can be matched with a wake-up voiceprint of the smart sound box to determine whether the user sending the current voice information has the capability of waking up the smart sound box.
It should be noted that, the above example only describes the smart speaker as the first device, and in practical applications, the mobile phone 1, the mobile phone 2 and the smart screen are also used as the first device to perform the operation described in step 301.
In addition, as described in the application scenario shown in fig. 1, matching the voice information with the wakeup voiceprint of the first device by the first device does not mean that the voice information is semantically matched with the wakeup voiceprint of the first device, but is matched in tone.
For example, the user a may previously input the voice information "i love my home" in the smart speaker as the wake-up voiceprint of the smart speaker, and the user B may previously input the voice information "marching forward courageously" in the smart speaker as the wake-up voiceprint of the smart speaker. After the user A says ' small E and small E ' in the space where the intelligent sound box is located and puts a song for me ', although the ' small E and the small E ' spoken by the user A put a song for me ' and the awakening voiceprint of the intelligent sound box ' i love my family ' and ' courage forward ' are not matched in semantics, the ' small E and the small E ' spoken by the user A put a song for me ' and the awakening voiceprint of the intelligent sound box ' i love my family ' are matched in timbre. In this case, the voice message is matched with the awakening voiceprint of the intelligent loudspeaker box. However, one electronic device may have multiple wake-up voiceprints. In order to improve the matching speed, it may be specified that when the voice information of the user is recorded as the wake-up voiceprint of the electronic device, it is specified that the voice information spoken by the user needs to be the specific voice information, and for example, when the user sets the wake-up voiceprint of the electronic device, it is necessary to speak the specific voice information "small E, small E".
Step 302, if the voice information is matched with the awakening voiceprint of the first device, the first device sends a bid value of the first device for the voice information to the second device, and monitors information sent by the second device, wherein the bid value of the first device for the voice information is the support capability of the first device for the awakening operation corresponding to the voice information.
In the embodiment of the present application, the second device is relative to the first device, for example, in the application scenario shown in fig. 1, the smart speaker may be used as the first device, and one or more of the wireless router, the mobile phone 1, the mobile phone 2, and the smart screen are the second device of the smart speaker. When the wisdom screen is as first equipment, one or more in wireless router, cell-phone 1, cell-phone 2 and the smart sound box then be the second equipment of wisdom screen.
Since a user may speak various voice messages (e.g., play a movie, watch a movie) when the user wants the electronic device to perform an operation (e.g., play a movie), in order to set the bid value of the electronic device for each voice message, the voice messages may be classified into a plurality of wake-up scenes, and the bid value of the electronic device for a wake-up scene is used as the bid value of the electronic device for the voice message belonging to the wake-up scene. The competitive value of each electronic device in different awakening scenes can be set when the electronic device leaves a factory; certainly, the user may also reset the bid value for the electronic device in each wake-up scene after the electronic device leaves the factory; and calculating the competitive value of each electronic device in different awakening scenes by the electronic device based on the respective device parameters and the dependence of different awakening scenes on the device parameters.
Referring to table 1, table 1 shows the competitive value of the electronic devices in different wake-up scenarios provided in the embodiment of the present application. As shown in table 1, the wake-up scenario includes: wake only, communication, music, video, storytelling, navigation, talk, etc. If the voice message received by the first device is matched with the wake-up voiceprint of the first device in terms of tone, the first device may determine, based on the semantic meaning of the voice message, a wake-up scene represented by the voice message.
For example, if the voice message is "small E, play song", the wake-up scene represented by the voice message is music; the voice information is 'small E, and calls mom', and the awakening scene represented by the voice information is a call. Of course, the correspondence between the wake-up scene, the voice information, and the wake-up scene is only used as an example, and in practical application, other wake-up scenes and other correspondence may also be set.
The larger the bid value in table 1 is, the larger the execution capability of the electronic device for the wake-up operation corresponding to the voice information in the wake-up scene is. The electronic device having a larger bid amount is easier to perform a wake-up operation corresponding to voice information as a final wake-up device. The value of 0 in table 1 indicates that the electronic device has no execution capability for the wake-up operation corresponding to the voice information in the wake-up scene. For example, if the voice message is "small E, playing movie", the wake-up scene represented by the voice message is a video, and the voice message corresponds to the wake-up operation: and opening the video playing software or opening the video playing software and playing the first movie in the list. Obviously, the smart sound box does not have the capability of playing a movie, so that the smart sound box has no response capability to the video of the awakening scene, and the competitive value of the smart sound box on the voice information corresponding to the video of the awakening scene is 0.
TABLE 1 Bid value of each electronic device under multiple wake-up scenarios
Wake-up scenario Mobile phone 1 Mobile phone 2 Intelligent screen Intelligent sound box
Wake-up only 3 3 2 2
Even communication 3 3 2 1
Music 2 2 1 3
Video 2 2 3 0
Story telling 2 2 1 3
Navigation 3 3 2 1
Communication system 3 3 2 1
…… …… …… …… ……
Note that, the higher the bid value in table 1 is, the stronger the execution capability of the electronic device for the wake-up operation corresponding to the voice information is. In practical applications, the lower the competitive value is, the higher the execution capability of the electronic device for the wake-up operation corresponding to the voice information is. The embodiment of the present application will be described below by taking as an example that the higher the bid value is, the stronger the execution capability of the electronic device for the wake-up operation corresponding to the voice information is. In addition, the specific values of the bid values in table 1 are only used as examples, and in practical applications, other values or other levels indicating the strength of the capability may also be used, which is not limited in the embodiment of the present application.
After the first device determines that the monitored voice information is matched with the awakening voiceprint of the first device, the first device also needs to monitor information sent by the second device. When the second device is an electronic device located in the same network as the first device, the information sent by the second device includes a bid value of the second device for the voice information. When the second device is a preset decision device, the decision device is used for receiving the competitive value of each electronic device to the voice information, which is sent by each electronic device in the network where the decision device is located, sending a first message to the electronic device with the highest competitive value, and sending a second message to the electronic device with the non-highest competitive value; the information sent by the second device includes the first message and the second message.
Step 303, in response to the monitored information sent by the second device, the first device determines whether the first device is a wake-up device based on the information sent by the second device.
In this embodiment of the application, as shown in step 302, the second device may be an electronic device other than the first device in a network where the first device is located, or may be a preset decision device. Similarly, when the second device has two roles, the information sent by the second device may not be available. When the second device has two roles, the specific step of the first device determining whether the first device is the wake-up device based on the information sent by the second device may also be different, and the specific step of the first device determining whether the first device is the wake-up device based on the information sent by the second device may refer to the description of the subsequent embodiments.
The wake-up device may be determined by the respective electronic devices for the bid value of the voice information, for example, the electronic device with the highest bid value in the current application scenario may be determined as the wake-up device.
Step 304, if the first device is a wake-up device, the first device executes a wake-up operation corresponding to the voice message.
In this embodiment of the application, the wake-up operation corresponding to the voice information may be preset, for example, when the voice information is a wake-up name "small E, small E" of the electronic device, the corresponding wake-up scene is pure wake-up, and the corresponding wake-up operation is to wake up the electronic device; when the voice information is 'small E, call', the corresponding awakening scene is a call; the corresponding awakening operation is entering a display interface of the electronic equipment for dialing the phone.
In practical application, the relationship among the wake-up scene, the voice message, and the wake-up operation can be expressed as follows:
and (3) awakening scene: conversation, voice information: small E, make a call, wake up operation: entering a display interface for dialing a call;
and (3) awakening scene: conversation, voice information: small E, call, wake-up operation: entering a display interface for dialing a call;
and (3) awakening scene: conversation, voice information: small E, call mother, wake up operation: dialing the telephone number indicated by 'mom' in the address book.
It can be understood from the above example that the multiple pieces of voice information may correspond to one wake-up scenario, two or more pieces of voice information in the multiple pieces of voice information may respectively correspond to different wake-up operations, and two or more pieces of voice information in the multiple pieces of voice information may correspond to the same wake-up operation.
In the embodiment of the application, the bidding value of the electronic equipment for the voice information is set for each electronic equipment, the awakening equipment is determined through the bidding value interaction process between the electronic equipment (between the first equipment and the second equipment), and the awakening operation corresponding to the voice information is executed through the determined awakening equipment, so that the problems that in the application scene of a plurality of electronic equipment, the electronic equipment can be awakened by the voice information of a user, and the electronic equipment can respond to the voice information of the user are solved.
As another embodiment of the present application, in the voice wake-up method shown in fig. 3, step 302 includes:
the first device acquires an awakening mode corresponding to the voice information, wherein the awakening mode comprises: waking up a scene;
the method comprises the steps that first equipment obtains a bidding value of the first equipment in an awakening mode, and the bidding value is used as the bidding value of the first equipment for voice information;
the first device sends the bid value of the first device for the voice information to the second device.
Correspondingly, the process of the first device executing the wakeup operation corresponding to the voice message includes:
the first equipment determines awakening operation corresponding to the voice information according to the semantic meaning of the voice information; the first device performs a wake-up operation corresponding to the voice information.
In the embodiment of the present application, as described in table 1, the bid value of the first device for the voice information is a bid value of the first device for an awake scene represented by the voice information. Therefore, the first device needs to first determine the wake-up scene corresponding to the voice information according to the semantic of the voice information. The first equipment determines the competitive value of the first equipment on the voice information by searching the competitive value of the first equipment in the awakening scene corresponding to the voice information. Of course, in this embodiment, the wake-up scenario is taken as an example, and in practical application, the wake-up mode may further include: a wake-up location. The specific description of the wake-up location can refer to the corresponding description of the schematic diagram shown in the following fig. 5.
As another embodiment of the present application, the obtaining, by the first device, a bid value of the first device in the wake mode includes:
in the first mode, the first device obtains a preset competitive value of the first device in the wake-up mode.
Or the like, or, alternatively,
in the second mode, the first device obtains the device parameters of the first device and the dependency of the wake-up mode on the device parameters.
The first device generates a bid value of the first device in the wake-up mode based on the device parameters of the first device and the dependence of the wake-up mode on the device parameters.
In the embodiment of the present application, a first way is to set, in advance, a bid value of each electronic device in different wake modes (e.g., wake scenarios) for each electronic device. The process of setting the bid value can be completed by a developer in the development stage of the electronic equipment; or at any stage before the electronic equipment leaves the factory; certainly, it is also possible that the user obtains the bid values in different wake-up modes from a preset server in the use process of the electronic device after leaving the factory, so as to complete the setting process; in the process of using the electronic device, it is also possible that the user sets the bid values in different wake-up modes for different electronic devices in the bid value setting interface of the electronic device to complete the setting process.
As an example of the first manner, the process of the first device setting the bid values of the electronic devices in different wake-up scenarios includes:
responding to a setting instruction of a user, displaying a first interface comprising at least one awakening scene by the first equipment to instruct the user to select one awakening scene in the first interface; responding to the awakening scene selected by the user, the first device displays a second interface comprising at least one electronic device and a bid value frame of the electronic device in the selected awakening scene so as to instruct the user to set the bid value of the electronic device in the selected awakening scene for the electronic device in the second interface, wherein the electronic device displayed in the second interface is the electronic device in the network where the first device is located.
As an example of this embodiment, referring to fig. 4, fig. 4 is a schematic diagram for setting a bid value of an electronic device in different wake-up scenarios according to an embodiment of the present application. Taking the example of setting on the mobile phone 1 in the application scenario shown in fig. 1, a user may input a setting instruction of a bid value on the mobile phone 1, for example, may click a certain virtual button in an interface displayed by the mobile phone 1 to implement; after receiving the setting instruction, the mobile phone 1 may display a first interface as shown in (a) in fig. 4, where a plurality of different wake-up scenes are displayed in the first interface, and a user may select any wake-up scene in the first interface as a wake-up scene to be set. For example, a user may select a video scene, after the mobile phone 1 receives a message or an instruction of the user selecting the video scene, the mobile phone 1 displays a second interface as shown in (b) of fig. 4, a plurality of electronic devices (e.g., the mobile phone 1, the mobile phone 2, the smart speaker, and the smart screen) and bid value frames of the plurality of electronic devices are displayed in the second interface, and the user may set a bid value of each electronic device in the video scene by adjusting an adjustment axis corresponding to each electronic device. For example, the user clicks a circular button in the adjustment axis with a finger and then slides left and right to control the circular button in the adjustment axis to move left and right in the adjustment axis, so that the number in the bid value frame changes; and when the number in the value frame to be bid is the target number which the user wants to set, stopping sliding and lifting the finger of the user. A plurality of electronic devices are displayed in the second interface of the mobile phone 1, and when the user completes setting of the bid value of one of the electronic devices in the video scene, the user can continue to set the bid values of the other electronic devices in the video scene for the other electronic devices.
It should be noted that, when the bid value of the electronic device is set through the second interface, the plurality of electronic devices (set a) and the plurality of electronic devices (set B) formed by the first device and the second device in the voice wake-up method provided by the embodiment of the present application may be completely the same, and there may be some portions that are the same and some portions that are different. For example, when a user sets a bid value of a plurality of electronic devices in different wake-up scenes through a first device, a mobile phone 1 and a smart screen 1 may exist in a network where the first device is located, that is, an electronic device set a when the bid value is set through a second interface includes the mobile phone 1 and the smart screen 1. After the setting is completed, in the user experiment stage, the smart screen 1 still exists in the network where the mobile phone 1 is located. At this time, when the user sends out voice information, the mobile phone 1 is the first device, the smart screen 1 is the second device, and when the smart screen is the first device, the mobile phone 1 is the second device, that is, the plurality of electronic device sets B formed by the first device and the second device participating in the voice wakeup method provided by the embodiment of the present application are the mobile phone 1 and the smart screen. In this case, the electronic devices in set a are identical to the electronic devices in set B. Perhaps, after a period of time (e.g., one month), the smart screen 1 is unavailable due to a failure, the smart screen 2 is newly purchased in a home scene, the smart sound box is newly purchased, and meanwhile, a friend carries the mobile phone 2 to make a visitor at home. After the user sends out voice information, the mobile phone 1 serves as a first device, and the mobile phone 2, the smart screen 2 and the smart sound box serve as a second device of the mobile phone 1. The electronic devices in set B then include: cell-phone 1, wisdom screen 2, intelligent audio amplifier, cell-phone 2. In this case, the electronic devices in set a are partially the same as the electronic devices in set B and are partially different. Certainly, in the voice wake-up bidding stage, the bidding value of the mobile phone 1 for the voice information is the bidding value set by the user in the setting stage, and the bidding values of the mobile phone 2, the smart screen 2 and the smart speaker for the voice information may be default bidding values of leaving the factory.
As an example of the second way, the dependency of each wake-up mode on device parameters of the electronic device may be set. When the wake-up mode is a wake-up scenario, table 2 shows the dependency of different wake-up scenarios on device parameters of the electronic device. The dependence of the wake-up scenario on device parameters is represented in table 2 by extremely high, medium, and none, and the device parameters in table 2 include: screen size, resolution, audio decoding capability, photographing capability, chip specification, speaker, storage space, user usage probability, and the like.
Of course, in practical applications, the device parameters may also include other device parameters not listed in table 2, or include some of the device parameters listed in table 2. The dependency of different wake-up scenarios on the device parameter can also be represented by other words representing different degrees of dependency, and can also be represented by other values representing different degrees of dependency. Table 2 is only used to show an example, and does not set any limit to the embodiments of the present application.
TABLE 2 dependency of different wake-up scenarios on device parameters of an electronic device
Figure BDA0002625580620000171
As an example, the process of the first device generating the bid value of the first device in the wake mode based on the device parameter of the first device and the dependency of the wake mode on the device parameter may be as follows:
the device parameter of the first device may also be set to several levels, for example, the screen size in a first preset range is set as a first level, and the score is set as 100; the screen size in a second preset range is a second grade, and the score is set to be 80; the screen size in a third preset range is a third grade, and the score is set to be 60; the screen size in the fourth preset range is in the fourth level, and the score is set to 0. The values represented by none, middle, high, and extremely high in the dependency are represented by a weight, for example, extremely high represents a weight of 1, high represents a weight of 0.5, middle represents a weight of 0.3, and none represents a weight of 0. And obtaining the competitive value of the first equipment in the awakening scene through the score corresponding to the equipment parameter of the first equipment and the corresponding weight. It should be noted that this example is only used for example, and the specific calculation process may also be other calculation manners, which is not limited in this application embodiment.
As another embodiment of the present application, the wake-up mode further includes: a wake-up location; the method for acquiring the awakening mode corresponding to the voice information by the first equipment comprises the following steps:
the first equipment determines a wake-up place corresponding to the voice information based on a network to which the first equipment is currently connected.
In the embodiment of the present application, as described above, the wake mode may include a wake place in addition to the wake scene. The wake-up location may include: in cars, companies, and homes, etc. The first device may determine a current wake-up location based on a network to which the first device is currently connected. For example, when the first device is connected to Wi-fi (wireless fidelity) in a home, the wake-up location is the home, and when the first device is connected to Wi-fi (wireless fidelity) in a vehicle, the wake-up location is in the vehicle.
In the case that the wake-up mode further includes a wake-up location, the setting of the bid value of the electronic device in different wake-up modes in the first manner can be shown in fig. 5, where fig. 5 adds a display interface, such as the interface shown in (b) of fig. 5, to the setting process shown in fig. 4, and the interface is denoted as a third interface. The first interface shown in (a) of fig. 5 may be the same as the first interface shown in (a) of fig. 4, after the user selects any one of the wake-up scenes through the first interface shown in (a) of fig. 5, the user enters the third interface shown in (b) of fig. 5, the third interface displays one or more wake-up places, after the user selects one of the wake-up places in the third interface, the user enters the second interface shown in (c) of fig. 5, and the second interface shown in (c) of fig. 5 may be the same as the second interface shown in fig. 4. The description of the first interface shown in (a) in fig. 5 and the second interface shown in (c) in fig. 5 may refer to the description about (a) in fig. 4 and (b) in fig. 4 in the application scenario shown in fig. 4, and is not repeated here.
Of course, in practical applications, the third interface shown in (b) in fig. 5 may also be displayed first, and the user selects one of the wake-up locations first, then enters the first interface shown in (a) in fig. 5, and then enters the second interface shown in (c) in fig. 5 after selecting one of the wake-up scenarios.
In the case that the wake-up pattern further includes wake-up locations, the dependency of different wake-up scenarios shown in table 2 on the device parameters of the electronic device may be represented as one of the wake-up locations, and then, for each wake-up location, the dependency of different wake-up scenarios in the wake-up location on the device parameters of the electronic device is set as shown in table 2. Or, a plurality of tables are set, each table represents the dependency degree of different wake-up places on the device parameters in one wake-up scene. And finally, determining the dependence of the specific awakening scene and the specific awakening place on the equipment parameters by searching, and calculating to obtain the competitive value of the first equipment in the awakening mode based on the dependence obtained by searching and the equipment parameters of the first equipment.
In order to clearly understand various implementation manners of the voice wake-up method provided in the embodiment of the present application, in the following embodiments, a process of determining the wake-up device based on the bid value of each electronic device in different application scenarios will be described in detail.
Referring to fig. 6, fig. 6 is a schematic flowchart of another voice wake-up method provided in the embodiment of the present application, and as shown in the drawing, the method includes:
step 601, responding to the monitored voice information, and matching the voice information with the awakening voiceprint of the first device by the first device.
Step 602, if the voice information is matched with the awakening voiceprint of the first device, the first device sends a bid value of the first device for the voice information to the second device, and monitors a bid value of the second device for the voice information sent by the second device, wherein the bid value of the first device for the voice information is a support capability of the first device for the awakening operation corresponding to the voice information.
The contents of step 601 to step 602 are the same as the contents of step 301 to step 302, and the description of step 301 to step 302 may be referred to specifically, which is not repeated herein.
Step 6031, if at least one of the bid values of the second device for the voice information, received by the first device within the first preset time, of the bid values of the second device for the voice information is higher than the bid value of the first device for the voice information, it is determined that the first device is not the wake-up device.
Step 6032, if the bid value of the second device for the voice information received by the first device within the first preset time is lower than the bid value of the first device for the voice information, determining that the first device is the wake-up device.
Step 6033, if the bid value of the second device for the voice information received by the first device within the first preset time is not higher than the bid value of the first device for the voice information and is equal to the bid value of the first device for the voice information, marking the second device with the bid value equal to the bid value of the first device for the voice information as a third device, and acquiring the state information of the third device.
Step 604, the first device determines whether the first device is a wake-up device based on the state information of the first device and the state information of the third device, where the state information includes at least one of the following state parameters: the total time length of the starting-up state, the position of the equipment, the residual electric quantity and whether the equipment is in the use state within the second preset time.
Step 605, if the first device is a wake-up device, the first device executes a wake-up operation corresponding to the voice message.
In this embodiment of the application, if the preset bid values of the electronic device are not equal, the bid value of the second device received by the first device and sent by the second device may be as follows:
in case 1-1, the bid values sent by the second device are all higher than the bid value of the first device.
In case 1-2, the bid values sent by the second device are all lower than the bid value of the first device.
In case 1-3, the second device sends a portion of the bid value higher than the first device and another portion lower than the first device.
Since the finally determined wake-up device is the electronic device with the highest bid value, the first device is not the wake-up device for the case 1-1, the first device is the wake-up device for the case 1-2, and the first device is not the wake-up device for the case 1-3.
In order to more clearly understand the above process, referring to fig. 7, fig. 7 is a flowchart of a voice wakeup method provided in this embodiment of the present application, as shown in the figure, the DSP module of the first device monitors sound in the environment, and when the voice information is monitored, matches the monitored voice information with a wakeup voiceprint of the first device, and if the voice information does not match the wakeup voiceprint, the voice information will not be responded to. If the voice information is matched with the awakening voiceprint, acquiring a bid value of the voice information, sending the bid value of the Socket client (Socket client) of the first device to a Socket Server (Socket Server) of the second device, monitoring the bid value of the second device sent by the second device by the Socket Server of the first device, if the second device with the bid value higher than the bid value of the second device exists within a specified time, quitting the Socket Server of the first device, wherein the accepted bid value of the second device sent by the second device is not monitored, and if the current bid value of the second device is lower than the bid value of the second device, the Socket Server of the first device keeps monitoring; and if the monitored bid values are all lower than the bid value of the device within the specified time, determining that the device is a wake-up device after the specified time, and responding. The predetermined time may be a preset time, such as 0.3s, 0.5s, 0.8s, etc., which is not limited in the present application. Among them, the Socket client and the Socket Server are located in a telephone management (Telephony manager) of a framework structure shown in fig. 8. In addition, it should be noted that the embodiment shown in fig. 8 is only used for example, and in practical applications, the application layer of the first device may further include other application software, and the application framework layer may further include other services or management.
Of course, if there is a possibility that the bid values of the electronic devices are equal, there may be the following cases:
in case 2-1, the bid values sent by the second device are all higher than the bid value of the first device.
Case 2-2, the bid values sent by the second device are all lower than the bid value of the first device.
In case 2-3, the second device sends a portion of the bid values that are higher than the first device, a portion that are lower than the first device, and another portion that are equal to the first device.
In case 2-4, some of the bid values sent by the second device are higher than the bid value of the first device, and some of the bid values are equal to the bid value of the first device.
In case 2-5, some of the bid values sent by the second device are lower than the bid value of the first device and some are equal to the bid value of the first device.
For the case 2-1, the first device is not a wake-up device, for the case 2-2, the first device is a wake-up device, for the case 2-3, the first device is not a wake-up device, for the case 2-4, the first device is not a wake-up device, for the case 2-5, both the first device and the second device having a bid value equal to the bid value of the first device may be used as wake-up devices, or one of the first device and the second device may be selected as a wake-up device. If one of the first device and the second device having a bid value equal to the bid value of the first device is taken as the wake-up device, reference may be made to the description of the subsequent embodiments.
In summary, if at least one of the bid values of the second device for the voice information received by the first device within the first preset time is higher than the bid value of the first device for the voice information, it is determined that the first device is not a wake-up device, for example, the above case 2-1, case 2-3, and case 2-4.
If the bid value of the second device for the voice information received by the first device within the first preset time is lower than the bid value of the first device for the voice information, the first device is determined to be a wake-up device, for example, in case 2-2 above.
If the bid value of the second equipment for the voice information received by the first equipment within the first preset time is not higher than the bid value of the first equipment for the voice information and is equal to the bid value of the first equipment for the voice information, marking the second equipment with the bid value equal to the bid value of the first equipment for the voice information as third equipment, and acquiring state information of the third equipment; the first device determines whether the first device is a wake-up device or not based on state information of the first device and state information of the third device, wherein the state information includes at least one of the following state parameters: the total duration of the power-on state within the second predetermined time, the location of the device, the amount of remaining power, and whether it is in use, such as in cases 2-5 above.
In this example, when there are at least two electronic devices in an application scenario that have equal bid values for voice information, the electronic device with the longest total on-state duration within a second preset time in the electronic devices with equal bid values may be used as a wake-up device, or the electronic device with the largest remaining power may be used as a wake-up device, or the electronic device currently in a use state may be used as a wake-up device, or the electronic device with the closest device location to the user may be used as a wake-up device. Of course, the electronic device whose device location is closest to the user may be determined by the loudness of the heard voice information, and the electronic device whose loudness of the heard voice information is highest represents the electronic device closest to the user. In addition, the total time of the starting state within the second preset time is represented as follows: the total time that the electronic device is kept in the power-on state within a second preset time, for example, the second preset time may be set to one week, and the state parameter is used to indicate whether the user has used the electronic device frequently recently.
Of course, in practical applications, the status information of the third device may be sent to the first device along with the bid value sent by the third device, and the status information of the first device may be sent to the second device along with the bid value sent by the first device.
Referring to fig. 9, fig. 9 is an application scenario of the embodiment shown in fig. 6, where the application scenario includes a plurality of electronic devices located under the same network, and the application scenario includes: cell-phone 1, cell-phone 2, wisdom screen and smart sound box. The user A inputs the voiceprint information of the user A in the mobile phone 1, the intelligent screen and the intelligent sound box in advance to be used as the awakening voiceprint of the mobile phone 1, the intelligent screen and the intelligent sound box.
In the application scene, the awakening voiceprints of the mobile phone 1, the smart screen and the smart sound box are voiceprint information of the user A, so that the awakening voiceprints of the mobile phone 1, the smart screen and the smart sound box are matched with the voiceprint information of the user A, and therefore the operation of the step 602 is executed by the mobile phone 1, the smart screen and the smart sound box, and after the operation of the step 601 is executed by the mobile phone 2, the awakening voiceprint of the mobile phone 2 is not matched with the voice information of the user A, namely the small E and the small E, the subsequent steps are not executed continuously.
As an example, the wake-up scene of the voice message "small E, play a song for me" is music, the competitive value in the music scene preset for the mobile phone 1 is 5, the competitive value in the music scene preset for the smart screen is 3, and the competitive value in the music scene preset for the smart speaker is 8. The mobile phone 1 serves as first equipment and sends the bid value 5 of the mobile phone 1 to the smart screen, the mobile phone 2 and the smart sound box; the intelligent screen is used as first equipment, and a bid value 3 of the intelligent screen is sent to the mobile phone 1, the mobile phone 2 and the intelligent sound box; the intelligent sound box serves as first equipment and sends a bid value 8 of the intelligent sound box to the mobile phone 1, the intelligent screen and the mobile phone 2; the handset 2, as the first device, ends after performing step 601. The mobile phone 1 serves as a first device, and in the received bid values, a bid value 8 of the smart speaker is higher than a bid value 5 of the smart speaker, so that the mobile phone 1 executes a step 6031; the smart screen is used as a first device, and the received bidding value is higher than the bidding value 5 of the smart sound box, so that the smart screen executes a step 6031; the smart speaker as the first device receives bid values that are all lower than its own bid value of 8, and therefore the smart speaker performs step 6032.
If another electronic device a equal to the bid value of the smart sound box exists in the application scenario, in the process of sending the bid value, each electronic device serves as a first device, in the process of sending the respective bid value to a second device, respective state information can also be sent, taking the state information as the total startup duration within the second preset time as an example, if the total startup duration of the electronic device a received by the smart sound box serving as the first device is longer than the total startup duration of the smart sound box, the smart sound box is not a wake-up device, and if the total startup duration of the smart sound box received by the electronic device a serving as the first device is longer than the total startup duration of the electronic device a, the electronic device a is a wake-up device.
Referring to fig. 10, fig. 10 is a schematic flowchart of another voice wake-up method provided in the embodiment of the present application, and as shown in the drawing, the method includes:
step 1001, in response to the monitored voice information, the first device matches the voice information with a wake-up voiceprint of the first device.
Step 1002, if the voice information matches the awakening voiceprint of the first device, the first device sends the bidding value of the first device for the voice information to the second device, and monitors the bidding value of the second device for the voice information sent by the second device.
The contents of steps 1001 to 1002 are the same as those of steps 301 to 302, and the descriptions of steps 301 to 302 may be referred to specifically, and are not repeated herein.
Step 1003, if the bid value of the second device for the voice information received by the first device within the first preset time is lower than the bid value of the first device for the voice information, the first device determines whether the bid value of the first device for the voice information is higher than a preset value.
In the embodiment of the present application, there may be a case where the electronic device in the application scenario that monitors the voice information, matches the voice information, and has the highest bid value for the voice information does not have the capability of performing the wake-up operation corresponding to the voice information.
For example, see the application scenario shown in fig. 11, in which the wake-up voiceprints of the mobile phone 1, the smart screen, and the smart speaker include voiceprint information of the user a. The user says "little E, makes a call" in the sitting room, however, the wisdom screen and the intelligent audio amplifier that can monitor user's speech information in the sitting room, all can not carry out the awakening operation that speech information "makes a call" corresponds: and entering an interface for making a call. Therefore, no matter which one of the intelligent screen and the intelligent voice box has a high bid value, the one with the high bid value cannot perform the awakening operation corresponding to the voice message. Therefore, the first device with the highest bidding value also needs to determine whether the bidding value of the first device for the voice information is higher than a preset value, and the first device is the awakening device only if the bidding value of the first device for the voice information is higher than the preset value. The preset value is a critical value which is preset for each wake-up mode and can perform a corresponding wake-up operation.
Step 10041, if the bid value of the first device for the voice information is higher than the preset value, determining that the first device is a wake-up device, and the first device executes a wake-up operation corresponding to the voice information.
Step 10042, if the bid value of the first device for the voice information is lower than or equal to the preset value, the first device determines whether a fourth device exists in the network where the first device is located, where the fourth device is an electronic device in which the voiceprint information matches the voice information in the second device where the first device does not monitor the bid value within the first preset time.
Step 1005, if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice information to the fourth device to instruct the fourth device to perform a wakeup operation corresponding to the voice information when the bid value of the fourth device for the voice information is higher than the preset value.
In the embodiment of the present application, as shown in fig. 11, after the user sends the voice message in the room 2, the mobile phone 1 and the mobile phone 2 located in the room 1 may not hear the voice message of the user due to a long distance. In this case, the smart screen with the highest bid value in the room 2 can find out whether the fourth device exists in the network. And when the fourth equipment is the electronic equipment with the highest bidding value and serves as the first equipment, the electronic equipment in the second equipment which does not receive the bidding value in the second equipment. For example, the smart screen is an electronic device with the highest bid value, the smart screen is a first device, and the second devices are a smart speaker, a mobile phone 1 and a mobile phone 2 which are located in the same network as the smart screen. Because the intelligent screen receives the competitive value of the intelligent sound box, the fourth device is the electronic device in the mobile phone 1 and the mobile phone 2. That is, the fourth device is an electronic device that wakes up the voiceprint matching the voice information in the mobile phone 1 and the mobile phone 2. Through comparison, the awakening voiceprint of the mobile phone 1 is matched with the voice information, the mobile phone 1 is the fourth device, the smart screen with the highest bidding value sends an awakening instruction to the mobile phone 1, after receiving the awakening instruction, the mobile phone 1 judges whether the bidding value of the mobile phone 1 to the voice information is higher than a preset value, and executes awakening operation corresponding to the voice information under the condition that the bidding value of the mobile phone 1 is higher than the preset value. It should be noted that, while the electronic device with the highest bidding value sends the wakeup instruction to the fourth device, the electronic device may also send the voice information or the text information corresponding to the voice information, so that the fourth device determines whether the bidding value of the fourth device for the voice information is higher than the preset value.
It should be noted that a second device corresponding to the bid value monitored by the first device may be denoted as a second bidding device, and a second device other than the second bidding device in the second device may be denoted as a fourth candidate device. Therefore, the fourth device is an electronic device in which the voiceprint information is matched with the voice information in candidate fourth devices, where the candidate fourth devices are other second devices in the second devices except for the second bidding device, and the second bidding device is a second device corresponding to the bid value monitored by the first device.
Referring to fig. 12, fig. 12 is a schematic flowchart of another voice wake-up method according to an embodiment of the present application, where as shown in the figure, the method includes:
step 1201, in response to the monitored voice information, the first device matches the voice information with a wake-up voiceprint of the first device.
Step 1202, if the voice information is matched with the awakening voiceprint of the first device, the first device sends the bidding value of the first device for the voice information to the second device, and monitors the bidding value of the second device for the voice information sent by the second device.
Steps 1201 to 1202 are the same as steps 301 to 302, and the description of steps 301 to 302 may be specifically referred to, and will not be repeated herein.
Step 1203, if the first device does not receive the information sent by the second device within the first preset time, the first device determines whether the bid value of the first device for the voice information is higher than a preset value.
Referring to fig. 13, which is an application scenario of the embodiment shown in fig. 12, the application scenario shown in fig. 13 is different from the application scenario shown in fig. 11 in that, in the application scenario shown in fig. 13, only one electronic device capable of monitoring the voice information of the user and matching the voice information exists, and in the application scenario shown in fig. 11, at least two electronic devices capable of monitoring the voice information of the user and matching the voice information exist. When two or more electronic devices capable of monitoring the voice information of the user exist and matched with the voice information, a comparison process of competitive value between the electronic devices capable of monitoring the voice information of the user and matched with the voice information is required. And the user voice information can be monitored, and under the condition that only one electronic device matched with the voice information is provided, the bidding value sent by other electronic devices cannot be received, so that the bidding value comparison process does not exist, and whether the bidding value of the first device for the voice information is higher than the preset value or not is directly judged.
Step 12041, if the bid value of the first device for the voice message is higher than the preset value, the first device determines that the first device is a wake-up device.
Step 12042, if the bid value of the first device for the voice information is lower than or equal to the preset value, the first device determines whether a fourth device exists in a network where the first device is located, where the fourth device is an electronic device in which the voice print information matches the voice information in a second device in which the first device does not monitor the bid value within the first preset time.
Step 1205, if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice information to the fourth device, so as to instruct the fourth device to perform a wakeup operation corresponding to the voice information when the bid value of the fourth device for the voice information is higher than the preset value.
The contents of step 1241 to step 1205 are the same as the contents of step 10041 to step 1005, and the descriptions of step 10041 to step 1005 may be referred to specifically, and are not described herein again.
In the embodiments shown in fig. 10 and 12, in the process of determining whether the fourth device exists in the second device, since the fourth device is a device whose wake-up voiceprint matches the voice information, the first device needs to store the wake-up voiceprint of each electronic device in advance.
As another embodiment of the present application, before the first device determines whether a fourth device exists in a network in which the first device is located, the method further includes:
the first device acquires a wakeup voiceprint of a second device, wherein the first device does not monitor the bid value within a first preset time.
As another embodiment of the present application, the acquiring, by a first device, a wake-up voiceprint of a second device, for which the first device does not monitor a bid value within a first preset time includes:
when the first equipment is preset voiceprint summarizing equipment, the first equipment receives awakening voiceprints of other electronic equipment except the first equipment in a network where the first equipment is located;
the first device generates a voiceprint characteristic summary based on the awakening voiceprint of the first device and the received awakening voiceprints of other electronic devices, and sends the voiceprint characteristic summary to the other electronic devices;
the method comprises the steps that a first device obtains awakening voiceprints of a second device, wherein the first device does not monitor the competitive value within a first preset time, from voiceprint characteristic collection;
when the first device is not the preset voiceprint collecting device, the first device sends the awakening voiceprint of the first device to the preset voiceprint collecting device so as to instruct the voiceprint collecting device to generate voiceprint characteristic collection of each electronic device in the network where the voiceprint collecting device is located;
the first equipment receives voiceprint characteristic summary sent by the voiceprint summary equipment;
the first equipment acquires the awakening voiceprint of the second equipment, which does not monitor the competitive value within the first preset time, from the voiceprint characteristic collection.
In this embodiment of the application, an application scenario shown in fig. 14 may be referred to, so that each electronic device obtains a wake-up voiceprint of another electronic device, in the application scenario, a voiceprint summarizing device needs to be preset, and an electronic device that often stays in the application scenario may be set as the voiceprint summarizing device, for example, a wireless router.
In this application scenario, each electronic device located in the same network with the voiceprint collecting device can send respective awakening voiceprints to the voiceprint collecting device, and after the voiceprint collecting device collects the awakening voiceprints of each electronic device, the collected awakening voiceprints of each electronic device are sent to each electronic device in the network.
Referring to fig. 15, fig. 15 is a schematic flowchart of another voice wake-up method provided in the embodiment of the present application, and as shown in the drawing, the method includes:
step 1501, in response to the monitored voice information, the first device matches the voice information with a wake-up voiceprint of the first device.
Step 1502, if the voice information is matched with the awakening voiceprint of the first device, the first device sends a bid value of the first device for the voice information to the second device, and monitors information sent by the second device, wherein the bid value of the first device for the voice information is the support capability of the first device for the awakening operation corresponding to the voice information, and the second device is a preset decision device.
The content of step 1501 to step 1502 is the same as the content of step 301 to step 302, and specific reference may be made to the description of step 301 to step 302, which is not described herein again.
Step 15031, if the information sent by the second device and received by the first device is the first message, the first device determines that the first device is the wake-up device.
Step 15041, if the first device is a wake-up device, the first device performs a wake-up operation corresponding to the voice message.
Step 15032, if the information sent by the second device and received by the first device is the second message, the first device determines that the first device is not the wake-up device.
Step 15042, if the first device is not a wake-up device, the first device abandons the response to the voice message.
In the embodiment of the application, when the second device is a decision device, other electronic devices in the network where the second device is located all send respective bid values for the voice information to the decision device, and after the decision device receives the bid values for the voice information sent by the electronic devices in the network where the decision device is located, the decision device compares the received bidding values of the electronic devices for the voice information, which are sent by the electronic devices within the third preset time, selects the electronic device with the highest bidding value as the awakening device, and sends a first message to the electronic device with the highest bidding value, the electronic equipment with the highest bidding value is indicated as awakening equipment, the awakening operation corresponding to the voice information is executed, a second message is sent to the electronic equipment with the highest bidding value, to indicate that the electronic device with the highest bid value has dropped the wake-up and corresponding response.
The first device is any one of other electronic devices in a network where the decision device is located, and the first device receives a first message sent by the second device, and indicates that the first device is a wake-up device, and the first device executes wake-up operation corresponding to the voice information. The first device receives a second message sent by the second device, and the second message indicates that the first device is not a wake-up device, and the first device gives up the wake-up and the corresponding response.
Of course, in this embodiment, if the number of the electronic devices with the highest bid value is two or more, the third device may be determined as the wake-up device according to the state information of each electronic device in the embodiment shown in fig. 6.
Certainly, in practical application, a plurality of decision devices may be set, however, priorities need to be set for the plurality of decision devices in advance, and the decision device with the highest decision priority can be determined as the final decision device by the plurality of decision devices in the same network through mutual information interaction.
In order to have a clearer understanding of the voice wake-up method provided by the embodiment shown in fig. 15, it is further explained by the application scenario shown in fig. 16.
In the application scenario shown in fig. 16, the wireless router is a preset decision device, and the wireless router does not act as a wake-up device. The application scene also comprises a mobile phone 1, a mobile phone 2, an intelligent screen and an intelligent sound box. The mobile phone 1, the mobile phone 2, the intelligent screen, the intelligent sound box and the wireless router are in the same wireless network, or the mobile phone 1, the mobile phone 2, the intelligent screen and the intelligent sound box are all connected with the wireless router. Wherein, the awakening voiceprint of the mobile phone 1 comprises: voiceprint information of user a; the wake-up voiceprint of the handset 2 comprises: voiceprint information of user B; the woke up voiceprint of wisdom screen and intelligent audio amplifier all includes: user a's voiceprint information and user B's voiceprint information.
In the application scenario, the user a speaks "small E, small E" to play a movie for me ", the wake-up voiceprint of the mobile phone 2 does not include the voiceprint information of the user a, and at this time, the voice information" small E, small E "to play a movie for me" is spoken for the user a, so that the voice information is not matched with the wake-up voiceprint of the mobile phone 2, and the mobile phone 2 gives up the round of response.
Cell-phone 1, wisdom screen and smart sound box also all match voice message "little E, put a film for me" with respective vocal print of awakening up. The awakening voiceprints matched with the voice information of the user A exist in the awakening voiceprints in the mobile phone 1, the intelligent screen and the intelligent loudspeaker box, namely the voice information is matched with the awakening voiceprints of the mobile phone 1, the intelligent screen and the intelligent loudspeaker box. The mobile phone 1, the intelligent screen and the intelligent sound box respectively send respective bidding values for voice information 'small E, and put a movie' to the wireless router, and monitor information sent by the decision-making equipment.
After the decision device receives respective bidding values of the voice information 'small E, small E' and the respective films played for me 'sent by the mobile phone 1, the smart screen and the smart sound box, the bidding values of the smart screen for the voice information' small E, small E 'and the respective films played for me' are the highest through comparison, at the moment, the decision device sends a first message to the smart screen, and sends a second message to the mobile phone 1 and the smart sound box.
After the first message is received by the smart screen, the smart screen is determined to be a wake-up device, and then wake-up operation corresponding to voice information 'small E, and release a movie' is executed: the first movie in the movie playlist is played.
After receiving the second message, the mobile phone 1 and the smart speaker give up the round of response.
Through the description of the application scenario, it can be understood that the smart screen as the first device performs as follows: step 1501, step 1502, step 15031 and step 15041. What cell-phone 1 and smart sound box carried out as first equipment does: step 1501, step 1502, step 15032, and step 15042.
As can be understood from the embodiment shown in fig. 15 and the application scenario shown in fig. 16, the decision device compares the competitive value of the electronic devices other than the decision device in the network where the decision device is located on the voice information in the application scenario. In practical application, the competitive value of the decision device for comparison may also include the competitive value of the decision device for the voice information.
For example, after the decision device receives the bid values of the voice information sent by the electronic devices in the network where the decision device is located, except for the decision device, the decision device compares the bid values of the voice information sent by the electronic devices sent by the other electronic devices received within the second preset time with the bid values of the voice information sent by the decision device, and selects the electronic device with the highest bid value as the wake-up device. If the bidding value of the decision device for the voice information is the highest, the decision device is a wake-up device and executes wake-up operation corresponding to the voice information; and sending a second message to other electronic equipment except the decision-making equipment in the network where the decision-making equipment is located to instruct the other electronic equipment to abandon the response. And if the bidding value of the decision device to the voice information is not the highest, the decision device abandons the response, and sends a first message to the electronic device with the highest bidding value in the other electronic devices to indicate the electronic device with the highest bidding value in the other electronic devices as a wake-up device, executes a wake-up operation corresponding to the voice information, and sends a second message to the electronic device with the highest bidding value in the other electronic devices to indicate the electronic device with the highest bidding value in the other electronic devices to abandon the response.
It should be understood that, the sequence numbers of the steps in the foregoing embodiments do not imply an execution sequence, and the execution sequence of each process should be determined by its function and inherent logic, and should not constitute any limitation to the implementation process of the embodiments of the present application.
In the embodiment of the present application, the electronic device may be divided into the functional units according to the above method example, for example, each functional unit may be divided for each function, or two or more functions may be integrated into one processing unit. The integrated unit can be realized in a form of hardware, and can also be realized in a form of a software functional unit. It should be noted that the division of the units in the embodiments of the present application is illustrative, and is only one logical function division, and in actual implementation, there may be another division manner. The following description will take the example of dividing each functional unit corresponding to each function:
referring to fig. 17, the electronic apparatus 1700 is described as a first apparatus as an execution subject, and the first apparatus includes:
a voiceprint matching unit 1701 for matching the voice information with a wake-up voiceprint of the first device in response to the monitored voice information;
an information interaction unit 1702, configured to send, to a second device, a bid value of the first device for the voice information if the voice information matches the wake-up voiceprint of the first device, and monitor information sent by the second device, where the bid value of the first device for the voice information is a support capability of the first device for a wake-up operation corresponding to the voice information;
a bidding unit 1703, configured to determine, in response to the monitored information sent by the second device, whether the first device is a wake-up device based on the information sent by the second device;
a wake-up unit 1704, configured to execute a wake-up operation corresponding to the voice message if the first device is a wake-up device.
As another embodiment of the present application, the second device is an electronic device located in the same network as the first device, and the information sent by the second device includes a bid value of the second device for the voice information;
accordingly, bidding unit 1703 is also used to:
if at least one bidding value of the second equipment on the voice information is higher than the bidding value of the first equipment on the voice information in the bidding values of the second equipment on the voice information received by the first equipment within the first preset time, determining that the first equipment is not the awakening equipment;
and if the bidding values of the second equipment to the voice information received by the first equipment within the first preset time are all lower than the bidding values of the first equipment to the voice information, determining that the first equipment is awakening equipment.
If the bid value of the second equipment for the voice information received by the first equipment within the first preset time is not higher than the bid value of the first equipment for the voice information and is equal to the bid value of the first equipment for the voice information, marking the second equipment with the bid value equal to the bid value of the first equipment for the voice information as third equipment, and acquiring state information of the third equipment;
determining whether the first device is a wake-up device based on state information of the first device and state information of a third device, wherein the state information includes at least one of the following state parameters: the total time of the starting-up state, the position of the equipment, the residual electric quantity and whether the equipment is in the use state within the second preset time.
As another embodiment of the present application, the bidding unit 1703 is further configured to:
if the bidding value of the second equipment to the voice information received by the first equipment within the first preset time is lower than the bidding value of the first equipment to the voice information, determining whether the bidding value of the first equipment to the voice information is higher than a preset value;
and if the bidding value of the first equipment on the voice information is higher than the preset value, determining that the first equipment is awakening equipment.
As another embodiment of the present application, the bidding unit 1703 is further configured to:
if the bidding value of the first equipment to the voice information is lower than or equal to a preset value, determining whether fourth equipment exists in a network where the first equipment is located, wherein the fourth equipment is electronic equipment, voice print information of which is matched with the voice information, in second equipment, the first equipment does not monitor the bidding value within first preset time;
and if the fourth equipment exists in the network where the first equipment is located, sending a wakeup instruction corresponding to the voice information to the fourth equipment so as to indicate the fourth equipment to execute wakeup operation corresponding to the voice information under the condition that the bid value of the fourth equipment on the voice information is higher than the preset value.
As another embodiment of the present application, the second device is an electronic device located in the same network as the first device.
Accordingly, bidding unit 1703 is also used to:
if the information sent by the second equipment is not received within the first preset time, determining whether the competitive value of the first equipment on the voice information is higher than a preset value;
if the bidding value of the first equipment on the voice information is higher than the preset value, determining that the first equipment is awakening equipment;
if the bidding value of the first equipment to the voice information is lower than or equal to a preset value, determining whether fourth equipment exists in a network where the first equipment is located, wherein the fourth equipment is electronic equipment, voice print information of which is matched with the voice information, in second equipment, the first equipment does not monitor the bidding value within first preset time;
and if the fourth equipment exists in the network where the first equipment is located, sending a wakeup instruction corresponding to the voice information to the fourth equipment so as to indicate the fourth equipment to execute wakeup operation corresponding to the voice information under the condition that the bid value of the fourth equipment on the voice information is higher than the preset value.
As another embodiment of the present application, the bidding unit 1703 is further configured to:
and acquiring a wakeup voiceprint of a second device, which does not monitor the competitive value in a first preset time, of the first device.
As another embodiment of the present application, the bidding unit is further configured to:
when the first equipment is preset voiceprint summarizing equipment, receiving awakening voiceprints of other electronic equipment except the first equipment in a network where the first equipment is located;
generating a voiceprint characteristic summary based on the awakening voiceprint of the first device and the received awakening voiceprint of the other electronic device, and sending the voiceprint characteristic summary to the other electronic device;
acquiring awakening voiceprints of second equipment, wherein the first equipment does not monitor the competitive value within first preset time, from the voiceprint characteristic collection;
when the first device is not the preset voiceprint collecting device, sending a wake-up voiceprint of the first device to the preset voiceprint collecting device so as to instruct the voiceprint collecting device to generate voiceprint characteristic collection of all electronic devices in a network where the voiceprint collecting device is located;
receiving voiceprint characteristic summary sent by voiceprint summary equipment;
and acquiring a wakeup voiceprint of the second equipment, which does not monitor the bid value within the first preset time, of the first equipment from the voiceprint characteristic collection.
As another embodiment of the present application, the second device is a preset decision device, and the decision device is configured to receive bid values of the voice information sent by each electronic device in a network where the decision device is located, send a first message to the electronic device with the highest bid value, and send a second message to the electronic device with the highest bid value;
bidding unit 1703 is also used to:
if the received information sent by the second device is a first message, determining that the first device is a wake-up device;
and if the received information sent by the second device is the second message, determining that the first device is not the awakening device.
As another embodiment of the present application, the information interaction unit 1702 is further configured to:
acquiring an awakening mode corresponding to the voice information, wherein the awakening mode comprises the following steps: waking up a scene;
acquiring the bidding value of the first equipment in the wake-up mode, and taking the bidding value as the bidding value of the first equipment on the voice information;
and sending the competitive value of the first equipment to the voice information to the second equipment.
As another embodiment of the present application, the information interaction unit 1702 is further configured to:
obtaining a preset competitive value of the first equipment in an awakening mode;
or the like, or, alternatively,
acquiring the device parameter of the first device and the dependence of the wake-up mode on the device parameter;
and generating a bid value of the first device in the wake mode based on the device parameters of the first device and the dependence of the wake mode on the device parameters.
As another embodiment of the present application, the wake-up mode further includes: a wake-up location;
the information interaction unit 1702 is further configured to:
and determining a wake-up place corresponding to the voice information based on the network to which the first equipment is currently connected.
It should be noted that, for the above-mentioned information interaction, execution process and other contents between the units in the first device, the specific functions and technical effects thereof are based on the same concept as that of the method embodiment of this application, and reference may be made to the method embodiment section specifically, and details are not described here again.
It will be clear to those skilled in the art that, for convenience and simplicity of description, the above division of the functional units is merely illustrated, and in practical applications, the above function distribution may be performed by different functional units according to needs, that is, the internal structure of the first device is divided into different functional units to perform all or part of the above described functions. Each functional unit in the embodiments may be integrated into one processing unit, or each unit may exist alone physically, or two or more units are integrated into one unit, and the integrated unit may be implemented in a form of hardware, or in a form of software functional unit. In addition, specific names of the functional units are only used for distinguishing one functional unit from another, and are not used for limiting the protection scope of the application. For the specific working process of each unit in the electronic device, reference may be made to the corresponding process in the foregoing method embodiment, which is not described herein again.
The embodiments of the present application further provide a computer-readable storage medium, where a computer program is stored, and when the computer program is executed by a processor, the steps in the above-mentioned method embodiments may be implemented.
Embodiments of the present application further provide a computer program product, which when run on an electronic device, enables the electronic device to implement the steps in the above method embodiments.
The integrated unit, if implemented in the form of a software functional unit and sold or used as a stand-alone product, may be stored in a computer readable storage medium. Based on such understanding, all or part of the processes in the methods of the embodiments described above can be implemented by a computer program, which can be stored in a computer readable storage medium and used by a processor to implement the steps of the embodiments of the methods described above. Wherein the computer program comprises computer program code which may be in the form of source code, object code, an executable file or some intermediate form, etc. The computer readable medium may include at least: any entity or apparatus capable of carrying computer program code to a first device, a recording medium, computer Memory, Read-Only Memory (ROM), Random-Access Memory (RAM), an electrical carrier signal, a telecommunications signal, and a software distribution medium. Such as a usb-disk, a removable hard disk, a magnetic or optical disk, etc. In certain jurisdictions, computer-readable media may not be an electrical carrier signal or a telecommunications signal in accordance with legislative and proprietary practices.
An embodiment of the present application further provides a chip system, where the chip system includes a processor, the processor is coupled to the memory, and the processor executes a computer program stored in the memory to implement the steps of any of the method embodiments of the present application. The chip system may be a single chip or a chip module composed of a plurality of chips.
In the above embodiments, the descriptions of the respective embodiments have respective emphasis, and reference may be made to the related descriptions of other embodiments for parts that are not described or recited in a certain embodiment.
Those of ordinary skill in the art will appreciate that the various illustrative elements and method steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, or combinations of computer software and electronic hardware. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the implementation. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present application.
The above embodiments are only used to illustrate the technical solutions of the present application, and not to limit the same; although the present application has been described in detail with reference to the foregoing embodiments, it should be understood by those skilled in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some technical features may be equivalently replaced; such modifications and substitutions do not substantially depart from the spirit and scope of the exemplary embodiments of the present application and are intended to be included within the scope of the present application.

Claims (15)

1. A voice wake-up method, comprising:
responding to the monitored voice information, and matching the voice information with a wake-up voiceprint of the first equipment by the first equipment;
if the voice information is matched with the awakening voiceprint of the first device, the first device sends a bidding value of the first device for the voice information to a second device, and monitors information sent by the second device, wherein the bidding value of the first device for the voice information is the supporting capacity of the first device for awakening operation corresponding to the voice information;
in response to the monitored information sent by the second device, the first device determines whether the first device is a wake-up device based on the information sent by the second device;
and if the first equipment is awakening equipment, the first equipment executes awakening operation corresponding to the voice information.
2. The method of claim 1, wherein the second device is an electronic device located in the same network as the first device, and wherein the information sent by the second device includes a bid value of the second device for the voice information;
correspondingly, the determining, by the first device, whether the first device is a wake-up device based on the information sent by the second device includes:
if the bidding value of the second device on the voice information is higher than the bidding value of the first device on the voice information in the bidding values of the second device on the voice information received by the first device within a first preset time, determining that the first device is not a wake-up device;
and if the bid value of the second equipment for the voice information received by the first equipment within the first preset time is lower than the bid value of the first equipment for the voice information, determining that the first equipment is awakening equipment.
3. The method of claim 2, wherein the first device determining whether the first device is a wake-up device based on information sent by the second device further comprises:
if the bid value of the second device for the voice information received by the first device within the first preset time is not higher than the bid value of the first device for the voice information and is equal to the bid value of the first device for the voice information, marking the second device with the bid value equal to the bid value of the first device for the voice information as a third device, and acquiring state information of the third device;
the first device determines whether the first device is a wake-up device based on state information of the first device and state information of the third device, wherein the state information includes at least one of the following state parameters: the total time length of the starting-up state, the position of the equipment, the residual electric quantity and whether the equipment is in the use state within the second preset time.
4. The method of claim 2, wherein the determining that the first device is a wake-up device if the bid value of the second device for the voice message received by the first device within the first preset time is lower than the bid value of the first device for the voice message comprises:
if the bid value of the second equipment for the voice information received by the first equipment within a first preset time is lower than the bid value of the first equipment for the voice information, the first equipment determines whether the bid value of the first equipment for the voice information is higher than a preset value;
and if the bidding value of the first equipment on the voice information is higher than a preset value, determining that the first equipment is awakening equipment.
5. The method of claim 4, wherein after the first device determines whether the bid value of the first device for the voice information is above a preset value, further comprising:
if the bidding value of the first device on the voice information is lower than or equal to the preset value, the first device determines whether a fourth device exists in a network where the first device is located, wherein the fourth device is an electronic device, in which the voice print information is matched with the voice information, in a second device, the first device does not monitor the bidding value within a first preset time;
if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice information to the fourth device to indicate the fourth device to execute wakeup operation corresponding to the voice information when the bid value of the fourth device for the voice information is higher than the preset value.
6. The method of any of claims 1 to 5, wherein the second device is an electronic device located in the same network as the first device, the method further comprising:
if the first equipment does not receive the information sent by the second equipment within first preset time, the first equipment determines whether the bidding value of the first equipment on the voice information is higher than a preset value or not;
if the bidding value of the first equipment on the voice information is higher than the preset value, the first equipment determines that the first equipment is awakening equipment;
if the bidding value of the first device on the voice information is lower than or equal to the preset value, the first device determines whether a fourth device exists in a network where the first device is located, wherein the fourth device is an electronic device, in which the voice print information is matched with the voice information, in a second device, the first device does not monitor the bidding value within a first preset time;
if the fourth device exists in the network where the first device is located, the first device sends a wakeup instruction corresponding to the voice message to the fourth device to indicate the fourth device to execute wakeup operation corresponding to the voice message when the bid value of the fourth device for the voice message is higher than a preset value.
7. The method of claim 5 or 6, before the first device determines whether a fourth device is present in a network in which the first device is located, further comprising:
the first equipment acquires a wakeup voiceprint of second equipment, wherein the first equipment does not monitor the bid value within first preset time.
8. The method of claim 7, wherein the obtaining, by the first device, a wake-up voiceprint of a second device to which the first device does not listen for a bid value within a first preset time comprises:
when the first equipment is preset voiceprint summarizing equipment, the first equipment receives awakening voiceprints of other electronic equipment except the first equipment in a network where the first equipment is located;
the first device generates a voiceprint characteristic summary based on the awakening voiceprint of the first device and the received awakening voiceprint of the other electronic device, and sends the voiceprint characteristic summary to the other electronic device;
the first equipment acquires awakening voiceprints of second equipment, which do not monitor the competitive value within first preset time, of the first equipment from the voiceprint characteristic collection;
when the first device is not a preset voiceprint summarizing device, the first device sends a wake-up voiceprint of the first device to the preset voiceprint summarizing device so as to indicate the voiceprint summarizing device to generate voiceprint characteristic summarization of all electronic devices in a network where the voiceprint summarizing device is located;
the first device receives the voiceprint characteristic summary sent by the voiceprint summary device;
and the first equipment acquires the awakening voiceprint of the second equipment which does not monitor the bid value within the first preset time from the voiceprint characteristic collection.
9. The method of claim 1, wherein the second device is a preset decision device, and the decision device is used for receiving bid values of the voice information sent by electronic devices in a network where the decision device is located, sending a first message to the electronic device with the highest bid value, and sending a second message to the electronic device with the lower bid value;
correspondingly, the determining, by the first device, whether the first device is a wake-up device based on the information sent by the second device includes:
if the information sent by the second device and received by the first device is a first message, the first device determines that the first device is a wake-up device;
if the information sent by the second device and received by the first device is a second message, the first device determines that the first device is not a wake-up device.
10. The method of any of claims 1 to 9, wherein the first device sending a bid value for the voice information by the first device to a second device comprises:
the first device obtains an awakening mode corresponding to the voice information, wherein the awakening mode comprises: waking up a scene;
the first equipment acquires the competitive value of the first equipment in the awakening mode, and the competitive value is used as the competitive value of the first equipment for the voice information;
and the first equipment sends the bid value of the first equipment for the voice information to the second equipment.
11. The method of claim 10, wherein the first device obtaining a bid value for the first device in the awake mode comprises:
the first equipment acquires preset competitive value of the first equipment in the awakening mode;
or the like, or, alternatively,
the first equipment acquires equipment parameters of the first equipment and the dependence of the wake-up mode on the equipment parameters;
and the first equipment generates a bid value of the first equipment in the wake-up mode based on the equipment parameters of the first equipment and the dependence of the wake-up mode on the equipment parameters.
12. The method of claim 10, wherein the awake mode further comprises: a wake-up location; the acquiring, by the first device, the wake-up mode corresponding to the voice information includes:
and the first equipment determines the awakening place corresponding to the voice information based on the network to which the first equipment is currently connected.
13. An electronic device, comprising:
the voiceprint matching unit is used for responding to the monitored voice information and matching the voice information with the awakening voiceprint of the electronic equipment;
the information interaction unit is used for sending a bidding value of the electronic equipment on the voice information to second equipment and monitoring information sent by the second equipment if the voice information is matched with the awakening voiceprint of the electronic equipment, wherein the bidding value of the electronic equipment on the voice information is the supporting capacity of the electronic equipment on the awakening operation corresponding to the voice information;
the bidding unit is used for responding to the monitored information sent by the second equipment and determining whether the electronic equipment is a wake-up equipment or not based on the information sent by the second equipment;
and the awakening unit is used for executing awakening operation corresponding to the voice information if the electronic equipment is awakening equipment.
14. An electronic device, characterized in that the electronic device comprises a processor for executing a computer program stored in a memory for implementing the method according to any of claims 1 to 12.
15. A chip system, characterized in that the chip system comprises a processor coupled with a memory for executing a computer program stored in the memory for implementing the method according to any of claims 1 to 12.
CN202010800181.1A 2020-08-10 2020-08-10 Voice wake-up method, electronic equipment and chip system Active CN114120987B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010800181.1A CN114120987B (en) 2020-08-10 2020-08-10 Voice wake-up method, electronic equipment and chip system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010800181.1A CN114120987B (en) 2020-08-10 2020-08-10 Voice wake-up method, electronic equipment and chip system

Publications (2)

Publication Number Publication Date
CN114120987A true CN114120987A (en) 2022-03-01
CN114120987B CN114120987B (en) 2023-01-13

Family

ID=80373566

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010800181.1A Active CN114120987B (en) 2020-08-10 2020-08-10 Voice wake-up method, electronic equipment and chip system

Country Status (1)

Country Link
CN (1) CN114120987B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115509627A (en) * 2022-11-22 2022-12-23 威海海洋职业学院 Electronic equipment awakening method and system based on artificial intelligence

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090043580A1 (en) * 2003-09-25 2009-02-12 Sensory, Incorporated System and Method for Controlling the Operation of a Device by Voice Commands
US8340975B1 (en) * 2011-10-04 2012-12-25 Theodore Alfred Rosenberger Interactive speech recognition device and system for hands-free building control
CN107004412A (en) * 2014-11-28 2017-08-01 微软技术许可有限责任公司 Equipment arbitration for audiomonitor
CN109391528A (en) * 2018-08-31 2019-02-26 百度在线网络技术(北京)有限公司 Awakening method, device, equipment and the storage medium of speech-sound intelligent equipment
CN110349578A (en) * 2019-06-21 2019-10-18 北京小米移动软件有限公司 Equipment wakes up processing method and processing device

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090043580A1 (en) * 2003-09-25 2009-02-12 Sensory, Incorporated System and Method for Controlling the Operation of a Device by Voice Commands
US8340975B1 (en) * 2011-10-04 2012-12-25 Theodore Alfred Rosenberger Interactive speech recognition device and system for hands-free building control
CN107004412A (en) * 2014-11-28 2017-08-01 微软技术许可有限责任公司 Equipment arbitration for audiomonitor
CN109391528A (en) * 2018-08-31 2019-02-26 百度在线网络技术(北京)有限公司 Awakening method, device, equipment and the storage medium of speech-sound intelligent equipment
CN110349578A (en) * 2019-06-21 2019-10-18 北京小米移动软件有限公司 Equipment wakes up processing method and processing device

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115509627A (en) * 2022-11-22 2022-12-23 威海海洋职业学院 Electronic equipment awakening method and system based on artificial intelligence

Also Published As

Publication number Publication date
CN114120987B (en) 2023-01-13

Similar Documents

Publication Publication Date Title
CN112289313A (en) Voice control method, electronic equipment and system
WO2020019176A1 (en) Method for updating wake-up voice of voice assistant by terminal, and terminal
CN111835907A (en) Method, equipment and system for switching service across electronic equipment
CN113676339B (en) Multicast method, device, terminal equipment and computer readable storage medium
CN114422340A (en) Log reporting method, electronic device and storage medium
CN112651510A (en) Model updating method, working node and model updating system
CN111930335A (en) Sound adjusting method and device, computer readable medium and terminal equipment
WO2022022319A1 (en) Image processing method, electronic device, image processing system and chip system
CN113467735A (en) Image adjusting method, electronic device and storage medium
CN113593567A (en) Method for converting video and sound into text and related equipment
CN114120987B (en) Voice wake-up method, electronic equipment and chip system
CN109285563B (en) Voice data processing method and device in online translation process
CN111819830B (en) Information recording and displaying method and terminal in communication process
WO2022206825A1 (en) Method and system for adjusting volume, and electronic device
CN113467747B (en) Volume adjusting method, electronic device and storage medium
WO2022161077A1 (en) Speech control method, and electronic device
WO2022095752A1 (en) Frame demultiplexing method, electronic device and storage medium
CN115393676A (en) Gesture control optimization method and device, terminal and storage medium
CN114116610A (en) Method, device, electronic equipment and medium for acquiring storage information
WO2020034104A1 (en) Voice recognition method, wearable device, and system
CN113364067B (en) Charging precision calibration method and electronic equipment
CN114079694B (en) Control labeling method and device
CN114449492B (en) Data transmission method and terminal equipment
US20240045651A1 (en) Audio Output Method, Media File Recording Method, and Electronic Device
WO2023197997A1 (en) Wearable device, and sound pickup method and apparatus

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant