WO2020001172A1 - 一种音频处理方法和电子设备 - Google Patents
一种音频处理方法和电子设备 Download PDFInfo
- Publication number
- WO2020001172A1 WO2020001172A1 PCT/CN2019/086193 CN2019086193W WO2020001172A1 WO 2020001172 A1 WO2020001172 A1 WO 2020001172A1 CN 2019086193 W CN2019086193 W CN 2019086193W WO 2020001172 A1 WO2020001172 A1 WO 2020001172A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- condition
- input data
- meets
- preset
- audio
- Prior art date
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F9/00—Arrangements for program control, e.g. control units
- G06F9/06—Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
- G06F9/44—Arrangements for executing specific programs
- G06F9/4401—Bootstrapping
- G06F9/4418—Suspend and resume; Hibernate and awake
Definitions
- the present application relates to the field of electronic devices, and more particularly, to an audio processing method and an electronic device.
- this application provides an audio processing method, which solves the problem that the device is prone to false wakeup in the prior art.
- An audio processing method applied to a first device includes:
- the input data satisfying the first condition does not satisfy the second condition, the input data satisfying the first condition is ignored.
- the input data satisfying the first condition is used to switch a state of a preset application to a preset running state, and after responding to the input data in a manner that meets the first condition, the method further includes:
- control data so that a preset application in a preset running state responds to the control data.
- responding to the input data includes:
- the response data is output in the first manner.
- the method when outputting multimedia content, after collecting input data, the method further includes:
- determining whether the input data satisfies a second condition includes:
- the first information includes at least one of the following:
- the second device collects the input data
- the second device performs an operation in response to the input data.
- determining whether the input data meets a second condition includes:
- determining whether the input data meets a second condition includes:
- the image meeting the preset condition includes at least one of the following:
- the identity of the person in the identified image meets a preset identity condition
- the identified person in the image faces the first device.
- An electronic device includes:
- a judging module configured to judge whether the input data satisfies a first condition, and whether the input data satisfies a second condition;
- a processing module configured to respond to the input data in a manner satisfying the first condition if the input data meets the first condition and the second condition; and if the input data meets the first condition and does not satisfy the second condition, Ignore the input data that meets the first condition.
- An electronic device includes:
- a processor configured to receive the collected input data, and if the input data meets the first condition and the second condition, respond to the input data in a manner that satisfies the first condition; and if the input data meets the first condition and The second condition is not satisfied, and the input data satisfying the first condition is ignored;
- a memory for storing the first condition and the second condition.
- the above electronic equipment preferably further includes:
- Audio collector for collecting voice audio
- the memory also stores preset voiceprint information
- the processor is specifically configured to determine whether the voice audio matches preset voiceprint information
- Audio collector for collecting voice audio
- An image acquisition module for acquiring images of an image acquisition area
- the memory further stores a preset condition
- the processor is specifically configured to analyze and determine whether the voice audio meets a first condition, and determine whether the image meets a preset condition.
- this application provides an audio processing method, including: collecting input data; if the input data satisfying the first condition satisfies the second condition, satisfying the first condition In a manner that responds to the input data; if the input data that meets the first condition does not satisfy the second condition, the input data that meets the first condition is ignored.
- this method by judging whether the input data satisfying the first condition meets the second condition, determining whether to respond to the input data in the manner of the first condition, two conditions of the input data are judged, and the judgment accuracy is high to prevent Wake up by mistake.
- FIG. 1 is a flowchart of Embodiment 1 of an audio processing method provided by this application;
- Embodiment 2 is a flowchart of Embodiment 2 of an audio processing method provided by this application;
- Embodiment 3 is a flowchart of Embodiment 3 of an audio processing method provided by the present application.
- FIG. 4 is a schematic diagram of display content in Embodiment 3 of an audio processing method provided by the present application.
- Embodiment 4 is a flowchart of Embodiment 4 of an audio processing method provided by the present application.
- Embodiment 5 is a flowchart of Embodiment 5 of an audio processing method provided by the present application.
- Embodiment 7 is a schematic diagram of a specific example in Embodiment 5 of an audio processing method provided by this application;
- Embodiment 8 is a flowchart of Embodiment 6 of an audio processing method provided by the present application.
- Embodiment 9 is a flowchart of Embodiment 7 of an audio processing method provided by this application.
- Embodiment 10 is a schematic diagram of a specific example in Embodiment 7 of an audio processing method provided by this application;
- Embodiment 1 is a schematic structural diagram of Embodiment 1 of an electronic device provided by this application.
- Embodiment 2 is a schematic structural diagram of Embodiment 2 of an electronic device according to the present application.
- Embodiment 3 is a schematic structural diagram of Embodiment 3 of an electronic device according to the present application.
- FIG. 14 is a schematic structural diagram of Embodiment 4 of an electronic device provided by the present application.
- Embodiment 1 of an audio processing method provided by the present application is applied to an electronic device.
- the electronic device serves as the first device, and the method includes the following steps:
- Step S101 collecting input data
- the input data is data input to the first device.
- the input data may be audio, video, image, or data transmitted from other devices.
- Step S102 if the input data satisfying the first condition satisfies the second condition, respond to the input data in a manner satisfying the first condition;
- the input data When the input data meets the first condition and the second condition at the same time, the input data is responded to in a manner that the first condition is satisfied.
- Step S103 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- the first condition is that the audio contains a wake-up word, for example, the wake-up word is "Hi, voice assistant", and the wake-up word is used to wake up the voice in the first device.
- Assistant then, in response to the input data, the voice assistant in the first device is woken up.
- the second condition is a supplement to the first condition.
- the input data is responded to in a manner that satisfies the first condition.
- the first device does not respond to the wake-up word, that is, the voice assistant.
- the second condition may be other conditions related to the first device, such as the audio conditions of the user who sent the audio, other feedback related to the first device, or the behavior of the user.
- the second condition will be explained in detail in subsequent embodiments, and will not be described in detail in this embodiment.
- an audio processing method includes: collecting input data; if the input data satisfying a first condition satisfies a second condition, responding to the input data in a manner satisfying the first condition; The input data of the first condition does not satisfy the second condition, and the input data of the first condition is ignored.
- this method by judging whether the input data satisfying the first condition meets the second condition, determining whether to respond to the input data in the manner of the first condition, two conditions of the input data are judged, and the judgment accuracy is high to prevent Wake up by mistake.
- the input data satisfying the first condition is used to switch a state of a preset application to a preset running state.
- FIG. 2 it is a flowchart of Embodiment 2 of an audio processing method provided by the present application.
- the method includes the following steps:
- Step S201 collecting input data
- Step S202 if the input data that meets the first condition meets the second condition, respond to the input data in a manner that satisfies the first condition;
- Steps S201-202 are consistent with steps S101-102 in Embodiment 1, and details are not described in this embodiment.
- Step S203 Collect control data so that a preset application in a preset running state responds to the control data
- the input data satisfies the first condition and the second condition, and responds to the input data in a manner that satisfies the first condition, so that a state of a preset application in the first device is switched to a preset running state.
- the preset operating state is a normal operating state or an activated state.
- the input control data is continuously collected, and the preset application responds to the data.
- the preset application is a voice assistant in the first device, and the preset running state is activated.
- the first device continues to collect input control data, such as a voice command "type Call Li Ming ", the voice assistant responds to the voice control instruction and executes the operation of controlling the phone software in the first device to" call Li Ming ".
- the control data that continues to be collected is a voice control instruction "open browser", and the voice assistant responds to the voice control instruction and performs an operation of controlling the opening of the browser software in the first device.
- Step S204 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Step S204 is the same as step S103 in Embodiment 1, and details are not described in this embodiment.
- the audio processing method provided in this embodiment further includes: collecting control data, so that a preset application in a preset running state responds to the control data.
- the method is adopted to respond to the input data in a manner that satisfies the first condition, to switch a state of a preset application in the first device to a preset running state, and to continue to collect input control data in a subsequent process, and Make the preset application respond to the control data to ensure that the preset application normally performs operations.
- the first device outputs multimedia content in a first manner.
- a flowchart of Embodiment 3 of an audio processing method provided by this application includes the following steps:
- Step S301 collecting input data
- Step S301 is the same as step S101 in Embodiment 1, and details are not described in this embodiment.
- Step S302 if the input data satisfying the first condition meets the second condition, output response data in the first manner;
- the first device outputs response data in a manner that affects the output of multimedia content, and outputting the response data will interfere with the output of the multimedia content.
- the second condition is used to determine whether the first device does not need to respond to the input data, and the first device needs to respond to the input data, then the input data meets the second condition, otherwise, the input data does not meet the second condition .
- the input data is collected, and the output mode of the multimedia content corresponds to the way in which the first device responds to the input data, which is the first mode.
- the first device When the first device outputs response data, it may affect its output of multimedia content. Therefore, it is necessary to determine that when the input data that meets the first condition also meets the second condition, the first device outputs response data and the user can receive The response.
- the first device displays content (such as a video or an image) on a screen
- content such as a video or an image
- an output response is achieved by displaying a prompt box on the screen, the prompt box occupies part of the screen and blocks the original display content in the part of the screen.
- the first device when the first device plays content (such as audio) through a speaker, the first device implements an output response by playing audio “starting a voice assistant”, which overlaps with the content being played.
- content such as audio
- starting a voice assistant the first device implements an output response by playing audio “starting a voice assistant”, which overlaps with the content being played.
- the display content diagram includes a display interface 401 that displays an image.
- a prompt box 402 is displayed on the display interface, and the prompt box prompts “Start Voice assistant.”
- Step S303 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Step S303 is the same as step S103 in Embodiment 1, and details are not described in this embodiment.
- responding to the input data includes outputting response data in the first manner.
- response data is output in the same manner as the device outputs multimedia content, ensuring that the user can understand that the first device has responded to the input data.
- a flowchart of Embodiment 4 of an audio processing method provided by the present application includes the following steps:
- Step S501 collecting input data
- Step S501 is the same as step S101 in Embodiment 1, and details are not described in this embodiment.
- Step S502 determine whether the input data meets a first condition
- Step S503 determine whether the input data satisfies a second condition based on the input data meeting the first condition
- the input data is audio
- the first condition is that the audio contains a wake-up word
- the device is a device awakened by a specific user purpose. It is also necessary to determine the information related to the first device / user according to the situation, that is, to determine whether the input data meets the second condition to prevent the device from being awakened by a specific user or the user's purpose. Not the first device, but causing the problem of false wake-up.
- this application does not limit the order of determining whether the input data meets the first condition and the second condition, and can determine whether the input data meets the first condition; based on the input data,
- the first condition is described to determine whether the input data satisfies the second condition; it is also possible to determine whether the input data meets the second condition; and based on the input data meeting the second condition, to determine whether the input data meets the first condition condition.
- Step S504 if the input data satisfying the first condition satisfies the second condition, respond to the input data in a manner satisfying the first condition;
- Step S505 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Steps S504-505 are the same as steps S102-103 in Embodiment 1, and details are not described in this embodiment.
- the audio processing method first determine whether the input data meets the first condition, and based on the input data meeting the first condition, determine whether the input data meets the second condition.
- this method by judging whether the input data satisfying the first condition meets the second condition, determining whether to respond to the input data in the manner of the first condition, two conditions of the input data are judged, and the judgment accuracy is high to prevent Wake up by mistake.
- a flowchart of Embodiment 5 of an audio processing method provided by this application includes the following steps:
- Step S601 collecting input data
- Step S602 determine whether the input data meets a first condition
- Steps S601-602 are consistent with steps S501-502 in Embodiment 4, and details are not described in this embodiment.
- Step S603 determine whether the first information fed back by the second device is received based on the input data meeting the first condition
- the second device and the first device form a networked system, and data in the networked system is shared.
- the first device and the second device can be in the same environment, and they can collect the same content in the environment, such as collecting the same input data, and the devices in the networked system can collect the input data after collecting them. Relevant collected information and / or information on the input data is fed back to other devices.
- the first information includes at least one of the following:
- the second device collects the input data
- the second device performs an operation in response to the input data.
- the system may include various electronic devices such as a mobile phone, a tablet computer, a television, a refrigerator, and an air conditioner.
- Step S604 determine whether the input data satisfies a second condition based on receiving the first information
- the first device may determine whether the input data collected by the first device meets the second condition by combining the first information.
- the first information is that when the second device collects input data, the first device collects the input data later than the second device, and then the second device can be analyzed to obtain that the second device is closer to the user.
- the second device is The device waken up by the user's purpose, then, the input data does not satisfy the second condition; when the first device does not receive the first information, the first device is the earliest to collect the input data, and the first data can be analyzed and obtained.
- the device is closest to the user, and the first device is the device awakened by the user's purpose. Then, the input data satisfies the second condition.
- the first information is the quality of the input data collected by the second device.
- the intensity of the input device collected by the second device is 9 and the intensity of the input data collected by the first device is 4. Then it can be analyzed that the second device is closer to the user, and the second device is a device awakened by the user's purpose. Then, the input data does not meet the second condition; the intensity of the input device collected by the second device is 2, and the The intensity of the input data collected by the first device is 8, and then the first device can be analyzed to obtain that the user is closer to the user. The first device is a device awakened by the user. Then, the input data meets the second condition.
- the first information is an operation performed by the second device in response to the input data
- the second device since the second device has responded to the input data before the first device collects the first information and has not responded, then It can be seen that the second device is a device waken up by the user's purpose, then the input data does not satisfy the second condition; if the first information is not received, then it can be seen that the first device has a faster acquisition speed, and the first device It is a device awakened by the user, then, the input data satisfies the second condition.
- the input data is audio
- the audio is generated when the user 701 speaks a specific wake-up word “Hi, voice assistant”, and the mobile phone 702, tablet computer 703, and The voice assistants in TV 704 can all be awakened by this particular wake word.
- the mobile phone, tablet computer, and TV are all capable of collecting audio in the environment, and the distance between the three and the user is mobile phone, TV, and tablet computer, respectively.
- the collection action is fed back to other devices.
- the collection speed of the three devices is from fast to slow: mobile phones, TVs, and tablet computers.
- the mobile phone collects audio, it feeds back the audio information it has collected to the TV and tablet computer.
- the mobile phone does not receive information from other devices.
- the mobile phone responds to the audio and wakes up its voice assistant; and the TV and the tablet computer get the feedback information, and it can be known that the mobile phone has collected audio before it, then the TV and the tablet computer do not respond to the collected audio .
- the audio quality of the device can be fed back to other devices.
- the collection intensity / definition of the three devices is from large to small: mobile phones, TVs, and tablet computers. After each device collects audio, it feeds back the quality of the collected audio to other devices. Since the audio quality in mobile phones is the best, then The mobile phone responds to the audio and wakes up its voice assistant; while the TV and tablet computer obtains the feedback information, it can be known that the audio quality of other devices is better than themselves, then the TV and tablet computer do not respond to the audio they collect.
- the audio is responded, and the information of the response operation is fed back to other devices.
- the three devices responded from fast to slow: mobile phones, TVs, and tablets.
- the mobile phone collects audio, it responds to the audio, wakes up its voice assistant, and feeds back the response operation information to the TV and tablet.
- the TV and the tablet computer obtain the feedback information, it can be known that the mobile phone has responded to the audio, then the TV and the tablet computer do not respond to the audio that they have collected.
- Step S605 if the input data that meets the first condition meets the second condition, respond to the input data in a manner that satisfies the first condition;
- Step S606 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Steps S605-606 are the same as steps S504-505 in Embodiment 5, and details are not described in this embodiment.
- determining whether the input data meets a second condition includes: determining whether the first information fed back by the second device is received; and based on receiving the first information, Determining whether the input data satisfies a second condition; wherein the first information includes at least one of the following: the second device collects the input data; or the second device collects the input data Quality; or the second device performs an operation in response to the input data.
- the first device and the second device are used to collect input data or input data quality or whether to respond to the input data for information feedback.
- Data is shared between the devices so that which device is determined based on the shared information. It is a device that is awakened by the user's purpose, ensuring that the device that the user wants to wake up is woken up, preventing the problem of accidental wakeup.
- the input data is voice audio.
- a flowchart of Embodiment 6 of an audio processing method provided by the present application includes the following steps:
- Step S801 collect input data
- Step S802 determine whether the input data meets a first condition
- Steps S801-802 are consistent with steps S501-502 in Embodiment 4, and details are not described in this embodiment.
- Step S803 Based on the input data meeting the first condition, determine whether the voice audio matches preset voiceprint information, and the preset voiceprint information is preset voiceprint information to wake people up;
- the input data meets the second condition
- the input data satisfies the first condition, that is, the voice audio contains a specific wake-up word.
- voiceprint information is preset in the first device, and the preset voiceprint information is voiceprint information preset to wake people up. Determine whether the voice audio matches the preset voiceprint information. If the two match, the person who issued the voice audio is the preset wake-up person who has the right to wake up the voice assistant of the first device. If the two do not match, then The person issuing the voice audio is not a preset wake-up person, and has no permission to wake up the voice assistant of the first device.
- user A uses a mobile phone and user B uses a tablet computer.
- the wake-up words of the voice assistant in both devices are "Hi, voice assistant".
- B speaks the voice "Hi, voice assistant”
- the phone will respond to the wake-up word after collecting the input data, and wake up the voice assistant, and user A of the phone does not want to wake up Voice assistant, which will cause A's experience to be poor.
- the second condition is set in the mobile phone. According to the voiceprint information, it can be determined that the voice is not sent by the user A, and the awake word is ignored and the voice assistant is not awakened.
- Step S804 if the input data that meets the first condition meets the second condition, respond to the input data in a manner that satisfies the first condition;
- Step S805 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Steps S804-805 are consistent with steps S504-505 in Embodiment 5, and details are not described in this embodiment.
- determining whether the input data satisfies a second condition includes determining whether the voice audio and preset voiceprint information Match, the preset voiceprint information is preset voiceprint information to wake people; based on the voice audio matching the preset voiceprint information, the input data meets the second condition; otherwise, the input data is not Meet the second condition.
- the input data includes images and audio.
- a flowchart of Embodiment 7 of an audio processing method provided by this application includes the following steps:
- Step S901 collect input data
- Step S902 determine whether the input data meets a first condition
- Steps S901-902 are consistent with steps S501-502 in Embodiment 4, and details are not described in this embodiment.
- Step S903 Based on the input data meeting the first condition, analyze and determine whether the image meets a preset condition;
- the input data meets a second condition
- the image meeting the preset condition includes at least one of the following:
- the identity of the person in the identified image meets a preset identity condition
- the identified person in the image faces the first device.
- the input data includes audio and images, and the first device can collect audio and images simultaneously.
- it may be determined whether the audio in the input data meets the first condition, and whether the image in the input data meets the preset condition.
- the first device when the user speaks the wake-up word, the first device, while acquiring audio information, also performs image acquisition on the image acquisition area, and the acquired image includes the user's image.
- the image is analyzed to obtain related information of the person in the image, such as features and postures.
- the character feature may include a face feature, a dynamic feature, and the like, and according to the character feature, whether the identity of the character is a specific wake-up person who meets a preset identity condition can be analyzed, and the specific wake-up person can wake up the device.
- the first device may preset information related to the character characteristics of the specific wake-up person.
- the specific wake-up person may be an authorized user, and only the authorized user can use the first device.
- the related information of the person in the image is a facial feature
- the image is recognized to obtain the facial feature of the person in the image, and whether the person is a specific wake-up person capable of waking up the device is determined based on the facial feature.
- the facial features match the facial features of a particular wake-up person, the input data satisfies the second condition, otherwise it does not.
- the consecutive frames of images are recognized to obtain the dynamic characteristics of the character (such as walking, waving, etc.) in the image, and it is determined whether the character is capable of When the specific wake-up person of the device is woken up, the input data meets the second condition when the dynamic feature matches the dynamic feature of the specific wake-up person, otherwise it is not satisfied.
- the character such as walking, waving, etc.
- the first device is provided with the personal characteristics of an authorized user.
- the first device obtains that the person who speaks the wake-up word matches the preset character characteristics according to the collected image analysis, and can respond to the wake-up word to wake up the first unprepared voice. assistant.
- the first device obtains that the person who speaks the wake-up word does not match the preset character characteristics according to the collected image analysis, so the wake-up word can be ignored and the first device is not woken up.
- Voice assistant When an authorized user speaks the wake-up word, the first device obtains that the person who speaks the wake-up word matches the preset character characteristics according to the collected image analysis, and can respond to the wake-up word to wake up the first unprepared voice. assistant.
- the image is identified to obtain whether the person in the image faces the first device. If the person faces the first device, the input data satisfies The second condition is not met otherwise.
- the device that they want to control / operate can be faced according to their own needs, so whether the user wants to operate / operate the device can be determined according to whether the user faces the device.
- FIG. 10 is a schematic diagram of a specific example.
- a user 1001 is surrounded by a mobile phone 1002, a tablet computer 1003, and a television 1004.
- the user faces the mobile phone 1002.
- Audio is generated when user 1001 speaks a specific wake-up word "Hi, voice assistant".
- the voice assistants in mobile phone 1002, tablet computer 1003, and TV 1004 can be awakened by the specific wake-up word.
- the mobile phone 1002, tablet computer 1003, and TV 1004 pair
- the image acquisition area performs image acquisition and analysis of the acquired images.
- the tablet computer 1003 analyzes the acquired images, and the result is that the user faces the tablet computer. If the input data meets the second condition, the tablet computer responds to the Wake word, wake up voice assistant.
- the user did not face himself, and the input data did not meet the second condition, and he did not respond to the wake word.
- Step S904 if the input data satisfying the first condition satisfies the second condition, respond to the input data in a manner satisfying the first condition;
- Step S905 If the input data satisfying the first condition does not satisfy the second condition, ignore the input data satisfying the first condition.
- Steps S904-905 are consistent with steps S504-505 in Embodiment 5, and details are not described in this embodiment.
- the input data includes an image and audio
- determining whether the input data meets a second condition includes analyzing and determining whether the image meets a preset condition; Based on that the image meets a preset condition, the input data meets a second condition; otherwise, the input data does not satisfy the second condition; wherein the image meets the preset condition includes at least one of the following: The identity of the person in the image meets a preset identity condition; or the identified person in the image faces the first device.
- the person in the image is analyzed to determine whether the person ’s identity meets the preset identity conditions or to determine whether the person is facing the device, determine whether the device is a device waken by the user ’s purpose, and prevent the device from waking up by the user ’s purpose. Wake up caused the problem of false wake up.
- this application also provides an embodiment of an electronic device to which the audio processing method is applied.
- FIG. 11 is a schematic structural diagram of Embodiment 1 of an electronic device provided by the present application.
- the electronic device has an audio collection function.
- the electronic device includes the following structures: a collection module 1101, a determination module 1102, and a processing module 1103.
- the acquisition module 1101 is used to collect input data.
- the determining module 1102 is configured to determine whether the input data meets a first condition, and whether the input data meets a second condition;
- the processing module 1103 is configured to respond to the input data in a manner that meets the first condition if the input data meets the first condition and the second condition; and if the input data meets the first condition and does not satisfy the first condition. Two conditions, ignoring the input data satisfying the first condition.
- the acquisition module may specifically adopt a device with an audio acquisition function such as a microphone; when the input data includes audio and images, the acquisition module may include an audio acquisition device (such as a microphone) and image acquisition Device (such as a camera).
- an audio acquisition device such as a microphone
- image acquisition Device such as a camera
- two conditions are imposed on the input data by judging whether the input data satisfying the first condition satisfies the second condition, and determining whether to respond to the input data in the manner of the first condition. Judgment, judgment accuracy is higher, to prevent false wake-up.
- At least some of the modules, sub-modules, units, or sub-units according to the embodiments of the present disclosure, or at least part of the functions of any of them, can be implemented in one module. Any one or more of the modules, sub-modules, units, and sub-units according to the embodiments of the present disclosure may be implemented by being divided into a plurality of modules.
- any one or more of the modules, sub-modules, units, and sub-units according to the embodiments of the present disclosure may be at least partially implemented as a hardware circuit, such as a field programmable gate array (FPGA), a programmable logic array (PLA), System-on-chip, system-on-substrate, system-on-package, application-specific integrated circuit (ASIC), or any other reasonable means of hardware or firmware that can integrate or package the circuit, or software, hardware, and firmware Any one of the implementation manners or an appropriate combination of any of them may be implemented.
- FPGA field programmable gate array
- PLA programmable logic array
- ASIC application-specific integrated circuit
- any one of the implementation manners or an appropriate combination of any of them may be implemented.
- one or more of the modules, submodules, units, and subunits according to the embodiments of the present disclosure may be at least partially implemented as computer program modules, and when the computer program modules are executed, corresponding functions may be performed.
- any number of the collection module 1101, the determination module 1102, and the processing module 1103 may be combined into one module, or any one of the modules may be split into multiple modules.
- at least part of the functions of one or more of these modules may be combined with at least part of the functions of other modules and implemented in one module.
- At least one of the acquisition module 1101, the determination module 1102, and the processing module 1103 may be at least partially implemented as a hardware circuit, such as a field programmable gate array (FPGA), a programmable logic array (PLA), System on a chip, system on a substrate, system on a package, application-specific integrated circuit (ASIC), or any other reasonable way to integrate or package a circuit such as hardware or firmware, or software, hardware, and firmware Any one of the implementation manners or an appropriate combination of any of them may be implemented.
- FPGA field programmable gate array
- PLA programmable logic array
- ASIC application-specific integrated circuit
- At least one of the acquisition module 1101, the determination module 1102, and the processing module 1103 may be implemented at least partially as a computer program module, and when the computer program module is executed, a corresponding function may be performed.
- FIG. 12 is a schematic structural diagram of Embodiment 2 of an electronic device provided by the present application.
- the electronic device includes the following structures: a processor 1201 and a memory 1202;
- the processor 1201 is configured to receive the collected input data, and if the input data meets the first condition and the second condition, respond to the input data in a manner that meets the first condition; and if the input data meets the first A condition that does not satisfy the second condition, and ignores the input data that meets the first condition;
- the memory 1202 is configured to store the first condition and the second condition.
- the processor may adopt a chip structure having data processing capabilities, such as a CPU (central processing unit, central processing unit) and the like.
- the first device outputs multimedia content in a first manner.
- the first mode may be a screen display mode or an audio playback mode.
- the first device when the first mode is a screen display mode, the first device further includes a display screen to display the multimedia content, and the response data corresponding to the input data is displayed on the display screen accordingly.
- the first device when the first mode is an audio playback mode, the first device further includes an audio player, such as a speaker, to implement audio playback of the multimedia content, and the response data corresponding to the input data is stored in the first device.
- the speaker plays.
- two conditions are imposed on the input data by judging whether the input data satisfying the first condition satisfies the second condition, and determining whether to respond to the input data in the manner of the first condition. Judgment, judgment accuracy is higher, to prevent false wake-up.
- the input data is voice audio.
- FIG. 13 is a schematic structural diagram of Embodiment 3 of an electronic device provided by the present application.
- the electronic device includes the following structures: a processor 1301, a memory 1302, and an audio collector 1303.
- the structural functions of the processor 1301 and the memory 1302 are consistent with the corresponding structural functions in Embodiment 2, and details are not described in this embodiment.
- the audio collector 1303 is used to collect voice audio
- the memory also stores preset voiceprint information
- the processor is specifically configured to determine whether the voice audio matches preset voiceprint information.
- the audio collector can adopt a device structure with an audio collection function such as a microphone.
- the input data is voice audio
- the voice audio by matching and judging the voice audio with the preset voiceprint information, it is determined whether the person sending the voice audio is a preset wake-up person, Prevents someone from waking up the device and causing it to wake up accidentally.
- the input data is voice audio and image.
- FIG. 14 is a schematic structural diagram of Embodiment 4 of an electronic device provided by the present application.
- the electronic device includes the following structures: a processor 1401, a memory 1402, an audio collector 1403, and an image acquisition module 1404.
- the structural functions of the processor 1401 and the memory 1402 are consistent with the corresponding structural functions in Embodiment 2, and details are not described in this embodiment.
- the audio collector 1403 is used to collect voice audio
- the image acquisition module 1404 is used to acquire an image of an image acquisition area, and the figure includes a person image.
- the memory further stores a preset condition
- the processor is specifically configured to analyze and determine whether the voice audio meets a first condition, and determine whether the image meets a preset condition.
- the image meeting the preset condition includes at least one of the following:
- the identity of the person in the identified image meets a preset identity condition
- the identified person in the image faces the first device.
- the person in the image is analyzed to determine whether the person ’s identity meets a preset identity condition or whether the person is facing the device, and whether the device is awakened for user purposes Device to prevent the user from unintentionally awakening the device from being awakened and causing the problem of false wakeup.
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Software Systems (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Computer Security & Cryptography (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- General Health & Medical Sciences (AREA)
- Human Computer Interaction (AREA)
- User Interface Of Digital Computer (AREA)
- Telephone Function (AREA)
- Telephonic Communication Services (AREA)
Abstract
本申请提供了一种音频处理方法,包括:采集输入数据;如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。采用该方法,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
Description
本申请涉及电子设备领域,更具体的说,是涉及一种音频处理方法和电子设备。
随着电子技术的发展,目前,很多设备支持语音功能,但是,由于使用固定的语音唤醒词,任何人说出该唤醒词都能够唤醒支持该唤醒词的设备,导致本不应唤醒的设备易被唤醒,出现误唤醒的问题。
发明内容
有鉴于此,本申请提供了一种音频处理方法,解决了现有技术中设备易出现误唤醒的问题。
为实现上述目的,本申请提供如下技术方案:
一种音频处理方法,应用于第一设备,所述方法包括:
采集输入数据;
如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
上述的方法,优选的,所述满足第一条件的输入数据用于切换预设应用的状态为预设运行状态,则所述以满足第一条件的方式响应所述输入数据之后,还包括:
采集控制数据,以使得处于预设运行状态的预设应用响应所述控制数据。
上述的方法,优选的,所述第一设备以第一方式输出多媒体内容时,则响应所述输入数据包括:
以所述第一方式输出响应数据。
上述的方法,优选的,所述输出多媒体内容时,采集输入数据之后,还包括:
判断所述输入数据是否满足第一条件;基于所述输入数据满足所述第一条件,判断所述输入数据是否满足第二条件;
或
判断所述输入数据是否满足第二条件;基于所述输入数据满足所述第二条件,判断所述 输入数据是否满足第一条件。
上述的方法,优选的,判断所述输入数据是否满足第二条件,包括:
判断是否接收到第二设备反馈的第一信息;
基于接收到所述第一信息,判断所述输入数据是否满足第二条件;
其中,所述第一信息包括下列中的至少一种:
所述第二设备采集到所述输入数据;或
所述第二设备采集到所述输入数据的质量;或
所述第二设备执行响应所述输入数据的操作。
上述的方法,优选的,所述输入数据是语音音频,则判断所述输入数据是否满足第二条件,包括:
判断所述语音音频是否与预设的声纹信息匹配,所述预设的声纹信息为预设唤醒人的声纹信息;
基于所述语音音频与预设的声纹信息的匹配,确定所述输入数据满足第二条件;否则,确定所述输入数据不满足第二条件。
上述的方法,优选的,所述输入数据包括图像和音频,则判断所述输入数据是否满足第二条件,包括:
分析判断所述图像是否满足预设条件;
基于所述图像满足所述预设条件,确定所述输入数据满足所述第二条件;否则,所述输入数据不满足所述第二条件;
其中,所述图像满足预设条件包括下列中的至少一种:
识别得到的所述图像中人物身份满足预设身份条件;或
识别得到的所述图像中的人物朝向所述第一设备。
一种电子设备,包括:
采集模块,用于采集输入数据;
判断模块,用于判断所述输入数据是否满足第一条件,以及所述输入数据是否满足第二条件;
处理模块,用于如果所述输入数据满足第一条件且满足第二条件,以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据。
一种电子设备,包括:
处理器,用于接收采集的输入数据,如果所述输入数据满足第一条件且满足第二条件, 以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据;
存储器,用于存储所述第一条件和第二条件。
上述的电子设备,优选的,还包括:
音频采集器,用于采集语音音频;
则,所述存储器中还存储有预设的声纹信息;
所述处理器具体用于判断所述语音音频是否与预设的声纹信息匹配;
或者,
还包括:
音频采集器,用于采集语音音频;
图像采集模组,用于采集图像采集区域的图像;
则,所述存储器中还存储有预设条件;
所述处理器具体用于分析判断所述语音音频是否满足第一条件,以及判断所述图像是否满足预设条件。
经由上述的技术方案可知,与现有技术相比,本申请提供了一种音频处理方法,包括:采集输入数据;如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。采用该方法,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
为了更清楚地说明本申请实施例或现有技术中的技术方案,下面将对实施例或现有技术描述中所需要使用的附图作简单地介绍,显而易见地,下面描述中的附图仅仅是本申请的实施例,对于本领域普通技术人员来讲,在不付出创造性劳动的前提下,还可以根据提供的附图获得其他的附图。
图1为本申请提供的一种音频处理方法实施例1的流程图;
图2为本申请提供的一种音频处理方法实施例2的流程图;
图3为本申请提供的一种音频处理方法实施例3的流程图;
图4为本申请提供的一种音频处理方法实施例3中显示内容示意图;
图5为本申请提供的一种音频处理方法实施例4的流程图;
图6为本申请提供的一种音频处理方法实施例5的流程图;
图7为本申请提供的一种音频处理方法实施例5中具体示例示意图;
图8为本申请提供的一种音频处理方法实施例6的流程图;
图9为本申请提供的一种音频处理方法实施例7的流程图;
图10为本申请提供的一种音频处理方法实施例7中具体示例示意图;
图11为本申请提供的一种电子设备实施例1的结构示意图;
图12为本申请提供的一种电子设备实施例2的结构示意图;
图13为本申请提供的一种电子设备实施例3的结构示意图;
图14为本申请提供的一种电子设备实施例4的结构示意图。
下面将结合本申请实施例中的附图,对本申请实施例中的技术方案进行清楚、完整地描述,显然,所描述的实施例仅仅是本申请一部分实施例,而不是全部的实施例。基于本申请中的实施例,本领域普通技术人员在没有做出创造性劳动前提下所获得的所有其他实施例,都属于本申请保护的范围。
如图1所示的,为本申请提供的一种音频处理方法实施例1的流程图,该方法应用于一电子设备,本申请中,该电子设备作为第一设备,该方法包括以下步骤:
步骤S101:采集输入数据;
其中,该输入数据为输入该第一设备的数据。
具体的,该输入数据可以为音频、视频、图像、其他设备传输来的数据等。
步骤S102:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
其中,该输入数据同时满足第一条件和第二条件时,才以满足该第一条件的方式响应该输入数据。
步骤S103:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,该输入数据满足第一条件但是不满足第二条件时,忽略该满足第一条件的输入数据,不对该输入数据进行响应。
作为一个具体示例,该输入数据是音频时,该第一条件是该音频中包含唤醒词,如该唤醒词为“嗨,语音助手”,而该唤醒词是用于唤醒该第一设备中语音助手的,那么,响应该输 入数据是唤醒该第一设备中的语音助手。
相应的,该第二条件是对该第一条件的补充,当该输入数据也满足第二条件时,才以满足该第一条件的方式响应该输入数据。
例如,即使该输入数据中包含有唤醒词“嗨,语音助手”,但是,由于该输入数据不满足第二条件,该第一设备也不响应该唤醒词,即不唤醒该第一设备中的语音助手。
需要说明的是,该第二条件可以是与该第一设备相关的其他条件,如发出音频的用户的音频条件、其他与该第一设备相关设备的反馈或者用户的行为等各个方面的条件,后续实施例中会针对该第二条件进行详细解释,本实施例中不做详述。
综上,本实施例提供的一种音频处理方法,包括:采集输入数据;如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。采用该方法,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
其中,该满足第一条件的输入数据用于切换预设应用的状态为预设运行状态。
如图2所示的,为本申请提供的一种音频处理方法实施例2的流程图,该方法包括以下步骤:
步骤S201:采集输入数据;
步骤S202:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
其中,步骤S201-202与实施例1中的步骤S101-102一致,本实施例中不做赘述。
步骤S203:采集控制数据,以使得处于预设运行状态的预设应用响应所述控制数据;
其中,该输入数据满足第一条件和第二条件,以满足该第一条件的方式响应该输入数据,实现该第一设备中的预设应用的状态切换为预设运行状态。
例如,该预设运行状态为正常运行状态,或者激活状态。
那么,该预设应用切换为预设运行状态后,继续采集输入的控制数据,该预设应用响应该数据。
作为一个具体示例,该预设应用为第一设备中的语音助手,该预设运行状态为激活状态,则该语音助手激活后,该第一设备继续采集输入的控制数据,如语音指令“打电话给李明”,则该语音助手响应该语音控制指令,执行控制第一设备中的电话软件进行“打电话给李明”的操作。又如,该继续采集输入的控制数据为语音控制指令“打开浏览器”,则该语音助手响 应该语音控制指令,执行控制第一设备中的浏览器软件打开的操作。
步骤S204:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S204与实施例1中的步骤S103一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,还包括:采集控制数据,以使得处于预设运行状态的预设应用响应所述控制数据。采用该方法,以满足该第一条件的方式响应该输入数据,实现将第一设备中的预设应用的状态切换为预设运行状态,并在后续过程中,继续采集输入的控制数据,并使得该预设应用响应该控制数据,保证该预设应用正常执行操作。
其中,该第一设备以第一方式输出多媒体内容。
如图3所示的,为本申请提供的一种音频处理方法实施例3的流程图,包括以下步骤:
步骤S301:采集输入数据;
其中,步骤S301与实施例1中的步骤S101一致,本实施例中不做赘述。
步骤S302:如果满足第一条件的所述输入数据满足第二条件,以所述第一方式输出响应数据;
需要说明的是,第一设备以影响多媒体内容输出的方式,输出响应数据,输出该响应数据会对该多媒体的内容输出产生干扰。
所以,该第二条件是用于判断是否该第一设备是否不需要响应该输入数据,第一设备需要响应该输入数据,则输入数据满足第二条件,否则,该输入数据不满足第二条件。
具体的,该第一设备中输出多媒体过程中,采集该输入数据,该多媒体内容的输出方式与该第一设备响应该输入数据的方式相应,都是第一方式。当该第一设备输出响应数据时,可能会对其输出多媒体内容产生影响,因此,需要确定该满足第一条件的输入数据也满足第二条件时,该第一设备输出响应数据,用户能够接收该响应。
例如,该第一设备通过屏幕显示内容(如视频或者图像)时,通过在该屏幕上显示一个提示框实现输出响应,该提示框占据了部分屏幕,遮挡部分屏幕中的原显示内容。
又如,该第一设备通过喇叭播放内容(如音频)时,通过播放音频“启动语音助手”实现输出响应,其与播放内容重叠。
如图4所示的为显示内容示意图,包括:显示界面401,该显示界面中显示图像,当设备对输入数据进行响应时,在显示界面上显示提示框402,该提示框中提示“启动语音助手。”。
步骤S303:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S303与实施例1中的步骤S103一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,所述第一设备以第一方式输出多媒体内容时,则响应所述输入数据包括:以所述第一方式输出响应数据。采用该方法,通过与设备输出多媒体内容相同的方式输出响应数据,保证用户能够了解该第一设备已经响应了该输入数据。
如图5所示的,为本申请提供的一种音频处理方法实施例4的流程图,包括以下步骤:
步骤S501:采集输入数据;
其中,步骤S501与实施例1中的步骤S101一致,本实施例中不做赘述。
步骤S502:判断所述输入数据是否满足第一条件;
步骤S503:基于所述输入数据满足所述第一条件,判断所述输入数据是否满足第二条件;
其中,先对该输入数据是否满足第一条件进行判断,如果该输入数据满足该第一条件,再判断其是否满足第二条件。
作为一个具体示例,该输入数据是音频,第一条件是该音频中包含唤醒词,那么判断该音频是否包含该唤醒词,如果包含,该输入数据满足该第一条件,而为了保证该第一设备是特定用户目的唤醒的设备,还需要根据情况对与该第一设备/用户相关的信息进行判断,即判断该输入数据是否满足第二条件,以防止不是特定用户唤醒设备或者用户目的唤醒的不是该第一设备,而导致出现误唤醒的问题。
需要说明的是,具体实施中,本申请对判断输入数据是否满足第一条件和第二条件的先后顺序不做限制,可以判断所述输入数据是否满足第一条件;基于所述输入数据满足所述第一条件,判断所述输入数据是否满足第二条件;也可以判断所述输入数据是否满足第二条件;基于所述输入数据满足所述第二条件,判断所述输入数据是否满足第一条件。
步骤S504:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
步骤S505:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S504-505与实施例1中的步骤S102-103一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,先判断该输入数据是否满足第一条件,基于所述输入数据满足所述第一条件,判断所述输入数据是否满足第二条件。采用该方法,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
如图6所示的,为本申请提供的一种音频处理方法实施例5的流程图,包括以下步骤:
步骤S601:采集输入数据;
步骤S602:判断所述输入数据是否满足第一条件;
其中,步骤S601-602与实施例4中的步骤S501-502一致,本实施例中不做赘述。
步骤S603:基于所述输入数据满足所述第一条件,判断是否接收到第二设备反馈的第一信息;
其中,该第二设备与该第一设备组成联网系统,该联网系统中的数据共享。
例如,该第一设备与第二设备可以处于同一环境中,二者能够对该环境中的相同内容进行采集,如采集相同的输入数据,而联网系统中的设备采集到输入数据后可以将其采集的相关信息和/或对该输入数据的信息反馈给其他设备。
具体的,所述第一信息包括下列中的至少一种:
所述第二设备采集到所述输入数据;或
所述第二设备采集到所述输入数据的质量;或
所述第二设备执行响应所述输入数据的操作。
需要说明的是,当用户在说出唤醒词时,由于联网系统中的各个设备处于与用户不同的相对位置,其能够采集的音频(输入数额)的质量不同,越靠近用户,输入数据的质量(如清晰度/强度等)越好,采集输入数据的速度越快,响应速度也越快。
例如,该联网系统为家电系统时,该系统中可以包括手机、平板电脑、电视、冰箱、空调等各种电子设备。
步骤S604:基于接收到所述第一信息,判断所述输入数据是否满足第二条件;
其中,该第一设备接收到第二设备反馈的第一信息后,可以结合该第一信息判断自己采集的输入数据是否满足第二条件。
具体的,该第一信息是第二设备采集到输入数据时,该第一设备采集该输入数据晚于该第二设备,则可以分析得到该第二设备更加靠近该用户,该第二设备是用户目的唤醒的设备,那么,该输入数据不满足第二条件;当该第一设备未接收到该第一信息时,该第一设备是最早采集该输入数据的,则可以分析得到该第一设备最靠近用户,该第一设备就是用户目的唤醒的设备,那么,该输入数据满足第二条件。
具体的,该第一信息是第二设备采集到输入数据的质量,以强度为例,该第二设备采集到输入设备的强度为9,而该第一设备采集到输入数据的强度为4,那么可以分析得到该第二设备更加靠近该用户,该第二设备是用户目的唤醒的设备,那么,该输入数据不满足第二条 件;该第二设备采集到输入设备的强度为2,而该第一设备采集到输入数据的强度为8,那么可以分析得到该第一设备更加靠近该用户,该第一设备是用户目的唤醒的设备,那么,该输入数据满足第二条件。
具体的,该第一信息是第二设备执行响应所述输入数据的操作时,由于该第一设备采集到该第一信息还未进行响应之前,该第二设备已经响应了该输入数据,那么可知,该第二设备是用户目的唤醒的设备,那么,该输入数据不满足第二条件;如果未接收到该第一信息时,那么可知,该第一设备采集速度较快,该第一设备是用户目的唤醒的设备,那么,该输入数据满足第二条件。
如图7所示的一个具体示例示意图,该输入数据为音频,该音频为用户701说出特定唤醒词“嗨,语音助手”时产生的,而该联网系统中的手机702、平板电脑703和电视704中的语音助手都能够被该特定唤醒词唤醒。该手机、平板电脑和电视都能够对环境中的音频进行采集,三者与用户的距离从近到远分别为手机、电视、平板电脑。
例如,任意一个设备采集完成后,将其采集动作反馈给其他设备。三个设备采集速度从快到慢是:手机、电视、平板电脑,手机采集到音频后,将其采集到音频的信息反馈给电视和平板电脑,该手机中未接收到其他设备反馈的信息,那么该手机对该音频进行响应,唤醒其语音助手;而电视和平板电脑得到该反馈的信息可知,已有手机在其之前采集到音频,那么,该电视和平板电脑不对该采集的音频进行响应。
又如,任意一个设备采集完成后,可以将其采集的音频质量反馈给其他设备。三个设备采集强度/清晰度从大到小是:手机、电视、平板电脑,各个设备采集到音频后,将其采集到音频的质量反馈给其他设备,由于手机中的音频质量最好,那么该手机对该音频进行响应,唤醒其语音助手;而电视和平板电脑得到该反馈的信息可知,有其他设备音频质量优于自己,那么,该电视和平板电脑不对其采集的音频进行响应。
又如,任意一个设备采集完成后,对该音频进行响应,并将响应操作的信息反馈给其他设备。三个设备响应的速度从快到慢是:手机、电视、平板电脑。手机采集到音频后,其对该音频进行响应,唤醒其语音助手,并将该响应操作的信息反馈给电视、平板电脑。而电视和平板电脑得到该反馈的信息可知,手机已经响应了该音频,那么,该电视和平板电脑不对其采集的音频进行响应。
步骤S605:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
步骤S606:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S605-606与实施例5中的步骤S504-505一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,判断所述输入数据是否满足第二条件,包括:判断是否接收到第二设备反馈的第一信息;基于接收到所述第一信息,判断所述输入数据是否满足第二条件;其中,所述第一信息包括下列中的至少一种:所述第二设备采集到所述输入数据;或所述第二设备采集到所述输入数据的质量;或所述第二设备执行响应所述输入数据的操作。采用该方法,通过第一设备与第二设备之间进行针对其采集输入数据或者输入数据质量或者是否响应该输入数据进行信息反馈,各个设备之间数据共享,使得根据该共享的信息确定哪个设备是用户目的唤醒的设备,保证了唤醒用户想要唤醒的设备被唤醒,防止出现误唤醒的问题。
其中,该输入数据是语音音频。
如图8所示的,为本申请提供的一种音频处理方法实施例6的流程图,包括以下步骤:
步骤S801:采集输入数据;
步骤S802:判断所述输入数据是否满足第一条件;
其中,步骤S801-802与实施例4中的步骤S501-502一致,本实施例中不做赘述。
步骤S803:基于所述输入数据满足所述第一条件,判断所述语音音频是否与预设的声纹信息匹配,所述预设的声纹信息为预设唤醒人的声纹信息;
基于所述语音音频与预设的声纹信息匹配,所述输入数据满足第二条件;
否则,所述输入数据不满足第二条件。
需要说明的是,不同人具有不同的声纹信息,根据声纹信息能够对发出声音的人身份进行判断。
其中,该输入数据满足第一条件,即语音音频中包含有特定的唤醒词。
为防止非特定用户唤醒第一设备,则还需要对发出该语音音频的人的身份进行判断,具体通过声纹信息进行判断。
具体的,该第一设备中预设声纹信息,该预设的声纹信息为预设唤醒人的声纹信息。判断该语音音频与预设的声纹信息是否匹配,如果二者匹配,该发出语音音频的人就是预设唤醒人,其具有唤醒该第一设备语音助手的权限;如果二者不匹配,那么发出语音音频的人就不是预设唤醒人,其没有唤醒该第一设备语音助手的权限。
作为一个具体示例,用户A使用手机,用户B使用平板电脑,两个设备中语音助手的唤醒词均是“嗨,语音助手”,那么,当A和B在同一环境中时,B说出语音“嗨,语音助手”,如果手机中未设置该第二条件,则该手机采集到输入数据后,就会对该唤醒词进行响应,唤 醒语音助手,而该手机的用户A并未想要唤醒语音助手,这会导致A的体验较差。而该手机中设置该第二条件,根据声纹信息即可判定该语音不是自己的用户A发出,则会忽略该唤醒词,不唤醒语音助手。
步骤S804:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
步骤S805:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S804-805与实施例5中的步骤S504-505一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,所述输入数据是语音音频,则判断所述输入数据是否满足第二条件,包括:判断所述语音音频是否与预设的声纹信息匹配,所述预设的声纹信息为预设唤醒人的声纹信息;基于所述语音音频与预设的声纹信息匹配,所述输入数据满足第二条件;否则,所述输入数据不满足第二条件。采用该方法,通过对语音音频与预设声纹信息进行匹配判断,确定发出该语音音频的人是否为预设唤醒人,防止出现其他人唤醒设备导致误唤醒的问题。
其中,该输入数据包括图像和音频。
如图9所示的,为本申请提供的一种音频处理方法实施例7的流程图,包括以下步骤:
步骤S901:采集输入数据;
步骤S902:判断所述输入数据是否满足第一条件;
其中,步骤S901-902与实施例4中的步骤S501-502一致,本实施例中不做赘述。
步骤S903:基于所述输入数据满足所述第一条件,分析判断所述图像是否满足预设条件;
基于所述图像满足预设条件,所述输入数据满足第二条件;
否则,所述输入数据不满足第二条件;
其中,图像满足预设条件包括下列中的至少一种:
识别得到的所述图像中人物身份满足预设身份条件;或
识别得到的所述图像中的人物朝向所述第一设备。
其中,该输入数据包括音频和图像,该第一设备能够同时对音频和图像进行采集。
具体实施中,可以对该输入数据中的音频进行判断是否满足第一条件,对该输入数据中的图像是否满足预设条件进行判断。
需要说明的是,当用户说出该唤醒词时,第一设备在采集得到音频信息同时,也对图像采集区域进行图像采集,采集的图像中包含用户的影像。
具体的,对该图像进行分析,得到该图像中人物的相关信息,如特征、姿势等。
具体的,该人物特征可以包括人脸特征、动态特征等,而根据该人物特征可以分析得到该人物的身份是否为满足预设身份条件的特定唤醒人,该特定唤醒人能够唤醒设备。
具体实施中,该第一设备中可以预设该特定唤醒人的人物特征的相关信息。该特定唤醒人可以为授权的用户,只有该授权的用户能够使用该第一设备。
具体的,该图像中人物的相关信息是人脸特征时,则对图像进行识别,得到图像中人物的脸部特征,根据该脸部特征确定该人物是否为能够唤醒设备的特定唤醒人,该脸部特征与特定唤醒人的脸部特征匹配时,该输入数据满足第二条件,否则不满足。
具体的,该图像中人物的相关信息是动态特征时,则对连续的几帧图像进行识别,得到图像中人物动态特征(如走路、挥手等动作),根据该动态特征确定该人物是否为能够唤醒设备的特定唤醒人,该动态特征与特定唤醒人的动态特征匹配时,该输入数据满足第二条件,否则不满足。
作为一个具体示例,第一设备中设置有授权的用户的人物特征。当有授权的用户说出唤醒词时,该第一设备根据采集的图像分析得到说出唤醒词的人的与预设的人物特征匹配,就可响应该唤醒词,唤醒第一没备的语音助手。当有非授权的用户说出唤醒词时,该第一设备根据采集的图像分析得到说出唤醒词的人的与预设的人物特征不匹配,就可忽略该唤醒词,不唤醒第一设备的语音助手。
具体的,该图像中人物的姿势是该人物面向该第一设备时,则对图像进行识别,得到图像中人物是否面对该第一设备,如果人物面对该第一设备,该输入数据满足第二条件,否则不满足。
具体应用中,当用户想要控制/操作某一设备时,会面向该设备,而当用户不面向该设备时,则可认为用户不想控制/操作该设备。
当用户周围有多个设备,可以根据自己的需求面对想要控制/操作的设备,所以,根据用户是否面对设备即可确定其是否想要操作/操作该设备。
如图10的为一个具体示例示意图,用户1001周围具有手机1002、平板电脑1003和电视1004,用户面对该手机1002。用户1001说出特定唤醒词“嗨,语音助手”时产生音频,手机1002、平板电脑1003和电视1004中的语音助手都能够被该特定唤醒词唤醒,该手机1002、平板电脑1003和电视1004对其图像采集区域进行图像采集,并对采集的图像进行分析,该平板电脑1003分析其采集到的图像,得到结果为用户面对该平板电脑,该输入数据满足第二条件,则平板电脑响应该唤醒词,唤醒语音助手。而手机和电视分析到的结果为用户均未面对自己,该输入数据不满足第二条件,则不响应该唤醒词。
步骤S904:如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;
步骤S905:如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
其中,步骤S904-905与实施例5中的步骤S504-505一致,本实施例中不做赘述。
综上,本实施例提供的一种音频处理方法中,该所述输入数据包括图像和音频,则判断所述输入数据是否满足第二条件,包括:分析判断所述图像是否满足预设条件;基于所述图像满足预设条件,所述输入数据满足第二条件;否则,所述输入数据不满足第二条件;其中,图像满足预设条件包括下列中的至少一种:识别得到的所述图像中人物身份满足预设身份条件;或识别得到的所述图像中的人物朝向所述第一设备。采用该方法,通过对图像中的人物进行分析,判断人物身份是否满足预设身份条件或者该确定该人物是否朝向设备,确定本设备是否为用户目的唤醒的设备,防止出现用户非目的唤醒的设备被唤醒导致误唤醒的问题。
与上述本申请提供的一种音频处理方法实施例相对应的,本申请还提供了应用该音频处理方法的电子设备实施例。
如图11所示的为本申请提供的一种电子设备实施例1的结构示意图,该电子设备中具有音频采集功能,该电子设备包括以下结构:采集模块1101、判断模块1102和处理模块1103;
其中,采集模块1101,用于采集输入数据;
其中,判断模块1102,用于判断所述输入数据是否满足第一条件,以及所述输入数据是否满足第二条件;
其中,处理模块1103,用于如果所述输入数据满足第一条件且满足第二条件,以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据。
其中,该输入数据包括音频时,该采集模块具体可以采用麦克风等具有音频采集功能的装置;该输入数据包括音频和图像时,该采集模块可以包括音频采集的装置(如麦克风)和图像采集的装置(如摄像头)。
综上,本实施例提供的一种电子设备中,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
根据本公开的实施例的模块、子模块、单元、子单元中的任意多个、或其中任意多个的至少部分功能可以在一个模块中实现。根据本公开实施例的模块、子模块、单元、子单元中 的任意一个或多个可以被拆分成多个模块来实现。根据本公开实施例的模块、子模块、单元、子单元中的任意一个或多个可以至少被部分地实现为硬件电路,例如现场可编程门阵列(FPGA)、可编程逻辑阵列(PLA)、片上系统、基板上的系统、封装上的系统、专用集成电路(ASIC),或可以通过对电路进行集成或封装的任何其他的合理方式的硬件或固件来实现,或以软件、硬件以及固件三种实现方式中任意一种或以其中任意几种的适当组合来实现。或者,根据本公开实施例的模块、子模块、单元、子单元中的一个或多个可以至少被部分地实现为计算机程序模块,当该计算机程序模块被运行时,可以执行相应的功能。
例如,采集模块1101、判断模块1102和处理模块1103中的任意多个可以合并在一个模块中实现,或者其中的任意一个模块可以被拆分成多个模块。或者,这些模块中的一个或多个模块的至少部分功能可以与其他模块的至少部分功能相结合,并在一个模块中实现。根据本公开的实施例,采集模块1101、判断模块1102和处理模块1103中的至少一个可以至少被部分地实现为硬件电路,例如现场可编程门阵列(FPGA)、可编程逻辑阵列(PLA)、片上系统、基板上的系统、封装上的系统、专用集成电路(ASIC),或可以通过对电路进行集成或封装的任何其他的合理方式等硬件或固件来实现,或以软件、硬件以及固件三种实现方式中任意一种或以其中任意几种的适当组合来实现。或者,采集模块1101、判断模块1102和处理模块1103中的至少一个可以至少被部分地实现为计算机程序模块,当该计算机程序模块被运行时,可以执行相应的功能。
如图12所示的为本申请提供的一种电子设备实施例2的结构示意图,该电子设备包括以下结构:处理器1201和存储器1202;
其中,处理器1201,用于接收采集的输入数据,如果所述输入数据满足第一条件且满足第二条件,以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据;
其中,存储器1202,用于存储所述第一条件和第二条件。
具体实施中,该处理器可以采用具有数据处理能力的芯片结构,如CPU(central processing unit,中央处理器)等。
具体实施中,该第一设备以第一方式输出多媒体内容。该第一方式可以为屏幕显示方式或者音频播放等。
具体的,该第一方式为屏幕显示方式时,该第一设备中还包含有显示屏,以实现对该多媒体内容进行显示,且将该响应该输入数据的响应数据在该显示屏幕中相应显示。
具体的,该第一方式为音频播放方式时,该第一设备中还包含有音频播放器,如喇叭, 以实现对该多媒体内容进行音频播放,且将该响应该输入数据的响应数据在该喇叭进行播放。
综上,本实施例提供的一种电子设备中,通过判断满足第一条件的输入数据是否满足第二条件,确定是否以第一条件的方式响应该输入数据,对输入数据进行了两个条件的判断,判断准确度较高,防止误唤醒。
其中,该输入数据是语音音频。
如图13所示的为本申请提供的一种电子设备实施例3的结构示意图,该电子设备包括以下结构:处理器1301、存储器1302和音频采集器1303;
其中,该处理器1301、存储器1302的结构功能与实施例2中的相应结构功能一致,本实施例中不做赘述。
其中,该音频采集器1303,用于采集语音音频;
则,所述存储器中还存储有预设的声纹信息;
所述处理器具体用于判断所述语音音频是否与预设的声纹信息匹配。
具体实施中,该音频采集器可以采用麦克风等具有音频采集功能的设备结构。
综上,本实施例提供的一种电子设备中,所述输入数据是语音音频,通过对语音音频与预设声纹信息进行匹配判断,确定发出该语音音频的人是否为预设唤醒人,防止出现其他人唤醒设备导致误唤醒的问题。
其中,该输入数据是语音音频和图像。
如图14所示的为本申请提供的一种电子设备实施例4的结构示意图,该电子设备包括以下结构:处理器1401、存储器1402、音频采集器1403和图像采集模组1404;
其中,该处理器1401、存储器1402的结构功能与实施例2中的相应结构功能一致,本实施例中不做赘述。
其中,该音频采集器1403,用于采集语音音频;
其中,图像采集模组1404,用于采集图像采集区域的图像,该图中包含有人物影像。
则,所述存储器中还存储有预设条件
所述处理器具体用于分析判断所述语音音频是否满足第一条件,以及判断所述图像是否满足预设条件。
其中,图像满足预设条件包括下列中的至少一种:
识别得到的所述图像中人物身份满足预设身份条件;或
识别得到的所述图像中的人物朝向所述第一设备。
综上,本实施例提供的一种电子设备中,通过对图像中的人物进行分析,判断人物身份是否满足预设身份条件或者该确定该人物是否朝向设备,确定本设备是否为用户目的唤醒的设备,防止出现用户非目的唤醒的设备被唤醒导致误唤醒的问题。
本说明书中各个实施例采用递进的方式描述,每个实施例重点说明的都是与其他实施例的不同之处,各个实施例之间相同相似部分互相参见即可。对于实施例提供的装置而言,由于其与实施例提供的方法相对应,所以描述的比较简单,相关之处参见方法部分说明即可。
对所提供的实施例的上述说明,使本领域专业技术人员能够实现或使用本发明。对这些实施例的多种修改对本领域的专业技术人员来说将是显而易见的,本文中所定义的一般原理可以在不脱离本发明的精神或范围的情况下,在其它实施例中实现。因此,本发明将不会被限制于本文所示的这些实施例,而是要符合与本文所提供的原理和新颖特点相一致的最宽的范围。
Claims (10)
- 一种音频处理方法,应用于第一设备,所述方法包括:采集输入数据;如果满足第一条件的所述输入数据满足第二条件,以满足第一条件的方式响应所述输入数据;如果满足所述第一条件的所述输入数据不满足所述第二条件,忽略所述满足第一条件的所述输入数据。
- 根据权利要求1所述的方法,所述满足第一条件的输入数据用于切换预设应用的状态为预设运行状态,则所述以满足第一条件的方式响应所述输入数据之后,还包括:采集控制数据,以使得处于预设运行状态的预设应用响应所述控制数据。
- 根据权利要求1所述的方法,所述第一设备以第一方式输出多媒体内容时,则响应所述输入数据包括:以所述第一方式输出响应数据。
- 根据权利要求1所述的方法,所述第一电子设备输出多媒体内容时,采集输入数据之后,还包括:判断所述输入数据是否满足所述第一条件;基于所述输入数据满足所述第一条件,判断所述输入数据是否满足第二条件;或判断所述输入数据是否满足所述第二条件;基于所述输入数据满足所述第二条件,判断所述输入数据是否满足所述第一条件。
- 根据权利要求4所述的方法,判断所述输入数据是否满足所述第二条件,包括:判断是否接收到第二设备反馈的第一信息;基于接收到所述第一信息,判断所述输入数据是否满足第二条件;其中,所述第一信息包括下列中的至少一种:所述第二设备采集到所述输入数据;或所述第二设备采集到所述输入数据的质量;或所述第二设备执行响应所述输入数据的操作。
- 根据权利要求4所述的方法,所述输入数据是语音音频,则判断所述输入数据是否满足所述第二条件,包括:判断所述语音音频是否与预设的声纹信息匹配,所述预设的声纹信息为预设唤醒人的声 纹信息;基于所述语音音频与预设的声纹信息的匹配,确定所述输入数据满足第二条件;否则,确定所述输入数据不满足第二条件。
- 根据权利要求4所述的方法,所述输入数据包括图像和音频,则判断所述输入数据是否满足第二条件,包括:分析判断所述图像是否满足预设条件;基于所述图像满足预设条件,确定所述输入数据满足第二条件;否则,确定所述输入数据不满足第二条件;其中,所述图像满足预设条件包括下列中的至少一种:识别得到的所述图像中人物身份满足预设身份条件;或识别得到的所述图像中的人物朝向所述第一设备。
- 一种电子设备,包括:采集模块,用于采集输入数据;判断模块,用于判断所述输入数据是否满足第一条件,以及所述输入数据是否满足第二条件;处理模块,用于如果所述输入数据满足第一条件且满足第二条件,以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据。
- 一种电子设备,包括:处理器,用于接收采集的输入数据,如果所述输入数据满足第一条件且满足第二条件,以满足第一条件的方式响应所述输入数据;以及如果所述输入数据满足第一条件且不满足第二条件,忽略所述满足第一条件的所述输入数据;存储器,用于存储所述第一条件和第二条件。
- 根据权利要求9所述的电子设备,还包括:音频采集器,用于采集语音音频;则,所述存储器中还存储有预设的声纹信息;所述处理器具体用于判断所述语音音频是否与预设的声纹信息匹配;或者,还包括:音频采集器,用于采集语音音频;图像采集模组,用于采集图像采集区域的图像;则,所述存储器中还存储有预设条件;所述处理器具体用于分析判断所述语音音频是否满足第一条件,以及判断所述图像是否满足预设条件。
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201810699716.3A CN109032554B (zh) | 2018-06-29 | 2018-06-29 | 一种音频处理方法和电子设备 |
CN201810699716.3 | 2018-06-29 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2020001172A1 true WO2020001172A1 (zh) | 2020-01-02 |
Family
ID=65522106
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2019/086193 WO2020001172A1 (zh) | 2018-06-29 | 2019-05-09 | 一种音频处理方法和电子设备 |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN109032554B (zh) |
WO (1) | WO2020001172A1 (zh) |
Families Citing this family (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109032554B (zh) * | 2018-06-29 | 2021-11-16 | 联想(北京)有限公司 | 一种音频处理方法和电子设备 |
CN109378000B (zh) * | 2018-12-19 | 2022-06-07 | 科大讯飞股份有限公司 | 语音唤醒方法、装置、系统、设备、服务器及存储介质 |
CN109979463B (zh) * | 2019-03-31 | 2022-04-22 | 联想(北京)有限公司 | 一种处理方法和电子设备 |
CN113407089A (zh) * | 2019-08-26 | 2021-09-17 | 华为技术有限公司 | 一种语音控制的分屏显示方法及电子设备 |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN105575395A (zh) * | 2014-10-14 | 2016-05-11 | 中兴通讯股份有限公司 | 语音唤醒方法及装置、终端及其处理方法 |
CN105898065A (zh) * | 2016-05-16 | 2016-08-24 | 深圳天珑无线科技有限公司 | 智能终端及其控制方法 |
US20170154519A1 (en) * | 2014-05-28 | 2017-06-01 | Xi'an Zhongxing New Software Co., Ltd. | Alarming Method, Terminal, and Storage Medium |
CN107181869A (zh) * | 2017-06-06 | 2017-09-19 | 上海传英信息技术有限公司 | 移动终端和利用语音识别开启移动终端应用的方法 |
CN109032554A (zh) * | 2018-06-29 | 2018-12-18 | 联想(北京)有限公司 | 一种音频处理方法和电子设备 |
Family Cites Families (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20140180457A1 (en) * | 2012-12-26 | 2014-06-26 | Anshuman Thakur | Electronic device to align audio flow |
CN105283836B (zh) * | 2013-07-11 | 2019-06-04 | 英特尔公司 | 用于设备唤醒的设备、方法、装置及计算机可读存储介质 |
US9713192B2 (en) * | 2015-03-27 | 2017-07-18 | Intel Corporation | Device and method for processing audio data |
KR20170027999A (ko) * | 2015-09-03 | 2017-03-13 | 삼성전자주식회사 | 사용자 단말 장치, 시스템 및 그 제어 방법 |
CN106815507A (zh) * | 2015-11-30 | 2017-06-09 | 中兴通讯股份有限公司 | 语音唤醒实现方法、装置及终端 |
CN105869637B (zh) * | 2016-05-26 | 2019-10-15 | 百度在线网络技术(北京)有限公司 | 语音唤醒方法和装置 |
CN107622652B (zh) * | 2016-07-15 | 2020-10-02 | 青岛海尔智能技术研发有限公司 | 家电系统的语音控制方法与家电控制系统 |
CN107749894A (zh) * | 2017-11-09 | 2018-03-02 | 吴章义 | 一种安全、简单、智能的物联网系统 |
CN107919119A (zh) * | 2017-11-16 | 2018-04-17 | 百度在线网络技术(北京)有限公司 | 多设备交互协同的方法、装置、设备及计算机可读介质 |
-
2018
- 2018-06-29 CN CN201810699716.3A patent/CN109032554B/zh active Active
-
2019
- 2019-05-09 WO PCT/CN2019/086193 patent/WO2020001172A1/zh active Application Filing
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170154519A1 (en) * | 2014-05-28 | 2017-06-01 | Xi'an Zhongxing New Software Co., Ltd. | Alarming Method, Terminal, and Storage Medium |
CN105575395A (zh) * | 2014-10-14 | 2016-05-11 | 中兴通讯股份有限公司 | 语音唤醒方法及装置、终端及其处理方法 |
CN105898065A (zh) * | 2016-05-16 | 2016-08-24 | 深圳天珑无线科技有限公司 | 智能终端及其控制方法 |
CN107181869A (zh) * | 2017-06-06 | 2017-09-19 | 上海传英信息技术有限公司 | 移动终端和利用语音识别开启移动终端应用的方法 |
CN109032554A (zh) * | 2018-06-29 | 2018-12-18 | 联想(北京)有限公司 | 一种音频处理方法和电子设备 |
Also Published As
Publication number | Publication date |
---|---|
CN109032554A (zh) | 2018-12-18 |
CN109032554B (zh) | 2021-11-16 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2020001172A1 (zh) | 一种音频处理方法和电子设备 | |
US20220343919A1 (en) | Voice-Controlled Split-Screen Display Method and Electronic Device | |
US10165391B2 (en) | Companion application for activity cooperation | |
WO2017016298A1 (zh) | 广告投放系统、装置及方法 | |
WO2019184946A1 (zh) | 人脸识别控制方法及移动终端 | |
CN108564943B (zh) | 语音交互方法及系统 | |
CN107870674B (zh) | 一种程序启动方法和移动终端 | |
CN112739507B (zh) | 一种交互沟通实现方法、设备和存储介质 | |
US20220405375A1 (en) | User identity verification method and electronic device | |
WO2022052791A1 (zh) | 一种多媒体流的播放方法和电子设备 | |
WO2019223493A1 (zh) | 对象识别方法及移动终端 | |
WO2020238451A1 (zh) | 终端控制方法和终端 | |
EP4199488A1 (en) | Voice interaction method and electronic device | |
US11930240B2 (en) | Media content recording with sensor data | |
WO2019129264A1 (zh) | 界面的显示方法和移动终端 | |
CN107809674A (zh) | 一种基于视频的用户反应获取、处理方法、终端及服务器 | |
KR102405307B1 (ko) | 전자 장치, 그 제어 방법 및 컴퓨터 판독가능 기록 매체 | |
US20210250433A1 (en) | Pressure-sensitive vibration processing method and apparatus, mobile terminal and electronic device | |
CN110597840B (zh) | 基于区块链的伴侣关系建立方法、装置、设备及存储介质 | |
US11443060B2 (en) | Note taking at a device based on whether the user is a known user, unknown user, or owner | |
CN113763932B (zh) | 语音处理方法、装置、计算机设备及存储介质 | |
CN116437155A (zh) | 一种直播互动方法、装置、计算机设备和存储介质 | |
CN116189674A (zh) | 一种语音控制方法及显示设备 | |
WO2021218730A1 (zh) | 信息显示方法及电子设备 | |
WO2024169473A1 (zh) | 一种多模式播放控制方法、装置、电子设备和存储介质 |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 19826816 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 300321) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 19826816 Country of ref document: EP Kind code of ref document: A1 |