EP3846162A1 - Smart audio device, calling method for audio device, electronic device and computer readable medium - Google Patents
Smart audio device, calling method for audio device, electronic device and computer readable medium Download PDFInfo
- Publication number
- EP3846162A1 EP3846162A1 EP20216951.2A EP20216951A EP3846162A1 EP 3846162 A1 EP3846162 A1 EP 3846162A1 EP 20216951 A EP20216951 A EP 20216951A EP 3846162 A1 EP3846162 A1 EP 3846162A1
- Authority
- EP
- European Patent Office
- Prior art keywords
- voice
- chip
- audio device
- calling
- communication
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Withdrawn
Links
- 238000000034 method Methods 0.000 title claims abstract description 40
- 238000004891 communication Methods 0.000 claims description 55
- 230000003993 interaction Effects 0.000 claims description 12
- 230000004044 response Effects 0.000 claims description 11
- 238000011946 reduction process Methods 0.000 claims description 10
- 238000004590 computer program Methods 0.000 claims description 6
- 230000008569 process Effects 0.000 claims description 5
- 230000002093 peripheral effect Effects 0.000 claims description 3
- 238000010586 diagram Methods 0.000 description 5
- 230000006870 function Effects 0.000 description 4
- 230000009467 reduction Effects 0.000 description 4
- 238000005516 engineering process Methods 0.000 description 3
- 230000002035 prolonged effect Effects 0.000 description 3
- 230000005236 sound signal Effects 0.000 description 3
- 230000005540 biological transmission Effects 0.000 description 2
- 238000013500 data storage Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000001514 detection method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
- 230000001360 synchronised effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F3/00—Input arrangements for transferring data to be processed into a form capable of being handled by the computer; Output arrangements for transferring data from processing unit to output unit, e.g. interface arrangements
- G06F3/16—Sound input; Sound output
- G06F3/167—Audio in a user interface, e.g. using voice commands for navigating, audio feedback
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/28—Constructional details of speech recognition systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/22—Interactive procedures; Man-machine interfaces
- G10L17/24—Interactive procedures; Man-machine interfaces the user being prompted to utter a password or a predefined phrase
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L21/0216—Noise filtering characterised by the method used for estimating noise
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L2015/088—Word spotting
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
- G10L2021/02082—Noise filtering the noise being echo, reverberation of the speech
-
- H—ELECTRICITY
- H04—ELECTRIC COMMUNICATION TECHNIQUE
- H04M—TELEPHONIC COMMUNICATION
- H04M3/00—Automatic or semi-automatic exchanges
- H04M3/002—Applications of echo suppressors or cancellers in telephonic connections
Definitions
- the embodiments of the present disclosure relate to the field of computer technology, and in particular, to a smart audio device, a calling method for audio device, an electronic device, and a computer readable medium.
- Embodiments of the present disclosure provide a smart audio device, a calling method for audio device, an electronic device and a computer readable medium.
- an embodiment of the present disclosure provides a smart audio device, including: a front chip provided therein with a plurality of voice algorithm modules; and a main control chip signally connected with the front chip and configured to call the voice algorithm modules in the front chip according to a user request in a multi-thread mode.
- the voice algorithm modules include: a voice wake-up algorithm module configured to wake up the main control chip according to the user request; and a front-end signal processing module configured to perform a noise reduction process on a front-end signal.
- the front-end signal processing module includes: a voice recognition front-end signal processing module configured to perform the noise reduction process on a front-end voice signal and obtain a voice recognition engine signal; and a communication front-end signal processing module configured to perform the noise reduction process on a front-end communication signal and obtain a communication application signal.
- the main control chip includes: a voice recognition module configured to recognize a voice signal; and/or a communication application module configured to communicate with other communication modules.
- the smart audio device further includes a communication interface configured to transmit signals between the front chip and the main control chip.
- the communication interface includes at least one of a general-purpose I/O interface, an Inter-Integrated Circuit bus interface, and a Serial Peripheral Interface.
- the smart audio device further includes a microphone signally connected with the front chip and configured to receive voice information and transmit the voice information to the front chip.
- the front chip further includes an echo cancellation module signally connected with the voice algorithm modules and configured to perform an echo cancellation process on the received voice information.
- an embodiment of the present disclosure provides a calling method for audio device, including: calling different voice algorithms in a front chip according to a user request in a multi-thread mode.
- the step of calling the different voice algorithms in the front chip according to the user request in the multi-thread mode includes: receiving a wake-up request, which is the user request; entering a working mode in response to the wake-up request; and sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to a voice recognition front-end signal processing algorithm mode.
- the calling method further includes: receiving a communication request, which is the user request; and sending an instruction of calling a communication front-end signal processing algorithm in the front chip in response to the communication request.
- the calling method further includes: sending an instruction of calling the voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to the voice recognition front-end signal processing algorithm mode.
- in response to no voice interaction sending an instruction of calling a voice wake-up algorithm in the front chip, so as to switch the front chip to a voice wake-up algorithm mode.
- an embodiment of the present disclosure provides an electronic device, including: one or more processors; a memory having one or more programs stored thereon, when the one or more programs are executed by the one or more processors, the one or more processors perform the above calling method for audio device; and one or more I/O interfaces connected between the processor and the memory and configured to enable information interaction between the processor and the memory.
- an embodiment of the present disclosure provides a computer readable medium having a computer program stored thereon.
- the computer program is executed by a processor, the above calling method for audio device is performed.
- the plurality of voice algorithm modules are arranged in the front chip, and the main control chip calls the voice algorithm modules in the front chip according to the user request in the multi-thread mode, so that a voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of the smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged.
- the voice algorithm modules in the front chip in the multi-thread mode a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- a single-thread working mode of an audio front chip is accepted, that is, the audio front chip works in a single mode after being powered on, and cannot adjust the working mode to optimize the working frequency according to actual applications (such as voice recognition, communication application, and voice wake-up) of a back end, which causes high overall power consumption of an intelligent hardware device.
- a voice wake-up algorithm cannot be adjusted according to the communication application of the back end, resulting in poor user experience.
- an embodiment of the present disclosure provides a smart audio device.
- Fig. 1 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure.
- the smart audio device includes: a front chip 10 provided therein with a plurality of voice algorithm modules, each corresponding to one voice algorithm mode; and a main control chip 20 signally connected with the front chip 10 and configured to call the voice algorithm modules in the front chip according to a user request in a multi-thread mode.
- the user request includes, but is not limited to, a voice wake-up request, a voice recognition request, and a communication request.
- the front chip 10 and the main control chip 20 are signally connected through a communication interface 30, which includes, but is not limited to, at least one of a general-purpose I/O interface, an Inter-Integrated Circuit (I2C) bus interface, and a Serial Peripheral Interface (SPI).
- a communication interface 30 includes, but is not limited to, at least one of a general-purpose I/O interface, an Inter-Integrated Circuit (I2C) bus interface, and a Serial Peripheral Interface (SPI).
- I2C Inter-Integrated Circuit
- SPI Serial Peripheral Interface
- the front chip 10 is in a voice wake-up mode, and the main control chip 20 is in a standby mode.
- Fig. 2 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure.
- the voice algorithm modules arranged in the front chip 10 include a voice wake-up algorithm module 11, a voice recognition front-end signal processing module 12, and a communication front-end signal processing module 13.
- the voice wake-up algorithm module 11 is configured to wake up the main control chip 20 according to a user request. For example, when the main control chip 20 is in the standby mode, the voice wake-up algorithm module 11 wakes up the main control chip 20 according to a wake-up request from a user to put the main control chip 20 into a working mode.
- the voice recognition front-end signal processing module 12 is configured to perform a noise reduction process on a front-end voice signal and obtain a voice recognition engine signal.
- the communication front-end signal processing module 13 is configured to perform the noise reduction process on a front-end communication signal and obtain a communication application signal.
- the front-end signal processing modules 12 and 13 are configured to perform the noise reduction process on a front-end signal, which is a signal received by the front chip 10.
- the front-end signal may be the front-end voice signal, the front-end communication signal, etc.
- the front-end signal processing modules 12 and 13 perform the noise reduction process on the front-end signal, and then transmits the processed front-end signal to the main control chip 20 via a communication interface.
- the main control chip 20 includes a voice recognition module 21 configured to recognize a voice signal, and a communication application module 22.
- the voice recognition module 21 is turned on in response to the voice recognition engine signal and recognizes an input voice signal.
- the voice recognition module 21 may recognize a letter, a word, a short sentence or a long sentence to determine the user request.
- the communication application module 22 is configured to communicate with other communication modules.
- the communication application module 22 is turned on in response to the communication application signal and communicates with the other communication modules, which may be other smart audio devices, and modules having a communication function, such as a smart phone, IPAD and so on.
- the main control chip 20 may send the voice signal to a cloud server having a stronger computing capability to let the cloud server to recognize the voice signal.
- the smart audio device further includes a microphone 40 signally connected with the front chip 10 and configured to receive voice information and transmit the voice information to the front chip 10.
- the microphone 40 is implemented in a form of a microphone array, that is, a plurality of microphones are connected to the front chip 10.
- an analog-to-digital (AD) converter 50 is disposed between the microphone 40 and the front chip 10 to convert an analog signal received by the microphone 40 into a digital signal.
- the front chip 10 further includes a Voice Activity Detection (VAD) module 14 and an echo cancellation module 15.
- VAD Voice Activity Detection
- the VAD module 14 is signally connected between the AD converter 50 and the voice wake-up algorithm module 11, and is configured to recognize and eliminate a long silent period from an audio signal flow.
- the echo cancellation module 15 is signally connected between the AD converter 50 and the front-end signal processing modules, and is configured to perform an echo cancellation process on the received voice information.
- the smart audio device further includes a power amplifier module 60 and a speaker 70.
- the power amplifier module 60 is signally connected with the main control chip 20, and is configured to amplify a voice signal output by the main control chip 20 and input the amplified voice signal into the speaker 70.
- the speaker 70 is configured to play the amplified voice signal.
- the front chip 10 collects the voice signal from the power amplifier module 60, and takes the voice signal as a reference signal for echo cancellation.
- the plurality of voice algorithm modules are arranged in the front chip, and the main control chip calls the voice algorithm modules in the front chip according to the user request in the multi-thread mode, so that a voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of the smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged.
- the voice algorithm modules in the front chip in the multi-thread mode a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- an embodiment of the present disclosure provides a calling method for audio device, which calls a voice algorithm module in a front chip according to a user request in a multi-thread mode.
- Fig. 3 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure.
- the calling method for audio device includes the following steps 301 to 303.
- step 301 receiving a wake-up request.
- a front chip after a smart audio device is powered on, a front chip is in a voice wake-up mode, and a VAD module and a voice wake-up algorithm module may obtain a wake-up message.
- the front chip sends the wake-up request to a main control chip.
- the front chip may send the wake-up request to the main control chip via a general-purpose I/O interface.
- step 302 entering a working mode in response to the wake-up request.
- the main control chip is in a standby mode after the smart audio device is powered on, and enters the working mode when receiving the wave-up request.
- step 303 sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip.
- the instruction of calling the voice recognition front-end signal processing algorithm may be sent to the front chip after the main control chip enters the working mode, so as to switch the front chip from a voice wake-up algorithm mode to a voice recognition front-end signal processing algorithm mode.
- the voice recognition front-end signal processing algorithm may perform an echo cancellation process on a voice signal, the voice information after the echo cancellation is sent to a voice recognition module of the main control chip, and is recognized by the voice recognition module.
- the main control chip sends the instruction of calling the voice recognition front-end signal processing algorithm to the front chip via an I2C bus interface or an SPI interface.
- Fig. 4 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure.
- the calling method for audio device includes the following steps 401 to 405.
- step 401 receiving a wake-up request.
- step 402 entering a working mode in response to the wake-up request.
- step 403 sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip.
- steps 401 to 403 are the same as the steps 301 to 303, and thus will not be repeated here.
- step 404 receiving a communication request.
- the main control chip After receiving the voice information sent from the front chip, the main control chip recognizes the voice information, and may perform voice interaction.
- step 405 sending an instruction of calling a communication front-end signal processing algorithm in the front chip in response to the communication request.
- the main control chip When the main control chip recognizes that a voice command is a communication command, the main control chip may send the instruction of calling the communication front-end signal processing algorithm to the front chip via an I2C bus interface or an SPI interface. After receiving the instruction of calling the communication front-end signal processing algorithm, the front chip switches the algorithm mode thereof to a communication front-end signal processing algorithm.
- the front chip performs an echo cancellation process on the voice information with the communication front-end signal processing algorithm, and sends the voice information after the echo cancellation to a communication application module of the main control chip.
- the communication with other communication modules is performed through the communication application module.
- the calling method for audio device further includes the following steps 406 to 407.
- step 406 sending an instruction of calling the voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to a voice recognition front-end signal processing algorithm mode.
- the main control chip sends the instruction of calling the voice recognition front-end signal processing algorithm to the front chip again via the I2C interface or the SPI interface, so as to switch the front chip back to the voice recognition front-end signal processing algorithm mode to perform voice interaction.
- step 407 when there is no voice interaction, sending an instruction of calling a voice wake-up algorithm in the front chip.
- the main control chip when there is no voice interaction within a preset period, sends the instruction of calling the voice wake-up algorithm to the front chip via the I2C interface or the SPI interface, so as to put the front chip into the voice wake-up mode again. Meanwhile, the main control chip is switched to the standby mode having a low power consumption.
- the main control chip calls the different voice algorithms in the front chip according to the user request; and since the main control chip calls the voice algorithms in the front chip in the multi-thread mode, the voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of a smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged.
- the voice algorithm modules in the front chip in the multi-thread mode a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- an embodiment of the present disclosure provides an electronic device, including: one or more processors 501; a memory 502 having one or more programs stored thereon, when the one or more programs are executed by the one or more processors, the one or more processors perform the above calling method for audio device; and one or more I/O interfaces 503 connected between the processor and the memory and configured to enable information interaction between the processor and the memory.
- the processor 501 is a device having a data processing capability, and includes, but is not limited to, a central processing unit (CPU), etc.
- the memory 502 is a device having a data storage capability, and includes, but is not limited to, a random access memory (RAM, more specifically, such as a synchronous dynamic RAM (SDRAM), a double data rate SDRAM (DDR SDRAM), etc.), a read-only memory (ROM), an electrically erasable programmable read-only memory (EEPROM), and a flash memory (FLASH).
- the I/O interface (read/write interface) 503 is connected between the processor 501 and the memory 502, enables the information interaction between the processor 501 and the memory 502, and includes, but is not limited to, a data bus.
- the processor 501, the memory 502, and the I/O interface 503 are connected to each other through a bus 504, so as to be further connected to the other components of the electronic device.
- an embodiment of the present disclosure provides a computer readable medium having a computer program stored thereon.
- the computer program is executed by a processor, the above calling method for audio device is performed.
- the functional modules/units in all or some of the steps, systems, and devices in the method disclosed above may be implemented as software, firmware, hardware, or suitable combinations thereof. If implemented as hardware, the division between the functional modules/units stated above is not necessarily corresponding to the division of physical components; for example, one physical component may have a plurality of functions, or one function or step may be performed through cooperation of several physical components. Some or all of the physical components may be implemented as software executed by a processor, such as a central processing unit, a digital signal processor, or a microprocessor, or may be implemented as hardware, or may be implemented as an integrated circuit, such as an application specific integrated circuit.
- a processor such as a central processing unit, a digital signal processor, or a microprocessor
- Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media).
- computer storage media includes volatile/nonvolatile and removable/non-removable media used in any method or technology for storing information (such as computer-readable instructions, data structures, program modules and other data).
- the computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory techniques, CD-ROM, digital versatile disk (DVD) or other optical discs, magnetic cassette, magnetic tape, magnetic disk or other magnetic storage devices, or any other media which can be used to store the desired information and can be accessed by a computer.
- the communication media generally include computer-readable instructions, data structures, program modules or other data in a modulated data signal, such as a carrier wave or other transmission mechanism, and may include any information delivery media.
Landscapes
- Engineering & Computer Science (AREA)
- Multimedia (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Physics & Mathematics (AREA)
- Health & Medical Sciences (AREA)
- Acoustics & Sound (AREA)
- Computational Linguistics (AREA)
- Quality & Reliability (AREA)
- Signal Processing (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Telephone Function (AREA)
Abstract
Description
- The embodiments of the present disclosure relate to the field of computer technology, and in particular, to a smart audio device, a calling method for audio device, an electronic device, and a computer readable medium.
- With the development of voice recognition technology, more and more intelligent hardware devices have a voice interaction function, and the quality of voice interaction is closely related to a noise reduction process of front-end audio signals. In general, there are two noise reduction methods: one achieves noise reduction with a front-end audio digital signal processing chip, such as a digital signal processor (DSP) or an application specific integrated circuit (ASIC), and the other one achieves the noise reduction with a software on a main control chip. However, both of the two noise reduction methods have defects to varying degrees.
- Embodiments of the present disclosure provide a smart audio device, a calling method for audio device, an electronic device and a computer readable medium.
- In a first aspect, an embodiment of the present disclosure provides a smart audio device, including: a front chip provided therein with a plurality of voice algorithm modules; and a main control chip signally connected with the front chip and configured to call the voice algorithm modules in the front chip according to a user request in a multi-thread mode.
- In some embodiments, the voice algorithm modules include: a voice wake-up algorithm module configured to wake up the main control chip according to the user request; and a front-end signal processing module configured to perform a noise reduction process on a front-end signal.
- In some embodiments, the front-end signal processing module includes: a voice recognition front-end signal processing module configured to perform the noise reduction process on a front-end voice signal and obtain a voice recognition engine signal; and a communication front-end signal processing module configured to perform the noise reduction process on a front-end communication signal and obtain a communication application signal.
- In some embodiments, the main control chip includes: a voice recognition module configured to recognize a voice signal; and/or a communication application module configured to communicate with other communication modules.
- In some embodiments, the smart audio device further includes a communication interface configured to transmit signals between the front chip and the main control chip.
- In some embodiments, the communication interface includes at least one of a general-purpose I/O interface, an Inter-Integrated Circuit bus interface, and a Serial Peripheral Interface.
- In some embodiments, the smart audio device further includes a microphone signally connected with the front chip and configured to receive voice information and transmit the voice information to the front chip.
- In some embodiments, the front chip further includes an echo cancellation module signally connected with the voice algorithm modules and configured to perform an echo cancellation process on the received voice information.
- In a second aspect, an embodiment of the present disclosure provides a calling method for audio device, including: calling different voice algorithms in a front chip according to a user request in a multi-thread mode.
- In some embodiments, the step of calling the different voice algorithms in the front chip according to the user request in the multi-thread mode includes: receiving a wake-up request, which is the user request; entering a working mode in response to the wake-up request; and sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to a voice recognition front-end signal processing algorithm mode.
- In some embodiments, after the step of sending the instruction of calling the voice recognition front-end signal processing algorithm, the calling method further includes: receiving a communication request, which is the user request; and sending an instruction of calling a communication front-end signal processing algorithm in the front chip in response to the communication request.
- In some embodiments, after the communication is finished, the calling method further includes: sending an instruction of calling the voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to the voice recognition front-end signal processing algorithm mode.
- In some embodiments, in response to no voice interaction, sending an instruction of calling a voice wake-up algorithm in the front chip, so as to switch the front chip to a voice wake-up algorithm mode.
- In a third aspect, an embodiment of the present disclosure provides an electronic device, including: one or more processors; a memory having one or more programs stored thereon, when the one or more programs are executed by the one or more processors, the one or more processors perform the above calling method for audio device; and one or more I/O interfaces connected between the processor and the memory and configured to enable information interaction between the processor and the memory.
- In a fourth aspect, an embodiment of the present disclosure provides a computer readable medium having a computer program stored thereon. When the computer program is executed by a processor, the above calling method for audio device is performed.
- In the smart audio device provided by the embodiments of the present disclosure, the plurality of voice algorithm modules are arranged in the front chip, and the main control chip calls the voice algorithm modules in the front chip according to the user request in the multi-thread mode, so that a voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of the smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged. In addition, by calling the voice algorithm modules in the front chip in the multi-thread mode, a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- The accompanying drawings are intended to provide further understanding of the embodiments of the present disclosure, and are incorporated in and constitute a part of the Specification. The drawings, together with the embodiments of the present disclosure, are intended to explain the present disclosure, rather than limiting the present disclosure. With the detailed description of exemplary embodiments with reference to the drawings, the above and other features and advantages will become more apparent to those skilled in the art. In the drawings:
-
Fig. 1 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure; -
Fig. 2 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure; -
Fig. 3 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure; -
Fig. 4 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure; and -
Fig. 5 is a block diagram of an electronic device according to an embodiment of the present disclosure. - In order to enable those skilled in the art to better understand the technical solutions of the present disclosure, a smart audio device, a calling method for audio device, an electronic device and a computer readable medium provided by the present disclosure are described in detail below with reference to the accompanying drawings.
- Although exemplary embodiments will be described in more detail below with reference to the drawings, the exemplary embodiments can be embodied in various forms and should not be interpreted as limitation to the present disclosure. Rather, these embodiments are provided for facilitating thorough and complete understanding of the present disclosure, and enabling those skilled in the art to fully understand the scope of the present disclosure.
- The embodiments and the features thereof in the present disclosure may be combined with one another if no conflict is incurred.
- As used herein, the term "and/or" includes any and all combinations of one or more of the associated listed items.
- The terms used herein are intended to describe specific embodiments, rather than limiting the present disclosure. Unless expressly indicated otherwise, the singular terms "a", "an" and "the" used herein are intended to include plural forms as well. It should also be understood that the terms "include" and/or "comprise", when used herein, specify the presence of the stated features, integers, steps, operations, elements, and/or components, but do not preclude the presence or addition of one or more other features, integers, steps, operations, elements, components, and/or the groups thereof.
- Unless defined otherwise, all the terms (including technical and scientific terms) used herein have the same meaning as commonly understood by those skilled in the art. Unless expressly defined herein, the terms defined in generally used dictionaries should be interpreted as having the meanings given in the context of the related art and the present disclosure, and should not be interpreted as having idealized or overly formal meanings.
- The embodiments of the present disclosure are provided in view of that a single-thread working mode of an audio front chip is accepted, that is, the audio front chip works in a single mode after being powered on, and cannot adjust the working mode to optimize the working frequency according to actual applications (such as voice recognition, communication application, and voice wake-up) of a back end, which causes high overall power consumption of an intelligent hardware device. In addition, due to the single-thread working mode, a voice wake-up algorithm cannot be adjusted according to the communication application of the back end, resulting in poor user experience.
- In a first aspect, an embodiment of the present disclosure provides a smart audio device.
Fig. 1 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure. With reference toFig. 1 , the smart audio device includes: afront chip 10 provided therein with a plurality of voice algorithm modules, each corresponding to one voice algorithm mode; and amain control chip 20 signally connected with thefront chip 10 and configured to call the voice algorithm modules in the front chip according to a user request in a multi-thread mode. - The user request includes, but is not limited to, a voice wake-up request, a voice recognition request, and a communication request.
- In some embodiments, the
front chip 10 and themain control chip 20 are signally connected through acommunication interface 30, which includes, but is not limited to, at least one of a general-purpose I/O interface, an Inter-Integrated Circuit (I2C) bus interface, and a Serial Peripheral Interface (SPI). In some embodiments, with the aid of thecommunication interface 30, thefront chip 10 and themain control chip 20 can achieve transmission of control signals, reset signals, wake-up terminal signals, and the audio signals processed by algorithms. - In some embodiments, after the smart audio device is powered on, the
front chip 10 is in a voice wake-up mode, and themain control chip 20 is in a standby mode. -
Fig. 2 is a schematic structural diagram of a smart audio device according to an embodiment of the present disclosure. As shown inFig. 2 , the voice algorithm modules arranged in thefront chip 10 include a voice wake-up algorithm module 11, a voice recognition front-endsignal processing module 12, and a communication front-endsignal processing module 13. The voice wake-up algorithm module 11 is configured to wake up themain control chip 20 according to a user request. For example, when themain control chip 20 is in the standby mode, the voice wake-up algorithm module 11 wakes up themain control chip 20 according to a wake-up request from a user to put themain control chip 20 into a working mode. The voice recognition front-endsignal processing module 12 is configured to perform a noise reduction process on a front-end voice signal and obtain a voice recognition engine signal. The communication front-endsignal processing module 13 is configured to perform the noise reduction process on a front-end communication signal and obtain a communication application signal. - The front-end
signal processing modules front chip 10. In some embodiments, the front-end signal may be the front-end voice signal, the front-end communication signal, etc. The front-endsignal processing modules main control chip 20 via a communication interface. - The
main control chip 20 includes avoice recognition module 21 configured to recognize a voice signal, and a communication application module 22. Thevoice recognition module 21 is turned on in response to the voice recognition engine signal and recognizes an input voice signal. For example, thevoice recognition module 21 may recognize a letter, a word, a short sentence or a long sentence to determine the user request. The communication application module 22 is configured to communicate with other communication modules. The communication application module 22 is turned on in response to the communication application signal and communicates with the other communication modules, which may be other smart audio devices, and modules having a communication function, such as a smart phone, IPAD and so on. - It should be noted that, when the processing capabilities of the
main control chip 20 cannot meet user's needs, themain control chip 20 may send the voice signal to a cloud server having a stronger computing capability to let the cloud server to recognize the voice signal. - The smart audio device further includes a
microphone 40 signally connected with thefront chip 10 and configured to receive voice information and transmit the voice information to thefront chip 10. In some embodiments, themicrophone 40 is implemented in a form of a microphone array, that is, a plurality of microphones are connected to thefront chip 10. In some embodiments, an analog-to-digital (AD)converter 50 is disposed between themicrophone 40 and thefront chip 10 to convert an analog signal received by themicrophone 40 into a digital signal. - In some embodiments, the
front chip 10 further includes a Voice Activity Detection (VAD)module 14 and anecho cancellation module 15. TheVAD module 14 is signally connected between theAD converter 50 and the voice wake-upalgorithm module 11, and is configured to recognize and eliminate a long silent period from an audio signal flow. Theecho cancellation module 15 is signally connected between theAD converter 50 and the front-end signal processing modules, and is configured to perform an echo cancellation process on the received voice information. - In some embodiments, the smart audio device further includes a
power amplifier module 60 and aspeaker 70. Thepower amplifier module 60 is signally connected with themain control chip 20, and is configured to amplify a voice signal output by themain control chip 20 and input the amplified voice signal into thespeaker 70. Thespeaker 70 is configured to play the amplified voice signal. - In some embodiments, the
front chip 10 collects the voice signal from thepower amplifier module 60, and takes the voice signal as a reference signal for echo cancellation. - In the smart audio device provided by the embodiments of the present disclosure, the plurality of voice algorithm modules are arranged in the front chip, and the main control chip calls the voice algorithm modules in the front chip according to the user request in the multi-thread mode, so that a voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of the smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged. In addition, by calling the voice algorithm modules in the front chip in the multi-thread mode, a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- In a second aspect, an embodiment of the present disclosure provides a calling method for audio device, which calls a voice algorithm module in a front chip according to a user request in a multi-thread mode.
-
Fig. 3 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure. With reference toFig. 3 , the calling method for audio device includes the followingsteps 301 to 303. - At
step 301, receiving a wake-up request. - In some embodiments, after a smart audio device is powered on, a front chip is in a voice wake-up mode, and a VAD module and a voice wake-up algorithm module may obtain a wake-up message. When receiving the wake-up request, the front chip sends the wake-up request to a main control chip. For example, the front chip may send the wake-up request to the main control chip via a general-purpose I/O interface.
- At
step 302, entering a working mode in response to the wake-up request. - In some embodiments, the main control chip is in a standby mode after the smart audio device is powered on, and enters the working mode when receiving the wave-up request.
- At
step 303, sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip. - As the front chip operating in a multi-thread mode, the instruction of calling the voice recognition front-end signal processing algorithm may be sent to the front chip after the main control chip enters the working mode, so as to switch the front chip from a voice wake-up algorithm mode to a voice recognition front-end signal processing algorithm mode. The voice recognition front-end signal processing algorithm may perform an echo cancellation process on a voice signal, the voice information after the echo cancellation is sent to a voice recognition module of the main control chip, and is recognized by the voice recognition module.
- In some embodiments, the main control chip sends the instruction of calling the voice recognition front-end signal processing algorithm to the front chip via an I2C bus interface or an SPI interface.
-
Fig. 4 is a flowchart illustrating a calling method for audio device according to an embodiment of the present disclosure. With reference toFig. 4 , the calling method for audio device includes the followingsteps 401 to 405. - At
step 401, receiving a wake-up request. - At
step 402, entering a working mode in response to the wake-up request. - At
step 403, sending an instruction of calling a voice recognition front-end signal processing algorithm in the front chip. - It should be noted that the
steps 401 to 403 are the same as thesteps 301 to 303, and thus will not be repeated here. - At
step 404, receiving a communication request. - After receiving the voice information sent from the front chip, the main control chip recognizes the voice information, and may perform voice interaction.
- At
step 405, sending an instruction of calling a communication front-end signal processing algorithm in the front chip in response to the communication request. - When the main control chip recognizes that a voice command is a communication command, the main control chip may send the instruction of calling the communication front-end signal processing algorithm to the front chip via an I2C bus interface or an SPI interface. After receiving the instruction of calling the communication front-end signal processing algorithm, the front chip switches the algorithm mode thereof to a communication front-end signal processing algorithm.
- The front chip performs an echo cancellation process on the voice information with the communication front-end signal processing algorithm, and sends the voice information after the echo cancellation to a communication application module of the main control chip. The communication with other communication modules is performed through the communication application module.
- In some embodiments, the calling method for audio device further includes the following
steps 406 to 407. - At
step 406, sending an instruction of calling the voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to a voice recognition front-end signal processing algorithm mode. - After the communication is finished, the main control chip sends the instruction of calling the voice recognition front-end signal processing algorithm to the front chip again via the I2C interface or the SPI interface, so as to switch the front chip back to the voice recognition front-end signal processing algorithm mode to perform voice interaction.
- At
step 407, when there is no voice interaction, sending an instruction of calling a voice wake-up algorithm in the front chip. - In some embodiments, when there is no voice interaction within a preset period, the main control chip sends the instruction of calling the voice wake-up algorithm to the front chip via the I2C interface or the SPI interface, so as to put the front chip into the voice wake-up mode again. Meanwhile, the main control chip is switched to the standby mode having a low power consumption.
- With the calling method for audio device provided by the embodiments of the present disclosure, the main control chip calls the different voice algorithms in the front chip according to the user request; and since the main control chip calls the voice algorithms in the front chip in the multi-thread mode, the voice wake-up algorithm can be arranged in the front chip, and even a main control chip having low computation and low cost can meet application requirements, thereby reducing a cost of main control chip, and further reducing a total cost of a smart audio device; moreover, the main control chip can adjust an operation frequency thereof as required, so that the power consumption of the main control chip can be reduced and the service life of the main control chip can be prolonged. In addition, by calling the voice algorithm modules in the front chip in the multi-thread mode, a flexibility of the smart audio device is enhanced, and different user's needs are met, thereby improving user experience.
- In a third aspect, with reference to
Fig. 5 , an embodiment of the present disclosure provides an electronic device, including: one ormore processors 501; amemory 502 having one or more programs stored thereon, when the one or more programs are executed by the one or more processors, the one or more processors perform the above calling method for audio device; and one or more I/O interfaces 503 connected between the processor and the memory and configured to enable information interaction between the processor and the memory. - The
processor 501 is a device having a data processing capability, and includes, but is not limited to, a central processing unit (CPU), etc. Thememory 502 is a device having a data storage capability, and includes, but is not limited to, a random access memory (RAM, more specifically, such as a synchronous dynamic RAM (SDRAM), a double data rate SDRAM (DDR SDRAM), etc.), a read-only memory (ROM), an electrically erasable programmable read-only memory (EEPROM), and a flash memory (FLASH). The I/O interface (read/write interface) 503 is connected between theprocessor 501 and thememory 502, enables the information interaction between theprocessor 501 and thememory 502, and includes, but is not limited to, a data bus. - In some embodiments, the
processor 501, thememory 502, and the I/O interface 503 are connected to each other through a bus 504, so as to be further connected to the other components of the electronic device. - In a fourth aspect, an embodiment of the present disclosure provides a computer readable medium having a computer program stored thereon. When the computer program is executed by a processor, the above calling method for audio device is performed.
- It should be understood by those skilled in the art that the functional modules/units in all or some of the steps, systems, and devices in the method disclosed above may be implemented as software, firmware, hardware, or suitable combinations thereof. If implemented as hardware, the division between the functional modules/units stated above is not necessarily corresponding to the division of physical components; for example, one physical component may have a plurality of functions, or one function or step may be performed through cooperation of several physical components. Some or all of the physical components may be implemented as software executed by a processor, such as a central processing unit, a digital signal processor, or a microprocessor, or may be implemented as hardware, or may be implemented as an integrated circuit, such as an application specific integrated circuit. Such software may be distributed on computer-readable media, which may include computer storage media (or non-transitory media) and communication media (or transitory media). As well known by those skilled in the art, the term "computer storage media" includes volatile/nonvolatile and removable/non-removable media used in any method or technology for storing information (such as computer-readable instructions, data structures, program modules and other data). The computer storage media include, but are not limited to, RAM, ROM, EEPROM, flash memory or other memory techniques, CD-ROM, digital versatile disk (DVD) or other optical discs, magnetic cassette, magnetic tape, magnetic disk or other magnetic storage devices, or any other media which can be used to store the desired information and can be accessed by a computer. In addition, it is well known by those skilled in the art that the communication media generally include computer-readable instructions, data structures, program modules or other data in a modulated data signal, such as a carrier wave or other transmission mechanism, and may include any information delivery media.
- It should be understood that both the exemplary embodiments and the specific terms disclosed in the present disclosure are for the purpose of illustration, rather than for limiting the present disclosure. It is obvious to those skilled in the art that the features, characteristics and/or elements described in connection with a particular embodiment may be used alone or in combination with the features, characteristics and/or elements described in connection with other embodiments in some examples, unless expressly indicated otherwise. Therefore, it should be understood by those skilled in the art that various changes in form and detail may be made without departing from the scope of the present disclosure as set forth in the appended claims.
Claims (14)
- A smart audio device, comprising:a front chip (10) provided therein with a plurality of voice algorithm modules; anda main control chip (20) signally connected with the front chip (10) and configured to call the voice algorithm modules in the front chip (10) according to a user request in a multi-thread mode.
- The smart audio device of claim 1, wherein the voice algorithm modules comprise:a voice wake-up algorithm module (11) configured to wake up the main control chip (20) according to the user request;a voice recognition front-end signal processing module (12) configured to perform a noise reduction process on a front-end voice signal and obtain a voice recognition engine signal; anda communication front-end signal processing module (13) configured to perform the noise reduction process on a front-end communication signal and obtain a communication application signal.
- The smart audio device of claim 1, wherein the main control chip (20) comprises:a voice recognition module (21) configured to recognize a voice signal; anda communication application module (22) configured to communicate with other communication modules.
- The smart audio device of claim 1, further comprising:
a communication interface (30) configured to transmit signals between the front chip (10) and the main control chip (20). - The smart audio device of claim 4, wherein the communication interface (30) comprises at least one of a general-purpose I/O interface, an Inter-Integrated Circuit bus interface, and a Serial Peripheral Interface.
- The smart audio device of claim 1, further comprising:
a microphone (40) signally connected with the front chip (10) and configured to receive voice information and transmit the voice information to the front chip (10). - The smart audio device of any one of claims 1 to 6, wherein the front chip (10) further comprises:
an echo cancellation module (50) signally connected with the voice algorithm modules and configured to perform an echo cancellation process on the received voice information. - A calling method for audio device, comprising:
calling different voice algorithms in a front chip according to a user request in a multi-thread mode. - The calling method of claim 8, wherein the step of calling the different voice algorithms in the front chip according to the user request in the multi-thread mode comprises:receiving (301, 401) a wake-up request, which is the user request;entering (302, 402) a working mode in response to the wake-up request; andsending (303, 403) an instruction of calling a voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to a voice recognition front-end signal processing algorithm mode.
- The calling method of claim 9, after the step of sending the instruction of calling the voice recognition front-end signal processing algorithm, the calling method further comprising:receiving (404) a communication request, which is the user request; andsending (405) an instruction of calling a communication front-end signal processing algorithm in the front chip in response to the communication request.
- The calling method of claim 10, after the communication is finished, the calling method further comprising:
sending (406) an instruction of calling the voice recognition front-end signal processing algorithm in the front chip, so as to switch the front chip to the voice recognition front-end signal processing algorithm mode. - The calling method of claim 11, in response to no voice interaction, sending (407) an instruction of calling a voice wake-up algorithm in the front chip, so as to switch the front chip to a voice wake-up algorithm mode.
- An electronic device, comprising:one or more processors (501);a memory (502) having one or more programs stored thereon, wherein when the one or more programs are executed by the one or more processors (501), the one or more processors (501) perform the calling method of any one of claims 8 to 12; andone or more I/O interfaces (503) connected between the processor (501) and the memory (502) and configured to enable information interaction between the processor (501) and the memory (502).
- A computer readable medium, having a computer program stored thereon, wherein when the computer program is executed by a processor, the calling method of any one of claims 8 to 12 is performed.
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010007936.2A CN111145752B (en) | 2020-01-03 | 2020-01-03 | Intelligent audio device, method, electronic device and computer readable medium |
Publications (1)
Publication Number | Publication Date |
---|---|
EP3846162A1 true EP3846162A1 (en) | 2021-07-07 |
Family
ID=70523616
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
EP20216951.2A Withdrawn EP3846162A1 (en) | 2020-01-03 | 2020-12-23 | Smart audio device, calling method for audio device, electronic device and computer readable medium |
Country Status (5)
Country | Link |
---|---|
US (1) | US20210210093A1 (en) |
EP (1) | EP3846162A1 (en) |
JP (1) | JP2021110945A (en) |
KR (1) | KR20210087880A (en) |
CN (1) | CN111145752B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115083404A (en) * | 2022-04-29 | 2022-09-20 | 中国第一汽车股份有限公司 | Vehicle-mounted voice noise reduction method and device, electronic equipment and storage medium |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020061012A1 (en) * | 1999-04-13 | 2002-05-23 | Thi James C. | Cable modem with voice processing capability |
US7162426B1 (en) * | 2000-10-02 | 2007-01-09 | Xybernaut Corporation | Computer motherboard architecture with integrated DSP for continuous and command and control speech processing |
WO2019015435A1 (en) * | 2017-07-19 | 2019-01-24 | 腾讯科技(深圳)有限公司 | Speech recognition method and apparatus, and storage medium |
US20190325888A1 (en) * | 2018-04-20 | 2019-10-24 | Baidu Online Network Technology (Beijing) Co., Ltd . | Speech recognition method, device, apparatus and computer-readable storage medium |
Family Cites Families (14)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN1174374C (en) * | 1999-06-30 | 2004-11-03 | 国际商业机器公司 | Method and device for parallelly having speech recognition, classification and segmentation of speaker |
KR101529918B1 (en) * | 2008-09-10 | 2015-06-18 | 엘지전자 주식회사 | Speech recognition apparatus using the multi-thread and methmod thereof |
JP5609182B2 (en) * | 2010-03-16 | 2014-10-22 | 日本電気株式会社 | Speech recognition apparatus, speech recognition method, and speech recognition program |
US9224404B2 (en) * | 2013-01-28 | 2015-12-29 | 2236008 Ontario Inc. | Dynamic audio processing parameters with automatic speech recognition |
CN105575395A (en) * | 2014-10-14 | 2016-05-11 | 中兴通讯股份有限公司 | Voice wake-up method and apparatus, terminal, and processing method thereof |
US10079015B1 (en) * | 2016-12-06 | 2018-09-18 | Amazon Technologies, Inc. | Multi-layer keyword detection |
US10540970B2 (en) * | 2017-12-12 | 2020-01-21 | Amazon Technologies, Inc. | Architectures and topologies for vehicle-based, voice-controlled devices |
US10838954B1 (en) * | 2017-12-14 | 2020-11-17 | Amazon Technologies, Inc. | Identifying user content |
US11315556B2 (en) * | 2019-02-08 | 2022-04-26 | Sonos, Inc. | Devices, systems, and methods for distributed voice processing by transmitting sound data associated with a wake word to an appropriate device for identification |
KR20200132613A (en) * | 2019-05-16 | 2020-11-25 | 삼성전자주식회사 | Method and apparatus for speech recognition with wake on voice |
KR20220008401A (en) * | 2019-06-07 | 2022-01-21 | 엘지전자 주식회사 | Speech Recognition Methods on Edge Computing Devices |
KR20210009596A (en) * | 2019-07-17 | 2021-01-27 | 엘지전자 주식회사 | Intelligent voice recognizing method, apparatus, and intelligent computing device |
WO2021045955A1 (en) * | 2019-09-04 | 2021-03-11 | Telepathy Labs, Inc. | Speech recognition systems and methods |
US11823659B2 (en) * | 2019-12-11 | 2023-11-21 | Amazon Technologies, Inc. | Speech recognition through disambiguation feedback |
-
2020
- 2020-01-03 CN CN202010007936.2A patent/CN111145752B/en active Active
- 2020-12-23 EP EP20216951.2A patent/EP3846162A1/en not_active Withdrawn
- 2020-12-23 US US17/132,112 patent/US20210210093A1/en not_active Abandoned
- 2020-12-28 JP JP2020218366A patent/JP2021110945A/en active Pending
- 2020-12-29 KR KR1020200185792A patent/KR20210087880A/en not_active Application Discontinuation
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20020061012A1 (en) * | 1999-04-13 | 2002-05-23 | Thi James C. | Cable modem with voice processing capability |
US7162426B1 (en) * | 2000-10-02 | 2007-01-09 | Xybernaut Corporation | Computer motherboard architecture with integrated DSP for continuous and command and control speech processing |
WO2019015435A1 (en) * | 2017-07-19 | 2019-01-24 | 腾讯科技(深圳)有限公司 | Speech recognition method and apparatus, and storage medium |
US20190325888A1 (en) * | 2018-04-20 | 2019-10-24 | Baidu Online Network Technology (Beijing) Co., Ltd . | Speech recognition method, device, apparatus and computer-readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN111145752A (en) | 2020-05-12 |
US20210210093A1 (en) | 2021-07-08 |
JP2021110945A (en) | 2021-08-02 |
CN111145752B (en) | 2022-08-02 |
KR20210087880A (en) | 2021-07-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
US11074924B2 (en) | Speech recognition method, device, apparatus and computer-readable storage medium | |
WO2020253715A1 (en) | Voice data processing method, device and system | |
JP6713035B2 (en) | Far-field voice function implementation method, equipment, system, storage medium, and program | |
CN110457078B (en) | Intelligent service method, device and equipment | |
US20190237070A1 (en) | Voice interaction method, device, apparatus and server | |
JP2019128939A (en) | Gesture based voice wakeup method, apparatus, arrangement and computer readable medium | |
JP2019128938A (en) | Lip reading based voice wakeup method, apparatus, arrangement and computer readable medium | |
WO2020038010A1 (en) | Intelligent device, voice wake-up method, voice wake-up apparatus, and storage medium | |
US20200219503A1 (en) | Method and apparatus for filtering out voice instruction | |
JP6785332B2 (en) | Bluetooth speaker Data processing method, device and bluetooth speaker | |
CN110211578B (en) | Sound box control method, device and equipment | |
US20200243085A1 (en) | Voice Processing Method, Apparatus and Device | |
EP3851952A2 (en) | Signal processing method, signal processing device, and electronic device | |
EP3846162A1 (en) | Smart audio device, calling method for audio device, electronic device and computer readable medium | |
CN112002320A (en) | Voice wake-up method and device, electronic equipment and storage medium | |
CN112581960A (en) | Voice wake-up method and device, electronic equipment and readable storage medium | |
JP2019191552A (en) | Cloud wake-up method and system, terminal and computer readable storage medium | |
CN112233676A (en) | Intelligent device awakening method and device, electronic device and storage medium | |
CN114724564A (en) | Voice processing method, device and system | |
CN113744732A (en) | Equipment wake-up related method and device and story machine | |
CN108806679A (en) | voice awakening method and device | |
US11366688B2 (en) | Do-not-disturb processing method and apparatus, and storage medium | |
CN111654782B (en) | Intelligent sound box and signal processing method | |
WO2023143544A1 (en) | Voice control method and apparatus, device, medium, and intelligent voice acquisition system | |
CN114400003B (en) | Control method and system for automatic switching microphone, electronic equipment and storage medium |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: REQUEST FOR EXAMINATION WAS MADE |
|
PUAI | Public reference made under article 153(3) epc to a published international application that has entered the european phase |
Free format text: ORIGINAL CODE: 0009012 |
|
17P | Request for examination filed |
Effective date: 20201223 |
|
AK | Designated contracting states |
Kind code of ref document: A1 Designated state(s): AL AT BE BG CH CY CZ DE DK EE ES FI FR GB GR HR HU IE IS IT LI LT LU LV MC MK MT NL NO PL PT RO RS SE SI SK SM TR |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: EXAMINATION IS IN PROGRESS |
|
17Q | First examination report despatched |
Effective date: 20221216 |
|
STAA | Information on the status of an ep patent application or granted ep patent |
Free format text: STATUS: THE APPLICATION HAS BEEN WITHDRAWN |
|
18W | Application withdrawn |
Effective date: 20230418 |