US20200013390A1 - Speech wakeup method, apparatus, and electronic device - Google Patents

Speech wakeup method, apparatus, and electronic device Download PDF

Info

Publication number
US20200013390A1
US20200013390A1 US16/571,468 US201916571468A US2020013390A1 US 20200013390 A1 US20200013390 A1 US 20200013390A1 US 201916571468 A US201916571468 A US 201916571468A US 2020013390 A1 US2020013390 A1 US 2020013390A1
Authority
US
United States
Prior art keywords
speech
speech wakeup
speech data
wakeup model
electronic device
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US16/571,468
Inventor
Zhiming Wang
Jun Zhou
Xiaolong Li
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Advanced New Technologies Co Ltd
Original Assignee
Alibaba Group Holding Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Alibaba Group Holding Ltd filed Critical Alibaba Group Holding Ltd
Assigned to ALIBABA GROUP HOLDING LIMITED reassignment ALIBABA GROUP HOLDING LIMITED ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: LI, XIAOLONG, ZHOU, JUN, WANG, ZHIMING
Publication of US20200013390A1 publication Critical patent/US20200013390A1/en
Priority to US16/774,422 priority Critical patent/US10748524B2/en
Assigned to ADVANTAGEOUS NEW TECHNOLOGIES CO., LTD. reassignment ADVANTAGEOUS NEW TECHNOLOGIES CO., LTD. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ALIBABA GROUP HOLDING LIMITED
Assigned to Advanced New Technologies Co., Ltd. reassignment Advanced New Technologies Co., Ltd. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ADVANTAGEOUS NEW TECHNOLOGIES CO., LTD.
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L15/16Speech classification or search using artificial neural networks
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/26Speech to text systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/02Feature extraction for speech recognition; Selection of recognition unit
    • G10L2015/025Phonemes, fenemes or fenones being the recognition units
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/06Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
    • G10L15/063Training
    • G10L2015/0631Creating reference templates; Clustering
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/08Speech classification or search
    • G10L2015/088Word spotting
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/22Procedures used during a speech recognition process, e.g. man-machine dialogue
    • G10L2015/223Execution procedure of a spoken command

Definitions

  • This specification relates to the field of computer technologies, and in particular, to a speech wakeup method, apparatus, and electronic device.
  • speech recognition is used in increasingly popular conversational assistants like Apple's Sin, Microsoft's Cortana, and Amazon's Alexa to enhance user experience and natural human-computer interaction level.
  • Keyword Spotting An important speech interaction technology is Keyword Spotting (KWS), which may also be generally referred to as speech wakeup or speech recognition. Based on the prior art, there is a need for a speech wakeup solution that may not rely on keyword-specific speech data.
  • a speech wakeup method, apparatus, and electronic device are provided in embodiments of this specification, for solving the following technical problem: there is a need for a speech wakeup solution that may not rely on keyword-specific speech data.
  • a speech wakeup method comprises: inputting speech data to a speech wakeup model trained with general speech data, and outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).
  • DNN Deep Neural Network
  • CTC Connectionist Temporal Classifier
  • an electronic device comprises: at least one processor; and a memory communicatively connected to the at least one processor; wherein the memory stores an instruction executable by the at least one processor, and the instruction is executed by the at least one processor to cause the electronic device to: input speech data to a speech wakeup model trained with general speech data, and output from the speech wakeup model a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • a non-transitory computer-readable storage medium has stored therein instructions that, when executed by a processor of an electronic device, cause the electronic device to perform a speech wakeup method, the method comprising: inputting speech data to a speech wakeup model trained with general speech data, and outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • the speech wakeup model may be trained with accessible general speech data, and then the trained speech wakeup model can be used for speech wakeup, which is conducive to improving the accuracy of speech wakeup.
  • FIG. 1 is a schematic diagram of an overall architecture for speech wakeup, according to an embodiment.
  • FIG. 2 is a flowchart of a speech wakeup method according to an embodiment.
  • FIG. 3 is a schematic diagram of a speech wakeup model according to an embodiment.
  • FIG. 4 is a schematic diagram of feature extraction performed by a feature extracting module according to an embodiment.
  • FIG. 5 is a schematic diagram of a Deep Neural Network (DNN) according to an embodiment.
  • DNN Deep Neural Network
  • FIG. 6 is a schematic diagram of a Connectionist Temporal Classifier (CTC) according to an embodiment.
  • CTC Connectionist Temporal Classifier
  • FIG. 7 is a schematic diagram of a speech wakeup apparatus according to an embodiment.
  • FIG. 8 is a schematic diagram of an electronic device according to an embodiment.
  • a speech wakeup method, apparatus, and electronic device are provided in embodiments of this specification.
  • a speech wakeup model including a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC) is trained with general speech data.
  • the trained speech wakeup model can be used for speech wakeup and support user-defined keywords triggered by speech wakeup.
  • the speech wakeup model can be used in low-power devices such as mobile phones and home appliances, because the DNN included in the speech wakeup model can be relatively not too complicated, and for example, can only have three or four layers with two or three hundred nodes in each layer.
  • the speech wakeup model can be referred to as CTC-KWS.
  • the DNN is a multi-layer perceptron, which has one or more hidden layers between an input layer and an output layer and can simulate complex nonlinear relationships.
  • the CTC is a classifier configured to perform a label tagging task, and does not require forced alignment between input and output.
  • FIG. 1 is a schematic diagram of an overall architecture 100 for speech wakeup, according to an embodiment.
  • the overall architecture 100 includes speech data 102 as a first part and a speech wakeup model 104 as a second part.
  • the speech wakeup model 104 includes a DNN 106 and a CTC 108 .
  • Speech wakeup can be implemented by inputting the speech data 102 to the speech wakeup model 104 for processing.
  • the embodiments described in detail in the following are based on the overall architecture 100 .
  • FIG. 2 is a flowchart of a speech wakeup method 200 according to an embodiment.
  • the method 200 may be performed by a server or a terminal, for example, through a model training program, a speech recognition program, a speech wakeup application, and so on the server or the terminal.
  • the server or terminal may be a mobile phone, a tablet computer, a smart wearable device, an automobile machine, a personal computer, a medium-sized computer, a computer cluster, and so on.
  • the method 200 may include the following steps.
  • speech data is input to a speech wakeup model trained with general speech data.
  • speech can be monitored by the server or the terminal to obtain the speech data.
  • a user can speak out a predetermined keyword to trigger the speech wakeup model to execute speech wakeup.
  • the speech wakeup model outputs a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • the general speech data described in step S 202 is less restricted and, thus, easily, accessible.
  • it may be a Urge Vocabulary Continuous Speech Recognition (LVCSR) corpus or the like.
  • the DNN included in the speech recognition model may predict a posterior probability distribution of a pronunciation phoneme sequence corresponding to input speech features.
  • the DNN can be followed by the CTC to give a confidence score corresponding to the predicted pronunciation phoneme sequence.
  • a result for determining whether to execute speech wakeup can be output based on the confidence score.
  • the speech wakeup model may be trained with the accessible general speech data, and further the trained speech wakeup model can be used for speech wakeup, which is conducive to improving the accuracy of speech wakeup.
  • the speech wakeup model also breaks through the restrictions of the keyword-specific speech data and supports user-defined triggered keywords. Therefore, it is more convenient and flexible in actual applications and conducive to improving user experience.
  • FIG. 3 is a schematic diagram of a speech wakeup model 300 , according to an embodiment.
  • the speech wakeup model 300 may include a feature extracting module 302 , a DNN 304 , and a CTC 306 in sequence.
  • step S 204 of the method 200 FIG.
  • the outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup may include: the feature extracting module 302 extracting acoustic features from the input speech data; inputting the acoustic features to the DNN 304 for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme; the DNN 304 inputting the class probability to the CTC 306 for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and the CTC 306 determining whether to execute wakeup according to the confidence score, and outputting a determination result.
  • the speech wakeup model 300 is described in detail below in further combination with FIG. 4 , FIG. 5 , and FIG. 6 .
  • FIG. 4 is a schematic diagram of feature extraction performed by a feature extracting module, such as the feature extracting module 302 ( FIG. 3 ), according to an embodiment.
  • a target label sequence corresponding thereto is a pronunciation phoneme sequence, which can be expressed as: “zhi1ma2kai1men2”, wherein the numbers represent tones.
  • tone phonemes are also taken into account as a modeling unit.
  • context-independent or context-dependent phonemes can all be taken into account, among which the latter is more numerous. However, in consideration of reducing subsequent computational burden of the DNN, only the context-independent phonemes may be considered, such as 72 context-independent phoneme units in Chinese, including a blank unit.
  • acoustic features can be extracted by the feature extracting module from the input speech data, which may include: extracting acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames may be multi-dimension log filter bank energies; stacking a plurality of adjacent acoustic feature frames respectively; taking the stacked acoustic feature frames respectively as acoustic features extracted from the general speech data; and further, the stacked acoustic feature frames can be used as inputs of the DNN respectively.
  • the log filter bank energies refer to energy signals extracted by a log filter bank, which can be expressed as a vector to facilitate model processing.
  • the multi-dimension in the foregoing represents multiple dimensions of the vector.
  • a specified length of a time window may be 25 milliseconds, each time window may move for 10 milliseconds, and the multi-dimension may be, for example, 40 dimensions.
  • milliseconds from 0 to 25 may be used as a window, and 40-dimension log filter bank energies are correspondingly extracted from the speech data to serve as a first acoustic feature frame; milliseconds from 10 to 35 can be used as a window, and 40-dimension log filter bank energies are correspondingly extracted from the speech data to serve as a second acoustic feature frame; and multiple acoustic feature frames can be extracted in the same way.
  • stacking a plurality of adjacent acoustic feature frames may allow more information from a context of a current frame, which is conducive to improving the accuracy of subsequent prediction results.
  • the current frame, the adjacent consecutive ten frames before the current frame, and the adjacent consecutive five frames after the current frame can be, for example, stacked to obtain a 640-dimension stacking feature for being inputted to the subsequent DNN.
  • cepstral mean and variance normalization can be carried out on the dimensions of the stacking feature, and then backward input can be carried out.
  • FIG. 5 is a schematic diagram of a DNN, such as the DNN 304 ( FIG. 3 ), according to an embodiment.
  • the DNN may include an input layer 502 , one or more hidden layers 504 , and an output layer 506 .
  • FIG. 5 various neurons, represented by circles, in the DNN are fully connected.
  • the acoustic features extracted by the feature extracting module 302 ( FIG. 3 ) are input to the DNN.
  • the DNN can describe a relationship between an input acoustic feature x 0 ⁇ R n 0 in the input layer 502 and a modeling unit j in the output layer 506 according to the following function mapping:
  • x i,i>0 ⁇ R n i is an output of a hidden layer
  • W i ⁇ R n i ⁇ n i ⁇ 1 and B i ⁇ RR n i are weights and offset parameters respectively, which may be predetermined based on, training or an application need
  • n i is the number of nodes on the i th layer
  • ⁇ W i , B i ⁇
  • T denotes transpose of a matrix
  • N is the number of the hidden layers
  • the formula III is a softmax function in the embodiment, representing an estimated posterior probability of a label unit j.
  • a Recurrent Neural Network can also be used in conjunction with the CTC.
  • RNN Recurrent Neural Network
  • the minimum computing and power consumption requirements of mobile devices can be more easily met by using the DNN in conjunction with the CTC.
  • the DNN with about hundreds of nodes in a hidden layer is more suitable.
  • FIG. 6 is a schematic diagram of a CTC, such as the CTC 306 ( FIG. 3 ), according to an embodiment.
  • the CTC is configured for sequential label tasks. Unlike the cross-entropy criterion for the frame-level alignment between input features and target labels, the CTC is aimed to automatically learn the alignment between speech data and label sequences (e.g., phonemes or characters, etc.), thereby eliminating the need for forced alignment of data, and the input is not necessarily the same as the label length.
  • speech data and label sequences e.g., phonemes or characters, etc.
  • a specified modeling unit is extracted from L, and the CTC is located on a softmax layer of the DNN.
  • the DNN is composed of an
  • y j t (j ⁇ [0,
  • An input sequence x T of a frame length T and a target label l ⁇ T are given, and l i ⁇ L.
  • a CTC path ⁇ ( ⁇ 0 , . . . , ⁇ r-1 ) is a frame-level label sequence, which is different from l in that the CTC path allows the appearance of repeated non-blank labels and blank units.
  • a many-to-one mapping function is defined as ⁇ , and “-” represents blank. If x T is given, and an output probability condition of each time step is assumed to be independent, the probability of the path ⁇ is:
  • the probability of the label sequence l can be calculated based on ⁇ by summing the probabilities of all the paths mapped to l. In some embodiments, it may be troublesome to sum up all the paths on the CTC in terms of calculation. With respect to this problem, a forward and backward dynamic programming algorithm can be adopted. All possible CTC paths are represented compactly as grids based on the algorithm, as shown in FIG. 6 .
  • a detection engine can make a positive decision accordingly, and it can be considered that corresponding keywords have been detected.
  • the set threshold can be fine-tuned based on a verification data set.
  • the model can be trained by a gradient descent method, such as an asynchronous random gradient descent method, to iteratively optimize parameters in the speech wakeup model until the training converges.
  • a gradient descent method such as an asynchronous random gradient descent method
  • the DNN and the CTC can be trained on a server having a Graphics Processing Unit (GPU).
  • Network parameters are randomly initialized to be uniformly distributed within a range of ( ⁇ 0.02, 0.02), an initial learning rate is 0.008, and a momentum is 0.9.
  • the learning rate is a parameter used in the gradient descent method.
  • a solution may be initialized first, and on the basis of this solution, a moving direction and a moving step size are determined, so that after the initial solution moves according to the direction and the step size, the output of a target function can be reduced. Then it is updated to a new solution, a next moving direction and a next step size are searched continuously, and after this process is performed iteratively, the target function is constantly decreased, to finally find a solution, such that the target function is relatively small.
  • the learning rate is used for the adjustment of the original step size. In the gradient descent method, the step size in each adjustment is equal to the learning rate multiplied by a gradient.
  • a verification data set may also be used to cross-verify the speech wakeup model to determine whether the training converges.
  • One measure is adaptive training. For example, a general model can be fine-tuned with speech data of some specific keywords and at a relatively low learning rate. Based on this consideration, when the speech wakeup model is trained, keyword-specific speech data can also be acquired, and the speech wakeup model can be trained with the keyword-specific speech data. A learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
  • the network parameters may not be randomly initialized, but refer to an existing corresponding network which has the same topology structure as the target network except for fine-grained units in the output layer, and may use a cross entropy criterion.
  • the transfer learning can be considered especially when the training data has a large scale.
  • architecture-related vector instructions e.g., ARM's NEON
  • NEON architecture-related vector instructions
  • a target label sequence corresponding to such user-defined keywords can be determined through a dictionary.
  • a speech wakeup method provided in the embodiment of this specification is as described in e foregoing.
  • a corresponding apparatus is further provided in an embodiment of this specification, as shown in FIG. 7 .
  • FIG. 7 is a schematic diagram of a speech wakeup apparatus 700 , according to an embodiment.
  • the apparatus 700 corresponds to the method 200 ( FIG. 2 ), and the dashed box in FIG. 7 represents an optional module.
  • the apparatus 700 may include an input module 701 and a speech wakeup model 702 .
  • Speech data is input by the input module 701 to the speech wakeup model 702 trained with general speech data, and the speech wakeup model 702 outputs a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • the general speech data includes a LVCSR corpus.
  • the apparatus further includes a training module 703 ; and training, by the training module 703 , the speech wakeup model with the general speech data includes: iteratively optimizing, by the training module 703 , parameters in the speech wakeup model with the general speech data by an asynchronous stochastic gradient descent method until the training converges.
  • the training module 703 further acquires keyword-specific speech data; and trains the speech wakeup model with the keyword-specific speech data, wherein a learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
  • the training module 703 cross-verifies the speech wakeup model with a verification data set in the training to determine whether the training converges.
  • the outputting, by the speech wakeup model 702 , a result for determining whether to execute speech wakeup specifically includes: extracting, by the speech wakeup model 702 , acoustic features from the input speech data; inputting the acoustic features to the DNN included in the speech wakeup model 702 for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme; inputting the class probability to the CTC included in the speech wakeup model 702 for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and determining whether to execute wakeup according to the confidence score, and outputting a determination result.
  • the extracting, by the speech wakeup model 702 , acoustic features from the input speech data specifically includes: extracting, by the speech wakeup model 702 , acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames is multi-dimension log filter bank energies; stacking a plurality of adjacent acoustic feature frames respectively; and taking the stacked acoustic feature frames respectively as acoustic features extracted from the monitored speech.
  • a corresponding electronic device is further provided in an embodiment of this specification, as shown FIG. 8 .
  • FIG. 8 is a schematic diagram of an electronic device 800 , according to an embodiment.
  • the electronic device 800 includes at least one processor 802 ; and a memory 804 communicatively connected to the at least one processor 802 .
  • the electronic device 800 may also include other hardware 806 , such as a network interface.
  • the at least one processor 802 may include one or more dedicated processing units, application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or various other types of processors or processing units.
  • the at least one processor 802 is coupled with e memory 804 and is configured to execute instructions stored in the memory 804 to perform the above described methods.
  • the memory 804 may include a non-permanent memory, a random access memory (RAM) and/or a non-volatile memory (such as a read-only memory (ROM) or a flash memory (flash RAM)), etc.
  • the memory 804 stores an instruction executable by the at least one processor 802 , and the instruction is executed by the at least one processor 802 to cause the electronic device 800 to: input speech data to a speech wakeup model trained with general speech data, and output, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • a corresponding non-transitory computer storage medium with a computer executable instruction stored thereon is further provided in an embodiment of this specification.
  • the computer executable instruction is configured to: input speech data to a speech wakeup model trained with general speech data, and output, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • the apparatus, the electronic device, and the computer storage medium provided in the embodiments of this specification are corresponding to the method. Therefore, the apparatus, the electronic device, and the non-volatile computer storage medium also have beneficial technical effects similar to those of the corresponding method. As the beneficial technical effects of the method have been described in detail in the foregoing, the beneficial technical effects of the apparatus, the electronic device, and the non-volatile computer storage medium will not be elaborated here.
  • PLD Programmable Logic Device
  • FPGA Field Programmable Gate Array
  • the logic compiler software is similar to a software complier used for developing and writing a program, and original codes before compiling also need to be written by using a specific programming language, which is referred to as a Hardware Description Language (HDL).
  • HDL Hardware Description Language
  • ABEL Advanced Boolean Expression Language
  • AHDL Altera Hardware Description Language
  • CUPL Cornell University Programming Language
  • HDCal Java Hardware Description Language
  • JHDL Java Hardware Description Language
  • Lava Lola
  • MyHDL MyHDL
  • PALASM Phase Change Language
  • RHDL Ruby Hardware Description Language
  • a controller can be implemented in any suitable manner in the above described devices.
  • the controller can employ a form of a microprocessor or a processor and a computer-readable storage medium that stores computer-readable program codes (such as software or firmware) executable by the microprocessor or processor, a logic gate, a switch, an Application Specific integrated Circuit (ASIC), a programmable logic controller, and an embedded microcontroller, Examples of the controller include, but are not limited to, the following microcontrollers: ARC 625D, Atmel AT91SAM, Microchip PIC18F26K20, and Silicone Labs C8051F320.
  • the controller of the memory can further be implemented as a part of control logic of the memory.
  • controller may be considered as a hardware component, and apparatuses included in the controller and configured to implement various functions may also be considered as structures inside the hardware component. Or, the apparatuses configured to implement various functions may even be considered as both software modules configured to implement the method and structures inside the hardware component.
  • the apparatuses, modules or models illustrated in the foregoing embodiments can be implemented by a computer chip or an entity, or implemented by a product having a specific function.
  • a typical implementation device is a computer.
  • the computer can be a personal computer, a laptop computer, a cellular phone, a camera phone, a smart phone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
  • the apparatus is divided into various units based on functions, and the units are described separately.
  • the functions of various units can also be implemented in one or more pieces of software and/or hardware.
  • the above described embodiments can be provided as a method, a system, or a computer program product. Therefore, the embodiments may be implemented in a form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, the embodiments of this specification can be in the form of a computer program product implemented on one or more computer usable storage media (including, but not limited to, a magnetic disk memory, a CD-ROM, an optical memory and the like) including computer usable program codes.
  • a computer usable storage media including, but not limited to, a magnetic disk memory, a CD-ROM, an optical memory and the like
  • the computer program instructions may also be stored in a computer-readable memory that can guide the computer or another programmable data processing device to work in a specific manner, such that the instruction stored in the computer-readable memory generates an article of manufacture including an instruction apparatus, and the instruction apparatus implements functions designated by one or more processes in a flowchart and/or one or more blocks in a block diagram.
  • the computer program instructions may also be loaded to the computer or another programmable data processing device, such that a series of operation steps are executed on the computer or another programmable device to generate a computer implemented processing, and therefore, the instruction executed in the computer or another programmable device provides steps for implementing functions designated in one or more processes in a flowchart and/or one or more blocks in a block diagram.
  • the computer-readable storage medium includes non-volatile and volatile media as well as movable and non-movable media, and can implement information storage by means of any, method or technology.
  • the information can be a computer-readable instruction, a data structure, and a module of a program or other data.
  • Examples of the computer-readable storage medium include, but are not limited to, a phase change memory (PRAM), a static random access memory (SRAM), a dynamic random access memory (DRAM), other types of RAM, a ROM, an electrically erasable programmable read-only memory (EEPROM), a flash memory or other memory technologies, a compact disk read-only memory (CD-ROM), a digital versatile disc (DVD) or other optical storages, a cassette tape, a magnetic tape/magnetic disk storage or other magnetic storage devices, or any other non-transmission medium, and can be used to store information accessible to the computing device.
  • the computer-readable storage medium does not include transitory media, such as a modulated data signal and a carrier.
  • the embodiments can be described in a general context of a computer executable instruction executed by a computer, for example, a program module.
  • the program module includes a routine, a program, an object, an assembly, a data structure, and the like used for executing a specific task or implementing a specific abstract data type.
  • the embodiments can also be implemented in distributed computing environments. In these distributed computing environments, a task is executed by using remote processing devices connected via a communications network. In the distributed computing environments, the program module may located in local and remote computer storage media including a storage device.

Abstract

A speech wakeup method, apparatus, and electronic device are disclosed in embodiments of this specification. The method includes: inputting speech data to a speech wakeup model trained with general speech data; and outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application is a continuation application of International Application No. PCT/CN2018/092899, filed on Jun. 26, 2018, which is based upon and claims priority to Chinese Patent Application No. 201710514348.6, filed on Jun. 29, 2017, the entire content of all of which is incorporated herein by reference.
  • TECHNICAL FIELD
  • This specification relates to the field of computer technologies, and in particular, to a speech wakeup method, apparatus, and electronic device.
  • TECHNICAL BACKGROUND
  • With the rapid development of mobile devices, speech-related technologies are also becoming increasingly common. For example, speech recognition is used in increasingly popular conversational assistants like Apple's Sin, Microsoft's Cortana, and Amazon's Alexa to enhance user experience and natural human-computer interaction level.
  • An important speech interaction technology is Keyword Spotting (KWS), which may also be generally referred to as speech wakeup or speech recognition. Based on the prior art, there is a need for a speech wakeup solution that may not rely on keyword-specific speech data.
  • SUMMARY
  • A speech wakeup method, apparatus, and electronic device are provided in embodiments of this specification, for solving the following technical problem: there is a need for a speech wakeup solution that may not rely on keyword-specific speech data.
  • According to one aspect, a speech wakeup method comprises: inputting speech data to a speech wakeup model trained with general speech data, and outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).
  • According to one aspect, an electronic device comprises: at least one processor; and a memory communicatively connected to the at least one processor; wherein the memory stores an instruction executable by the at least one processor, and the instruction is executed by the at least one processor to cause the electronic device to: input speech data to a speech wakeup model trained with general speech data, and output from the speech wakeup model a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • According to one aspect, a non-transitory computer-readable storage medium has stored therein instructions that, when executed by a processor of an electronic device, cause the electronic device to perform a speech wakeup method, the method comprising: inputting speech data to a speech wakeup model trained with general speech data, and outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • At least one of the foregoing technical solutions adopted in the embodiments of this specification can achieve the following beneficial effects: instead of relying on keyword-specific speech data, the speech wakeup model may be trained with accessible general speech data, and then the trained speech wakeup model can be used for speech wakeup, which is conducive to improving the accuracy of speech wakeup.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • The accompanying drawings, which are incorporated in and constitute a part of this specification, illustrate embodiments and, together with the description, serve to explain the principles of the specification.
  • FIG. 1 is a schematic diagram of an overall architecture for speech wakeup, according to an embodiment.
  • FIG. 2 is a flowchart of a speech wakeup method according to an embodiment.
  • FIG. 3 is a schematic diagram of a speech wakeup model according to an embodiment.
  • FIG. 4 is a schematic diagram of feature extraction performed by a feature extracting module according to an embodiment.
  • FIG. 5 is a schematic diagram of a Deep Neural Network (DNN) according to an embodiment.
  • FIG. 6 is a schematic diagram of a Connectionist Temporal Classifier (CTC) according to an embodiment.
  • FIG. 7 is a schematic diagram of a speech wakeup apparatus according to an embodiment.
  • FIG. 8 is a schematic diagram of an electronic device according to an embodiment.
  • DETAILED DESCRIPTION
  • A speech wakeup method, apparatus, and electronic device are provided in embodiments of this specification.
  • The technical solutions in the embodiments of this specification will be clearly and completely described in the following with reference to the accompanying drawings. The embodiments described are merely exemplary, rather than all, of the embodiments of this application. All other embodiments obtained by those of ordinary skill in the art based on the embodiments of this specification without creative efforts should all fall within the protection scope of this application.
  • In the embodiments described below, a speech wakeup model including a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC) is trained with general speech data. The trained speech wakeup model can be used for speech wakeup and support user-defined keywords triggered by speech wakeup. Further, the speech wakeup model can be used in low-power devices such as mobile phones and home appliances, because the DNN included in the speech wakeup model can be relatively not too complicated, and for example, can only have three or four layers with two or three hundred nodes in each layer. The speech wakeup model can be referred to as CTC-KWS.
  • In the embodiments, the DNN is a multi-layer perceptron, which has one or more hidden layers between an input layer and an output layer and can simulate complex nonlinear relationships. The CTC is a classifier configured to perform a label tagging task, and does not require forced alignment between input and output.
  • FIG. 1 is a schematic diagram of an overall architecture 100 for speech wakeup, according to an embodiment. The overall architecture 100 includes speech data 102 as a first part and a speech wakeup model 104 as a second part. The speech wakeup model 104 includes a DNN 106 and a CTC 108. Speech wakeup can be implemented by inputting the speech data 102 to the speech wakeup model 104 for processing. The embodiments described in detail in the following are based on the overall architecture 100.
  • FIG. 2 is a flowchart of a speech wakeup method 200 according to an embodiment. For example, the method 200 may be performed by a server or a terminal, for example, through a model training program, a speech recognition program, a speech wakeup application, and so on the server or the terminal. Also for example, the server or terminal may be a mobile phone, a tablet computer, a smart wearable device, an automobile machine, a personal computer, a medium-sized computer, a computer cluster, and so on. The method 200 may include the following steps.
  • In S202, speech data is input to a speech wakeup model trained with general speech data.
  • In the embodiment, speech can be monitored by the server or the terminal to obtain the speech data. When the speech is monitored, a user can speak out a predetermined keyword to trigger the speech wakeup model to execute speech wakeup.
  • In S204, the speech wakeup model outputs a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • In the embodiment, compared with the keyword-specific speech data mentioned in the Background, the general speech data described in step S202 is less restricted and, thus, easily, accessible. For example, it may be a Urge Vocabulary Continuous Speech Recognition (LVCSR) corpus or the like.
  • In the embodiment, the DNN included in the speech recognition model may predict a posterior probability distribution of a pronunciation phoneme sequence corresponding to input speech features. The DNN can be followed by the CTC to give a confidence score corresponding to the predicted pronunciation phoneme sequence. A result for determining whether to execute speech wakeup can be output based on the confidence score.
  • With the method 200, instead of relying on keyword-specific speech data, the speech wakeup model may be trained with the accessible general speech data, and further the trained speech wakeup model can be used for speech wakeup, which is conducive to improving the accuracy of speech wakeup.
  • The speech wakeup model also breaks through the restrictions of the keyword-specific speech data and supports user-defined triggered keywords. Therefore, it is more convenient and flexible in actual applications and conducive to improving user experience.
  • Based on the method 200, some specific implementations are further provided in the embodiments below.
  • FIG. 3 is a schematic diagram of a speech wakeup model 300, according to an embodiment. Referring to FIG. 3, the speech wakeup model 300 may include a feature extracting module 302, a DNN 304, and a CTC 306 in sequence. In step S204 of the method 200 (FIG. 2), the outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup may include: the feature extracting module 302 extracting acoustic features from the input speech data; inputting the acoustic features to the DNN 304 for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme; the DNN 304 inputting the class probability to the CTC 306 for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and the CTC 306 determining whether to execute wakeup according to the confidence score, and outputting a determination result.
  • The speech wakeup model 300 is described in detail below in further combination with FIG. 4, FIG. 5, and FIG. 6.
  • FIG. 4 is a schematic diagram of feature extraction performed by a feature extracting module, such as the feature extracting module 302 (FIG. 3), according to an embodiment.
  • In FIG. 4, when a section of speech “Zhi Ma Kai Men” is monitored at present, a target label sequence corresponding thereto is a pronunciation phoneme sequence, which can be expressed as: “zhi1ma2kai1men2”, wherein the numbers represent tones. In addition to phonemes such as initials and finals, tone phonemes are also taken into account as a modeling unit. In an embodiment, context-independent or context-dependent phonemes can all be taken into account, among which the latter is more numerous. However, in consideration of reducing subsequent computational burden of the DNN, only the context-independent phonemes may be considered, such as 72 context-independent phoneme units in Chinese, including a blank unit.
  • In an embodiment, acoustic features can be extracted by the feature extracting module from the input speech data, which may include: extracting acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames may be multi-dimension log filter bank energies; stacking a plurality of adjacent acoustic feature frames respectively; taking the stacked acoustic feature frames respectively as acoustic features extracted from the general speech data; and further, the stacked acoustic feature frames can be used as inputs of the DNN respectively.
  • The log filter bank energies refer to energy signals extracted by a log filter bank, which can be expressed as a vector to facilitate model processing. The multi-dimension in the foregoing represents multiple dimensions of the vector.
  • For example, a specified length of a time window may be 25 milliseconds, each time window may move for 10 milliseconds, and the multi-dimension may be, for example, 40 dimensions. On a time axis of the speech data, milliseconds from 0 to 25 may be used as a window, and 40-dimension log filter bank energies are correspondingly extracted from the speech data to serve as a first acoustic feature frame; milliseconds from 10 to 35 can be used as a window, and 40-dimension log filter bank energies are correspondingly extracted from the speech data to serve as a second acoustic feature frame; and multiple acoustic feature frames can be extracted in the same way.
  • Further, stacking a plurality of adjacent acoustic feature frames may allow more information from a context of a current frame, which is conducive to improving the accuracy of subsequent prediction results. Following the above example, the current frame, the adjacent consecutive ten frames before the current frame, and the adjacent consecutive five frames after the current frame can be, for example, stacked to obtain a 640-dimension stacking feature for being inputted to the subsequent DNN. In addition, cepstral mean and variance normalization can be carried out on the dimensions of the stacking feature, and then backward input can be carried out.
  • It should be noted that the feature extracting manner and parameters adopted in the foregoing embodiment are only exemplary, and they can be adjusted as needed in actual applications.
  • FIG. 5 is a schematic diagram of a DNN, such as the DNN 304 (FIG. 3), according to an embodiment. Referring to FIG. 5, the DNN may include an input layer 502, one or more hidden layers 504, and an output layer 506.
  • In FIG. 5, various neurons, represented by circles, in the DNN are fully connected. The acoustic features extracted by the feature extracting module 302 (FIG. 3) are input to the DNN. The DNN can describe a relationship between an input acoustic feature x0∈Rn 0 in the input layer 502 and a modeling unit j in the output layer 506 according to the following function mapping:
  • z i = x i - 1 W i T + B i , 1 i N + 1 ; ( formula I ) x i = σ ( z i ) , 1 i N ; ( formula II ) y j = exp ( z N + 1 , j ) k exp ( z N + 1 , k ) ; ( formula III )
  • where xi,i>0∈Rn i is an output of a hidden layer, Wi∈Rn i ×n i−1 and Bi∈RRn i are weights and offset parameters respectively, which may be predetermined based on, training or an application need, ni is the number of nodes on the ith layer, θ={Wi, Bi}, “T” denotes transpose of a matrix, N is the number of the hidden layers, and σ is a nonlinear activation function, for example, a Rectified Linear Unit (ReLU) function σ(z)=max(z, 0). The formula III is a softmax function in the embodiment, representing an estimated posterior probability of a label unit j.
  • In an embodiment, a Recurrent Neural Network (RNN) can also be used in conjunction with the CTC. However, in the case where the training data is restricted such as in the Background, the minimum computing and power consumption requirements of mobile devices can be more easily met by using the DNN in conjunction with the CTC. In order to reduce the complexity in computation, the DNN with about hundreds of nodes in a hidden layer is more suitable.
  • FIG. 6 is a schematic diagram of a CTC, such as the CTC 306 (FIG. 3), according to an embodiment.
  • The CTC is configured for sequential label tasks. Unlike the cross-entropy criterion for the frame-level alignment between input features and target labels, the CTC is aimed to automatically learn the alignment between speech data and label sequences (e.g., phonemes or characters, etc.), thereby eliminating the need for forced alignment of data, and the input is not necessarily the same as the label length.
  • In FIG. 6, a specified modeling unit is extracted from L, and the CTC is located on a softmax layer of the DNN. The DNN is composed of an |L| unit and a blank unit. The introduction of the blank unit alleviates the burden of label prediction, as symbols are not output correspondingly during uncertainty.
  • In the embodiment, yj t(j∈[0, |L|], t∈[0, T)) is defined as a probability that the DNN outputs j in a time step t. An input sequence xT of a frame length T and a target label l≤T are given, and li∈L. A CTC path π(π0, . . . , πr-1) is a frame-level label sequence, which is different from l in that the CTC path allows the appearance of repeated non-blank labels and blank units.
  • The CTC path π can be mapped to its corresponding label sequence l by deleting the repeated labels and the blank units. For example, τ(“aa-b-c”)=τ(“abb-cc-”)=“abc”. A many-to-one mapping function is defined as τ, and “-” represents blank. If xT is given, and an output probability condition of each time step is assumed to be independent, the probability of the path π is:

  • p(π|x;θ)=Πt=0 T−1 y π t t;  (formula IV)
  • Then, the probability of the label sequence l can be calculated based on τ by summing the probabilities of all the paths mapped to l. In some embodiments, it may be troublesome to sum up all the paths on the CTC in terms of calculation. With respect to this problem, a forward and backward dynamic programming algorithm can be adopted. All possible CTC paths are represented compactly as grids based on the algorithm, as shown in FIG. 6.
  • At the time of training, the CTC is aimed to enable θ*=argminθΣ(x,l)∈s−log(p(l|x; θ)), where S represents training data used. At the time of decoding, when the confidence score output by the CTC, which may be expressed as log(p(l|x; θ*)), is greater than a set threshold, a detection engine can make a positive decision accordingly, and it can be considered that corresponding keywords have been detected. The set threshold can be fine-tuned based on a verification data set.
  • In the embodiment, the model can be trained by a gradient descent method, such as an asynchronous random gradient descent method, to iteratively optimize parameters in the speech wakeup model until the training converges.
  • For example, the DNN and the CTC can be trained on a server having a Graphics Processing Unit (GPU). Network parameters are randomly initialized to be uniformly distributed within a range of (−0.02, 0.02), an initial learning rate is 0.008, and a momentum is 0.9.
  • The learning rate is a parameter used in the gradient descent method. In the gradient descent method, a solution may be initialized first, and on the basis of this solution, a moving direction and a moving step size are determined, so that after the initial solution moves according to the direction and the step size, the output of a target function can be reduced. Then it is updated to a new solution, a next moving direction and a next step size are searched continuously, and after this process is performed iteratively, the target function is constantly decreased, to finally find a solution, such that the target function is relatively small. In the process of finding the solution, if the step size is too large and the search is not careful enough, a good solution may be bypassed, and if the step size is too small, the process of finding the solution will proceed too slowly. Therefore, it is important to set the step size appropriately. The learning rate is used for the adjustment of the original step size. In the gradient descent method, the step size in each adjustment is equal to the learning rate multiplied by a gradient.
  • In the foregoing training, a verification data set may also be used to cross-verify the speech wakeup model to determine whether the training converges.
  • In order to improve the performance and robustness of the speech wakeup model, more measures are further provided in the solutions of this specification.
  • One measure is adaptive training. For example, a general model can be fine-tuned with speech data of some specific keywords and at a relatively low learning rate. Based on this consideration, when the speech wakeup model is trained, keyword-specific speech data can also be acquired, and the speech wakeup model can be trained with the keyword-specific speech data. A learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
  • Another measure is transfer learning. For example, the network parameters may not be randomly initialized, but refer to an existing corresponding network which has the same topology structure as the target network except for fine-grained units in the output layer, and may use a cross entropy criterion. The transfer learning can be considered especially when the training data has a large scale.
  • Other measures include, for example, using architecture-related vector instructions (e.g., ARM's NEON) to further accelerate multiplication, and so on.
  • As mentioned previously, the solution of this specification can easily support user-defined speech wakeup keywords. A target label sequence corresponding to such user-defined keywords can be determined through a dictionary.
  • A speech wakeup method provided in the embodiment of this specification is as described in e foregoing. A corresponding apparatus is further provided in an embodiment of this specification, as shown in FIG. 7.
  • FIG. 7 is a schematic diagram of a speech wakeup apparatus 700, according to an embodiment. For example, the apparatus 700 corresponds to the method 200 (FIG. 2), and the dashed box in FIG. 7 represents an optional module. The apparatus 700 may include an input module 701 and a speech wakeup model 702.
  • Speech data is input by the input module 701 to the speech wakeup model 702 trained with general speech data, and the speech wakeup model 702 outputs a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • In an embodiment, the general speech data includes a LVCSR corpus.
  • In an embodiment, the apparatus further includes a training module 703; and training, by the training module 703, the speech wakeup model with the general speech data includes: iteratively optimizing, by the training module 703, parameters in the speech wakeup model with the general speech data by an asynchronous stochastic gradient descent method until the training converges.
  • In an embodiment, the training module 703 further acquires keyword-specific speech data; and trains the speech wakeup model with the keyword-specific speech data, wherein a learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
  • In an embodiment, the training module 703 cross-verifies the speech wakeup model with a verification data set in the training to determine whether the training converges.
  • In an embodiment, the outputting, by the speech wakeup model 702, a result for determining whether to execute speech wakeup specifically includes: extracting, by the speech wakeup model 702, acoustic features from the input speech data; inputting the acoustic features to the DNN included in the speech wakeup model 702 for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme; inputting the class probability to the CTC included in the speech wakeup model 702 for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and determining whether to execute wakeup according to the confidence score, and outputting a determination result.
  • In an embodiment, the extracting, by the speech wakeup model 702, acoustic features from the input speech data specifically includes: extracting, by the speech wakeup model 702, acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames is multi-dimension log filter bank energies; stacking a plurality of adjacent acoustic feature frames respectively; and taking the stacked acoustic feature frames respectively as acoustic features extracted from the monitored speech.
  • A corresponding electronic device is further provided in an embodiment of this specification, as shown FIG. 8.
  • FIG. 8 is a schematic diagram of an electronic device 800, according to an embodiment. For example, the electronic device 800 includes at least one processor 802; and a memory 804 communicatively connected to the at least one processor 802. The electronic device 800 may also include other hardware 806, such as a network interface. The at least one processor 802 may include one or more dedicated processing units, application-specific integrated circuits (ASICs), field-programmable gate arrays (FPGAs), or various other types of processors or processing units. The at least one processor 802 is coupled with e memory 804 and is configured to execute instructions stored in the memory 804 to perform the above described methods. The memory 804 may include a non-permanent memory, a random access memory (RAM) and/or a non-volatile memory (such as a read-only memory (ROM) or a flash memory (flash RAM)), etc.
  • The memory 804 stores an instruction executable by the at least one processor 802, and the instruction is executed by the at least one processor 802 to cause the electronic device 800 to: input speech data to a speech wakeup model trained with general speech data, and output, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • A corresponding non-transitory computer storage medium with a computer executable instruction stored thereon is further provided in an embodiment of this specification. The computer executable instruction is configured to: input speech data to a speech wakeup model trained with general speech data, and output, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a DNN and a CTC.
  • Specific embodiments of this specification have been described in the foregoing. Other embodiments fall within the scope of the appended claims Under some circumstances, the actions or steps described in the claims may be performed in a sequence different from that in the embodiments and still can achieve a desired result. In addition, the processes depicted in the accompanying drawings are not necessarily required to achieve the desired result according to the specific sequence or consecutive sequence shown. Multitask processing and parallel processing are also possible or may be advantageous in some implementation manners.
  • The embodiments in the specification are described progressively, identical or similar parts of the embodiments may be obtained with reference to each other, and each embodiment emphasizes a part different from other embodiments. Especially, the apparatus, electronic device, and computer storage medium embodiments are similar to the method embodiments. For related parts, refer to the descriptions of the parts in the method embodiments.
  • The apparatus, the electronic device, and the computer storage medium provided in the embodiments of this specification are corresponding to the method. Therefore, the apparatus, the electronic device, and the non-volatile computer storage medium also have beneficial technical effects similar to those of the corresponding method. As the beneficial technical effects of the method have been described in detail in the foregoing, the beneficial technical effects of the apparatus, the electronic device, and the non-volatile computer storage medium will not be elaborated here.
  • Each of the above described methods, modules and models may be implemented as software, or hardware, or a combination of software and hardware. For example, a Programmable Logic Device (PLD) (e.g., a Field Programmable Gate Array (FPGA)) is an integrated circuit, and its logic functions are determined by a user programming device. Designers program by themselves to integrate a digital system into a PLD, without asking a chip manufacturer to design and manufacture a dedicated integrated circuit chip. Moreover, at present, the programming may be implemented by using logic compiler software instead of manually manufacturing an integrated circuit chip. The logic compiler software is similar to a software complier used for developing and writing a program, and original codes before compiling also need to be written by using a specific programming language, which is referred to as a Hardware Description Language (HDL). There are many types of HDLs, such as Advanced Boolean Expression Language (ABEL), Altera Hardware Description Language (AHDL), Confluence, Cornell University Programming Language (CUPL), HDCal, Java Hardware Description Language (JHDL), Lava, Lola, MyHDL, PALASM, and Ruby Hardware Description Language (RHDL), among which Very-High-Speed Integrated Circuit Hardware Description Language (VHDL) and Verilog are most commonly used now. Those skilled in the art should also know that a hardware circuit for implementing the logic method procedure can be easily obtained by slightly logically programming the method procedure using the above several hardware description languages and programming it into an integrated circuit.
  • A controller can be implemented in any suitable manner in the above described devices. For example, the controller can employ a form of a microprocessor or a processor and a computer-readable storage medium that stores computer-readable program codes (such as software or firmware) executable by the microprocessor or processor, a logic gate, a switch, an Application Specific integrated Circuit (ASIC), a programmable logic controller, and an embedded microcontroller, Examples of the controller include, but are not limited to, the following microcontrollers: ARC 625D, Atmel AT91SAM, Microchip PIC18F26K20, and Silicone Labs C8051F320. The controller of the memory can further be implemented as a part of control logic of the memory. Those skilled in the art also know that in addition to implementing the controller by using pure computer-readable program codes, it is completely feasible to logically program the method steps to enable the controller to implement the same function in a form of a logic gate, a switch, an ASIC, a programmable logic controller, and an embedded microcontroller. Therefore, such a controller may be considered as a hardware component, and apparatuses included in the controller and configured to implement various functions may also be considered as structures inside the hardware component. Or, the apparatuses configured to implement various functions may even be considered as both software modules configured to implement the method and structures inside the hardware component.
  • The apparatuses, modules or models illustrated in the foregoing embodiments can be implemented by a computer chip or an entity, or implemented by a product having a specific function. A typical implementation device is a computer. For example, the computer can be a personal computer, a laptop computer, a cellular phone, a camera phone, a smart phone, a personal digital assistant, a media player, a navigation device, an email device, a game console, a tablet computer, a wearable device, or a combination of any of these devices.
  • For ease of description, the apparatus is divided into various units based on functions, and the units are described separately. In an implementation of this specification, the functions of various units can also be implemented in one or more pieces of software and/or hardware.
  • Those skilled in the art should understand that the above described embodiments can be provided as a method, a system, or a computer program product. Therefore, the embodiments may be implemented in a form of a complete hardware embodiment, a complete software embodiment, or an embodiment combining software and hardware. Moreover, the embodiments of this specification can be in the form of a computer program product implemented on one or more computer usable storage media (including, but not limited to, a magnetic disk memory, a CD-ROM, an optical memory and the like) including computer usable program codes.
  • The embodiments are described with reference to flowcharts and/or block diagrams of the method, the device and the computer program product according to the embodiments of this specification. It should be understood that computer program instructions may be used to implement each process and/or block in the flowcharts and/or block diagrams and combinations of processes and/or blocks in the flowcharts and/or block diagrams. The computer program instructions may be provided to a general-purpose computer, a special-purpose computer, an embedded processor or a processor of another programmable data processing device to generate a machine, such that the computer or the processor of another programmable data processing device executes an instruction to generate an apparatus configured to implement functions designated in one or more processes in a flowchart and/or one or more blocks in a block diagram.
  • The computer program instructions may also be stored in a computer-readable memory that can guide the computer or another programmable data processing device to work in a specific manner, such that the instruction stored in the computer-readable memory generates an article of manufacture including an instruction apparatus, and the instruction apparatus implements functions designated by one or more processes in a flowchart and/or one or more blocks in a block diagram.
  • The computer program instructions may also be loaded to the computer or another programmable data processing device, such that a series of operation steps are executed on the computer or another programmable device to generate a computer implemented processing, and therefore, the instruction executed in the computer or another programmable device provides steps for implementing functions designated in one or more processes in a flowchart and/or one or more blocks in a block diagram.
  • The computer-readable storage medium includes non-volatile and volatile media as well as movable and non-movable media, and can implement information storage by means of any, method or technology. The information can be a computer-readable instruction, a data structure, and a module of a program or other data. Examples of the computer-readable storage medium include, but are not limited to, a phase change memory (PRAM), a static random access memory (SRAM), a dynamic random access memory (DRAM), other types of RAM, a ROM, an electrically erasable programmable read-only memory (EEPROM), a flash memory or other memory technologies, a compact disk read-only memory (CD-ROM), a digital versatile disc (DVD) or other optical storages, a cassette tape, a magnetic tape/magnetic disk storage or other magnetic storage devices, or any other non-transmission medium, and can be used to store information accessible to the computing device. The computer-readable storage medium does not include transitory media, such as a modulated data signal and a carrier.
  • It should be further rioted that the terms “include”, “comprise” or any other variations thereof are intended to cover non-exclusive inclusion, so that a process, method, commodity or device including a series of elements not only includes the elements, but also includes other elements not expressly listed, or further includes elements inherent to the process, method, commodity or device. In the absence of more limitations, an element defined by “including a/an . . . ” does not exclude that the process, method, commodity or device including the element further has other identical elements.
  • The embodiments can be described in a general context of a computer executable instruction executed by a computer, for example, a program module. Generally, the program module includes a routine, a program, an object, an assembly, a data structure, and the like used for executing a specific task or implementing a specific abstract data type. The embodiments can also be implemented in distributed computing environments. In these distributed computing environments, a task is executed by using remote processing devices connected via a communications network. In the distributed computing environments, the program module may located in local and remote computer storage media including a storage device.
  • Although the specification has been described in conjunction with specific embodiments, many alternatives, modifications and variations will be apparent to those skilled in the art. Accordingly, the following claims embrace all such alternatives, modifications and variations that fall within the terms of the claims.

Claims (15)

1. A speech wakeup method, comprising:
inputting speech data to a speech wakeup model trained with general speech data; and
outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).
2. The method of claim 1, wherein the general speech data comprises a Large Vocabulary Continuous Speech Recognition (LVCSR) corpus.
3. The method of claim 1, further comprising:
training the speech wakeup model with the general speech data, wherein the training includes:
iteratively optimizing parameters in the speech wakeup model with the general speech data by means of an asynchronous stochastic gradient descent method until the training converges.
4. The method of claim 3, further comprising:
acquiring keyword-specific speech data; and
training the speech wakeup model with the keyword-specific speech data, wherein a learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
5. The method of claim 3, further comprising:
cross-verifying the speech wakeup model with a verification data set in the training to determine whether the training converges.
6. The method of claim 1, wherein the outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup comprises:
extracting acoustic features from the input speech data;
inputting the acoustic features to the DNN included in the speech wakeup model for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme;
inputting the class probability to the CTC included in the speech wakeup model for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and
determining whether to execute speech wakeup according to the confidence score, and outputting a determination result.
7. The method of claim 6, wherein the extracting acoustic features from the input speech data comprises:
extracting acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames is multi-dimension log filter bank energies;
stacking a plurality of adjacent acoustic feature frames respectively; and
taking the stacked acoustic feature frames respectively as acoustic features extracted from the input speech data.
8. An electronic device, comprising:
at least one processor; and
a memory communicatively connected to the at least one processor;
wherein the memory stores an instruction executable by the at least one processor, and the instruction is executed by the at least one processor to cause the electronic device to:
input speech data to a speech wakeup model trained with general speech data; and
output from the speech wakeup model a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).
9. The electronic device of claim 8, wherein the general speech data comprises a large vocabulary continuous speech recognition (LVCSR) corpus.
10. The electronic device of claim 8, wherein the instruction is further executed by the at least one processor to cause the electronic device to:
train the speech wakeup model with the general speech data, including iteratively optimizing parameters in the speech wakeup model with the general speech data by means of an asynchronous stochastic gradient descent method until the training converges.
11. The electronic device of claim 10, wherein the instruction is further executed by the at least one processor to cause the electronic device to:
acquire keyword-specific speech data; and
train the speech wakeup model with the keyword-specific speech data, wherein a learning rate used in the training is less than that used in the training of the speech wakeup model with the general speech data.
12. The electronic device of claim 10, wherein the instruction is further executed by the at least one processor to cause the electronic device to: cross-verify the speech wakeup model with a verification data set in the training to determine whether the training converges.
13. The electronic device of claim 8, wherein in outputting from the speech wakeup model a result for determining whether to execute speech wakeup, the instruction is further executed by the at least one processor to cause the electronic device to:
extract, by the speech wakeup model, acoustic features from the input speech data;
input the acoustic features to the DNN included in the speech wakeup model for processing to obtain a class probability of the acoustic features respectively corresponding to each pronunciation phoneme;
input the class probability to the CTC included in the speech wakeup model for processing to obtain a confidence score of a speech wakeup term corresponding to a pronunciation phoneme sequence; and
determine whether to execute wakeup according to the confidence score, and output a determination result.
14. The electronic device of claim 13, wherein in extracting, by the speech wakeup model, acoustic features from the input speech data, the instruction is further executed by the at least one processor to cause the electronic device to:
extract, by the speech wakeup model, acoustic feature frames of the input speech data from a window according to a specified time interval, wherein each of the acoustic feature frames is multi-dimension log filter bank energies;
stack a plurality of adjacent acoustic feature frames respectively; and
take the stacked acoustic feature frames respectively as acoustic features extracted from the input speech data.
15. A non-transitory computer-readable storage medium having stored therein instructions that, when executed by a processor of an electronic device, cause the electronic device to perform a speech wakeup method, the method comprising:
inputting speech data to a speech wakeup model trained with general speech data; and
outputting, by the speech wakeup model, a result for determining whether to execute speech wakeup, wherein the speech wakeup model includes a Deep Neural Network (DNN) and a Connectionist Temporal Classifier (CTC).
US16/571,468 2017-06-29 2019-09-16 Speech wakeup method, apparatus, and electronic device Abandoned US20200013390A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US16/774,422 US10748524B2 (en) 2017-06-29 2020-01-28 Speech wakeup method, apparatus, and electronic device

Applications Claiming Priority (3)

Application Number Priority Date Filing Date Title
CN201710514348.6A CN107358951A (en) 2017-06-29 2017-06-29 A kind of voice awakening method, device and electronic equipment
CN201710514348.6 2017-06-29
PCT/CN2018/092899 WO2019001428A1 (en) 2017-06-29 2018-06-26 Voice wake-up method and device and electronic device

Related Parent Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2018/092899 Continuation WO2019001428A1 (en) 2017-06-29 2018-06-26 Voice wake-up method and device and electronic device

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US16/774,422 Continuation US10748524B2 (en) 2017-06-29 2020-01-28 Speech wakeup method, apparatus, and electronic device

Publications (1)

Publication Number Publication Date
US20200013390A1 true US20200013390A1 (en) 2020-01-09

Family

ID=60274110

Family Applications (2)

Application Number Title Priority Date Filing Date
US16/571,468 Abandoned US20200013390A1 (en) 2017-06-29 2019-09-16 Speech wakeup method, apparatus, and electronic device
US16/774,422 Active US10748524B2 (en) 2017-06-29 2020-01-28 Speech wakeup method, apparatus, and electronic device

Family Applications After (1)

Application Number Title Priority Date Filing Date
US16/774,422 Active US10748524B2 (en) 2017-06-29 2020-01-28 Speech wakeup method, apparatus, and electronic device

Country Status (11)

Country Link
US (2) US20200013390A1 (en)
EP (1) EP3579227B1 (en)
JP (1) JP6877558B2 (en)
KR (1) KR102181836B1 (en)
CN (1) CN107358951A (en)
ES (1) ES2878137T3 (en)
PH (1) PH12019501674A1 (en)
PL (1) PL3579227T3 (en)
SG (1) SG11201906576WA (en)
TW (1) TWI692751B (en)
WO (1) WO2019001428A1 (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111883121A (en) * 2020-07-20 2020-11-03 北京声智科技有限公司 Awakening method and device and electronic equipment
CN112233655A (en) * 2020-09-28 2021-01-15 上海声瀚信息科技有限公司 Neural network training method for improving voice command word recognition performance
CN112669818A (en) * 2020-12-08 2021-04-16 北京地平线机器人技术研发有限公司 Voice wake-up method and device, readable storage medium and electronic equipment
CN113113007A (en) * 2021-03-30 2021-07-13 北京金山云网络技术有限公司 Voice data processing method and device, electronic equipment and storage medium
US11081102B2 (en) * 2019-08-16 2021-08-03 Ponddy Education Inc. Systems and methods for comprehensive Chinese speech scoring and diagnosis
US20220293088A1 (en) * 2021-03-12 2022-09-15 Samsung Electronics Co., Ltd. Method of generating a trigger word detection model, and an apparatus for the same
US11587550B2 (en) * 2020-06-10 2023-02-21 Apollo Intelligent Connectivity (Beijing) Technology Co., Ltd. Method and apparatus for outputting information
WO2023085699A1 (en) * 2021-11-10 2023-05-19 삼성전자주식회사 Electronic device and control method thereof
US11967322B2 (en) 2021-05-06 2024-04-23 Samsung Electronics Co., Ltd. Server for identifying false wakeup and method for controlling the same

Families Citing this family (27)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107358951A (en) * 2017-06-29 2017-11-17 阿里巴巴集团控股有限公司 A kind of voice awakening method, device and electronic equipment
CN108320733B (en) * 2017-12-18 2022-01-04 上海科大讯飞信息科技有限公司 Voice data processing method and device, storage medium and electronic equipment
CN108182937B (en) * 2018-01-17 2021-04-13 出门问问创新科技有限公司 Keyword recognition method, device, equipment and storage medium
US11488002B2 (en) * 2018-02-15 2022-11-01 Atlazo, Inc. Binary neural network accelerator engine methods and systems
CN108597523B (en) * 2018-03-23 2019-05-17 平安科技(深圳)有限公司 Identified by speaking person method, server and computer readable storage medium
WO2019222996A1 (en) * 2018-05-25 2019-11-28 Beijing Didi Infinity Technology And Development Co., Ltd. Systems and methods for voice recognition
CN110619871B (en) * 2018-06-20 2023-06-30 阿里巴巴集团控股有限公司 Voice wakeup detection method, device, equipment and storage medium
US11257481B2 (en) * 2018-10-24 2022-02-22 Tencent America LLC Multi-task training architecture and strategy for attention-based speech recognition system
CN111276138B (en) * 2018-12-05 2023-07-18 北京嘀嘀无限科技发展有限公司 Method and device for processing voice signal in voice wake-up system
CN109886386B (en) * 2019-01-30 2020-10-27 北京声智科技有限公司 Method and device for determining wake-up model
CN109872713A (en) * 2019-03-05 2019-06-11 深圳市友杰智新科技有限公司 A kind of voice awakening method and device
CN110310628B (en) * 2019-06-27 2022-05-20 百度在线网络技术(北京)有限公司 Method, device and equipment for optimizing wake-up model and storage medium
JP7098587B2 (en) * 2019-08-29 2022-07-11 株式会社東芝 Information processing device, keyword detection device, information processing method and program
CN110634468B (en) * 2019-09-11 2022-04-15 中国联合网络通信集团有限公司 Voice wake-up method, device, equipment and computer readable storage medium
CN110648659B (en) * 2019-09-24 2022-07-01 上海依图信息技术有限公司 Voice recognition and keyword detection device and method based on multitask model
CN110648668A (en) * 2019-09-24 2020-01-03 上海依图信息技术有限公司 Keyword detection device and method
CN110970016B (en) * 2019-10-28 2022-08-19 苏宁云计算有限公司 Awakening model generation method, intelligent terminal awakening method and device
CN110853629A (en) * 2019-11-21 2020-02-28 中科智云科技有限公司 Speech recognition digital method based on deep learning
CN110992929A (en) * 2019-11-26 2020-04-10 苏宁云计算有限公司 Voice keyword detection method, device and system based on neural network
US11341954B2 (en) * 2019-12-17 2022-05-24 Google Llc Training keyword spotters
JP7438744B2 (en) 2019-12-18 2024-02-27 株式会社東芝 Information processing device, information processing method, and program
CN112733272A (en) * 2021-01-13 2021-04-30 南昌航空大学 Method for solving vehicle path problem with soft time window
KR102599480B1 (en) * 2021-05-18 2023-11-08 부산대학교 산학협력단 System and Method for automated training keyword spotter
CN113160823A (en) * 2021-05-26 2021-07-23 中国工商银行股份有限公司 Voice awakening method and device based on pulse neural network and electronic equipment
CN113990296B (en) * 2021-12-24 2022-05-27 深圳市友杰智新科技有限公司 Training method and post-processing method of voice acoustic model and related equipment
US20240119925A1 (en) * 2022-10-10 2024-04-11 Samsung Electronics Co., Ltd. System and method for post-asr false wake-up suppression
CN115862604B (en) * 2022-11-24 2024-02-20 镁佳(北京)科技有限公司 Voice awakening model training and voice awakening method and device and computer equipment

Family Cites Families (23)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JPH05128286A (en) * 1991-11-05 1993-05-25 Ricoh Co Ltd Keyword spotting system by neural network
JP2007179239A (en) * 2005-12-27 2007-07-12 Kenwood Corp Schedule management device and program
US9117449B2 (en) * 2012-04-26 2015-08-25 Nuance Communications, Inc. Embedded system for construction of small footprint speech recognition with user-definable constraints
US9177547B2 (en) * 2013-06-25 2015-11-03 The Johns Hopkins University System and method for processing speech to identify keywords or other information
CN104378723A (en) * 2013-08-16 2015-02-25 上海耐普微电子有限公司 Microphone with voice wake-up function
US9715660B2 (en) * 2013-11-04 2017-07-25 Google Inc. Transfer learning for deep neural network based hotword detection
US9443522B2 (en) * 2013-11-18 2016-09-13 Beijing Lenovo Software Ltd. Voice recognition method, voice controlling method, information processing method, and electronic apparatus
CN105096935B (en) * 2014-05-06 2019-08-09 阿里巴巴集团控股有限公司 A kind of pronunciation inputting method, device and system
US10783900B2 (en) * 2014-10-03 2020-09-22 Google Llc Convolutional, long short-term memory, fully connected deep neural networks
EP3282445A4 (en) * 2015-04-10 2018-05-02 Huawei Technologies Co. Ltd. Voice recognition method, voice wake-up device, voice recognition device and terminal
CN106297774B (en) * 2015-05-29 2019-07-09 中国科学院声学研究所 A kind of the distributed parallel training method and system of neural network acoustic model
TWI639153B (en) * 2015-11-03 2018-10-21 絡達科技股份有限公司 Electronic apparatus and voice trigger method therefor
JP6679898B2 (en) * 2015-11-24 2020-04-15 富士通株式会社 KEYWORD DETECTION DEVICE, KEYWORD DETECTION METHOD, AND KEYWORD DETECTION COMPUTER PROGRAM
US10755698B2 (en) 2015-12-07 2020-08-25 University Of Florida Research Foundation, Inc. Pulse-based automatic speech recognition
CN106887227A (en) * 2015-12-16 2017-06-23 芋头科技(杭州)有限公司 A kind of voice awakening method and system
CN105632486B (en) * 2015-12-23 2019-12-17 北京奇虎科技有限公司 Voice awakening method and device of intelligent hardware
US10229672B1 (en) * 2015-12-31 2019-03-12 Google Llc Training acoustic models using connectionist temporal classification
CN105931633A (en) * 2016-05-30 2016-09-07 深圳市鼎盛智能科技有限公司 Speech recognition method and system
CN106098059B (en) * 2016-06-23 2019-06-18 上海交通大学 Customizable voice awakening method and system
CN106611597B (en) * 2016-12-02 2019-11-08 百度在线网络技术(北京)有限公司 Voice awakening method and device based on artificial intelligence
CN106782536B (en) * 2016-12-26 2020-02-28 北京云知声信息技术有限公司 Voice awakening method and device
CN107221326B (en) * 2017-05-16 2021-05-28 百度在线网络技术(北京)有限公司 Voice awakening method and device based on artificial intelligence and computer equipment
CN107358951A (en) * 2017-06-29 2017-11-17 阿里巴巴集团控股有限公司 A kind of voice awakening method, device and electronic equipment

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US11081102B2 (en) * 2019-08-16 2021-08-03 Ponddy Education Inc. Systems and methods for comprehensive Chinese speech scoring and diagnosis
US11587550B2 (en) * 2020-06-10 2023-02-21 Apollo Intelligent Connectivity (Beijing) Technology Co., Ltd. Method and apparatus for outputting information
CN111883121A (en) * 2020-07-20 2020-11-03 北京声智科技有限公司 Awakening method and device and electronic equipment
CN112233655A (en) * 2020-09-28 2021-01-15 上海声瀚信息科技有限公司 Neural network training method for improving voice command word recognition performance
CN112669818A (en) * 2020-12-08 2021-04-16 北京地平线机器人技术研发有限公司 Voice wake-up method and device, readable storage medium and electronic equipment
US20220293088A1 (en) * 2021-03-12 2022-09-15 Samsung Electronics Co., Ltd. Method of generating a trigger word detection model, and an apparatus for the same
CN113113007A (en) * 2021-03-30 2021-07-13 北京金山云网络技术有限公司 Voice data processing method and device, electronic equipment and storage medium
US11967322B2 (en) 2021-05-06 2024-04-23 Samsung Electronics Co., Ltd. Server for identifying false wakeup and method for controlling the same
WO2023085699A1 (en) * 2021-11-10 2023-05-19 삼성전자주식회사 Electronic device and control method thereof

Also Published As

Publication number Publication date
EP3579227A1 (en) 2019-12-11
CN107358951A (en) 2017-11-17
SG11201906576WA (en) 2019-08-27
JP2020517977A (en) 2020-06-18
US10748524B2 (en) 2020-08-18
KR20190134594A (en) 2019-12-04
US20200168207A1 (en) 2020-05-28
TWI692751B (en) 2020-05-01
TW201905897A (en) 2019-02-01
WO2019001428A1 (en) 2019-01-03
ES2878137T3 (en) 2021-11-18
JP6877558B2 (en) 2021-05-26
EP3579227B1 (en) 2021-06-09
PL3579227T3 (en) 2021-10-18
EP3579227A4 (en) 2020-02-26
PH12019501674A1 (en) 2020-06-01
KR102181836B1 (en) 2020-11-25

Similar Documents

Publication Publication Date Title
US10748524B2 (en) Speech wakeup method, apparatus, and electronic device
Ravanelli et al. Light gated recurrent units for speech recognition
US11423233B2 (en) On-device projection neural networks for natural language understanding
US20230410796A1 (en) Encoder-decoder models for sequence to sequence mapping
US11450312B2 (en) Speech recognition method, apparatus, and device, and storage medium
US11501154B2 (en) Sensor transformation attention network (STAN) model
US11158305B2 (en) Online verification of custom wake word
Lee et al. High-level feature representation using recurrent neural network for speech emotion recognition
KR102323046B1 (en) Speech emotion detection method and apparatus, computer device and storage medium
KR102167719B1 (en) Method and apparatus for training language model, method and apparatus for recognizing speech
US9728183B2 (en) System and method for combining frame and segment level processing, via temporal pooling, for phonetic classification
Dahl et al. Context-dependent pre-trained deep neural networks for large-vocabulary speech recognition
US9235799B2 (en) Discriminative pretraining of deep neural networks
CN108885870A (en) For by combining speech to TEXT system with speech to intention system the system and method to realize voice user interface
US11423884B2 (en) Device with convolutional neural network for acquiring multiple intent words, and method thereof
US20200152179A1 (en) Time-frequency convolutional neural network with bottleneck architecture for query-by-example processing
CN111833866A (en) Method and system for high accuracy key phrase detection for low resource devices
KR20220130565A (en) Keyword detection method and apparatus thereof
JP7178394B2 (en) Methods, apparatus, apparatus, and media for processing audio signals
US20230031733A1 (en) Method for training a speech recognition model and method for speech recognition
Zhang et al. Wake-up-word spotting using end-to-end deep neural network system
KR20200120595A (en) Method and apparatus for training language model, method and apparatus for recognizing speech
US20230045790A1 (en) Sensor transformation attention network (stan) model
Bovbjerg Self-supervised Keyword Spotting
CN114627860A (en) Model training method, voice processing method, device, equipment and medium

Legal Events

Date Code Title Description
AS Assignment

Owner name: ALIBABA GROUP HOLDING LIMITED, CAYMAN ISLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:WANG, ZHIMING;ZHOU, JUN;LI, XIAOLONG;SIGNING DATES FROM 20190806 TO 20190808;REEL/FRAME:050382/0704

STPP Information on status: patent application and granting procedure in general

Free format text: DOCKETED NEW CASE - READY FOR EXAMINATION

AS Assignment

Owner name: ADVANTAGEOUS NEW TECHNOLOGIES CO., LTD., CAYMAN ISLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ALIBABA GROUP HOLDING LIMITED;REEL/FRAME:053713/0665

Effective date: 20200826

AS Assignment

Owner name: ADVANCED NEW TECHNOLOGIES CO., LTD., CAYMAN ISLANDS

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNOR:ADVANTAGEOUS NEW TECHNOLOGIES CO., LTD.;REEL/FRAME:053761/0338

Effective date: 20200910

STPP Information on status: patent application and granting procedure in general

Free format text: NON FINAL ACTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE TO NON-FINAL OFFICE ACTION ENTERED AND FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: FINAL REJECTION MAILED

STPP Information on status: patent application and granting procedure in general

Free format text: RESPONSE AFTER FINAL ACTION FORWARDED TO EXAMINER

STPP Information on status: patent application and granting procedure in general

Free format text: ADVISORY ACTION MAILED

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION