WO2021012734A1 - Audio separation method and apparatus, electronic device and computer-readable storage medium - Google Patents
Audio separation method and apparatus, electronic device and computer-readable storage medium Download PDFInfo
- Publication number
- WO2021012734A1 WO2021012734A1 PCT/CN2020/086757 CN2020086757W WO2021012734A1 WO 2021012734 A1 WO2021012734 A1 WO 2021012734A1 CN 2020086757 W CN2020086757 W CN 2020086757W WO 2021012734 A1 WO2021012734 A1 WO 2021012734A1
- Authority
- WO
- WIPO (PCT)
- Prior art keywords
- voice
- speech
- preset
- text
- voiceprint
- Prior art date
Links
- 238000000926 separation method Methods 0.000 title claims abstract description 74
- 238000013145 classification model Methods 0.000 claims abstract description 52
- 238000001914 filtration Methods 0.000 claims abstract description 36
- 238000000034 method Methods 0.000 claims abstract description 33
- 238000012545 processing Methods 0.000 claims abstract description 21
- 238000012549 training Methods 0.000 claims description 48
- 230000007613 environmental effect Effects 0.000 claims description 29
- 238000004590 computer program Methods 0.000 claims description 24
- 239000013598 vector Substances 0.000 claims description 22
- 238000010801 machine learning Methods 0.000 claims description 19
- 238000013136 deep learning model Methods 0.000 claims description 17
- 230000008569 process Effects 0.000 claims description 17
- 238000004422 calculation algorithm Methods 0.000 claims description 16
- 238000012216 screening Methods 0.000 claims description 12
- 238000012795 verification Methods 0.000 claims description 10
- 230000009467 reduction Effects 0.000 claims description 9
- 230000000717 retained effect Effects 0.000 claims description 9
- 230000003595 spectral effect Effects 0.000 claims description 8
- NGVDGCNFYWLIFO-UHFFFAOYSA-N pyridoxal 5'-phosphate Chemical compound CC1=NC=C(COP(O)(O)=O)C(C=O)=C1O NGVDGCNFYWLIFO-UHFFFAOYSA-N 0.000 claims description 7
- 238000001228 spectrum Methods 0.000 claims description 4
- 239000002245 particle Substances 0.000 description 28
- 230000006870 function Effects 0.000 description 13
- 238000013135 deep learning Methods 0.000 description 12
- 238000000605 extraction Methods 0.000 description 8
- 239000011159 matrix material Substances 0.000 description 8
- 238000010586 diagram Methods 0.000 description 7
- 230000003321 amplification Effects 0.000 description 6
- 238000003199 nucleic acid amplification method Methods 0.000 description 6
- 238000005516 engineering process Methods 0.000 description 4
- 238000012360 testing method Methods 0.000 description 4
- 230000003044 adaptive effect Effects 0.000 description 2
- 238000004458 analytical method Methods 0.000 description 2
- 238000012790 confirmation Methods 0.000 description 2
- 238000009432 framing Methods 0.000 description 2
- 238000003062 neural network model Methods 0.000 description 2
- 210000000056 organ Anatomy 0.000 description 2
- 239000004065 semiconductor Substances 0.000 description 2
- 230000001755 vocal effect Effects 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 230000010365 information processing Effects 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 230000003287 optical effect Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N20/00—Machine learning
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/02—Feature extraction for speech recognition; Selection of recognition unit
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/14—Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
- G10L15/142—Hidden Markov Models [HMMs]
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/26—Speech to text systems
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L17/00—Speaker identification or verification techniques
- G10L17/04—Training, enrolment or model building
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0208—Noise filtering
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
- G10L21/00—Speech or voice signal processing techniques to produce another audible or non-audible signal, e.g. visual or tactile, in order to modify its quality or its intelligibility
- G10L21/02—Speech enhancement, e.g. noise reduction or echo cancellation
- G10L21/0272—Voice signal separating
- G10L21/028—Voice signal separating using properties of sound source
Definitions
- This application relates to the field of speech processing, and in particular to an audio separation method, device, electronic equipment, and computer-readable storage medium.
- the recognized text contains the content of multiple people's speech.
- the inventor realizes that it is impossible to distinguish who said the text, which affects the recognition effect and accuracy.
- the first aspect of the present application provides an audio separation method, the method including:
- Extract voiceprint feature data from the filtered voice input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result.
- the second aspect of the application provides an audio separation device, which includes:
- Acquisition module used to acquire voice
- a noise filtering module for performing noise filtering on the voice
- the voice separation module is used to extract voiceprint feature data from the filtered voice, input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and classify the same voiceprint feature data according to the classification result
- the corresponding speech is encoded and stored as a separate speech file to separate the speech;
- the text recognition module is used to recognize the speech after the separation process to obtain the recognized text of the speech.
- a third aspect of the present application provides an electronic device including a processor configured to implement the audio separation method when executing a computer program stored in a memory, and the audio separation method includes:
- Extract voiceprint feature data from the filtered voice input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result.
- a fourth aspect of the present application provides a computer-readable storage medium on which a computer program is stored, and when the computer program is executed by a processor, the audio separation method is implemented, and the audio separation method includes:
- Extract voiceprint feature data from the filtered voice input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result.
- This application uses a preset voice classification model to separate the filtered voice according to the voiceprint characteristics of the voice, and to recognize the voice after the separation process to obtain the recognized text of the voice, which can identify different people in the voice
- the speech text of spoken words improves the accuracy of speech recognition.
- Fig. 1 is a flowchart of an audio separation method in an embodiment of the present application.
- Fig. 2 is a schematic diagram of an application environment of an audio separation method in an embodiment of the present application.
- Fig. 3 is a schematic diagram of a page audio separation device in an embodiment of the present application.
- Fig. 4 is a schematic diagram of an electronic device in an embodiment of the present application.
- the audio separation method of the present application is applied to one or more electronic devices.
- the electronic device is a device that can automatically perform numerical calculation and/or information processing in accordance with pre-set or stored instructions. Its hardware includes, but is not limited to, a microprocessor and an application specific integrated circuit (ASIC) , Field-Programmable Gate Array (FPGA), Digital Processor (Digital Signal Processor, DSP), embedded equipment, etc.
- ASIC application specific integrated circuit
- FPGA Field-Programmable Gate Array
- DSP Digital Processor
- embedded equipment etc.
- the electronic device may be a computing device such as a desktop computer, a notebook computer, a tablet computer, and a cloud server.
- the device can interact with the user through a keyboard, a mouse, a remote control, a touch panel, or a voice control device.
- Fig. 1 is a flowchart of an audio separation method in an embodiment of the present application. According to different needs, the order of the steps in the flowchart can be changed, and some steps can be omitted.
- the audio separation method specifically includes the following steps:
- Step S11 Acquire voice.
- FIG. 2 shows an application environment diagram of an audio separation method in an embodiment of this application.
- the method is applied in a terminal device 1.
- the terminal device 1 includes a voice acquisition unit 11.
- the terminal device 1 acquires voice through the voice acquiring unit 11.
- the voice acquisition unit 11 may include, but is not limited to, an electric, capacitive, piezoelectric, electromagnetic, and semiconductor microphone.
- the terminal device 1 can receive a voice sent by an external device 2 communicatively connected with the terminal device 1.
- the terminal device 1 obtains voice from the storage device of the terminal device 1.
- Step S12 Perform noise filtering on the voice.
- the terminal device 1 filters the environmental noise in the voice. For example, when the terminal device 1 acquires voice through the voice acquiring unit 11 from a noisy environment, the voice includes environmental noise of the environment where the terminal device 1 is located. In a specific embodiment, the terminal device 1 detects whether the voice decibel of the acquired voice is within a preset decibel threshold range, and when the voice decibel of the voice is not within the preset decibel threshold range, the terminal device 1 The speech is noise filtered.
- the preset decibel threshold can be set as required, and in this embodiment, the preset decibel threshold range can be set to 70-80db.
- the terminal device 1 selects the voice information whose voice decibel exceeds the first decibel threshold as the environmental noise from the voice, and deletes the environmental noise whose voice decibel exceeds the first decibel threshold, so as to realize the Voice noise filtering.
- the first decibel threshold can be set as required, for example, the first decibel threshold can be set to 80db.
- the terminal device 1 filters the environmental noise in the speech through a deep learning voiceprint noise reduction method.
- the voiceprint noise reduction method through deep learning to filter the environmental noise in the voice includes: establishing a machine learning and deep learning model; establishing a voiceprint recognition model; and passing the acquired voice through Machine learning and deep learning models are used for learning, and the environmental noise in the speech is recognized and distinguished; the speech after recognition by the machine learning and deep learning models is filtered to remove the speech that does not belong to human speech audio Ambient noise, get the voice that has undergone preliminary screening; determine whether the voice that has undergone the preliminary screening reaches the preset threshold; when it is determined that the voice that has undergone the preliminary screening reaches the preset threshold, the voice that has reached the preset threshold and the voiceprint
- the recognition model is compared and extracted, the voice frequency and the spectral image that are consistent with the voiceprint recognition model are retained, the voices that are not consistent with the voiceprint recognition model are eliminated, and the voice processed by voiceprint noise reduction is obtained.
- the terminal device 1 uses a large amount of obtained speech environment audio and a large amount of specific person's speech audio to build a machine learning and deep learning model; all the speech environment audio and specific person's speech audio are converted into the form of pop chart It is imported into the terminal device 1, and through a large amount of repeated training, the environmental noise (environmental sound) is distinguished from the voice pop chart of a specific person's speech through machine learning and deep learning.
- each individual’s unique voiceprint can be observed using the grammar.
- Acquire the voiceprint of a specific speaker perform feature extraction operation on the voiceprint of the person, use the existing voiceprint of the specific speaker to build a voiceprint spectrogram, and perform the features on the voiceprint spectrogram
- a voiceprint recognition model that only belongs to the person can be established.
- the modeling methods of voiceprint recognition models are divided into three types: text-related, text-independent and text prompt. Since the input voice content cannot be determined, the text-independent type is selected for voiceprint modeling, thereby obtaining the voiceprint recognition model.
- the text irrelevant includes: GMM-UBM, GMM-SVM, GMM-UBM-LF, i-vector/PLDA).
- GMM-UBM is selected to build the voiceprint modeling of the speaker confirmation system.
- the MFCC feature vector is extracted, and after a large amount of human voiceprint data is repeatedly trained and MAP adaptive Process and confirm the decision to obtain a human voiceprint recognition model with a high voiceprint recognition rate.
- the MFCC feature vector extraction process includes input sample speech, the sample speech pre-emphasis, framing, and windowing, the processed sample speech is subjected to Fourier transform, Mel frequency filtering, and Log logarithmic energy , Calculate the cepstrum of the sample and output the MFCC image.
- the terminal device 1 filters out white noise in the speech.
- white noise refers to noise with equal noise energy contained in frequency bands of equal bandwidth within a wider frequency range.
- the white noise in the speech can be removed by the wavelet transform algorithm or the Kalman filter algorithm.
- Step S13 Extract voiceprint feature data from the filtered voice, input the voiceprint feature data into the preset voice classification model for classification to obtain a classification result, and assign the same voiceprint feature data to the corresponding
- the speech is encoded and stored as a separate speech file, and the speech is processed separately.
- the voiceprint feature can be used to verify the speaker's identity and distinguish the speaker's voice.
- the voiceprint feature data includes, but is not limited to, Mel cepstrum coefficient MFCC, perceptual linear prediction coefficient PLP, depth feature Deep Feature, and energy regularization spectrum coefficient PNCC.
- the terminal device 1 uses wavelet transform technology to extract the Mel cepstrum coefficient MFCC, the perceptual linear prediction coefficient PLP, the depth feature Deep Feature, or the energy normalized spectral coefficient PNCC, etc.
- the voiceprint feature data is input into the preset voice classification model to obtain the classification result.
- the voice corresponding to the same voiceprint feature data is encoded and stored as a separate voice file.
- the preset speech classification model includes at least one of the following: a vector machine model, a random model, and a neural network model.
- the terminal device uses a pre-trained preset voice classification model to determine the category of the voiceprint feature data according to the extracted voiceprint feature data.
- the categories of the voiceprint feature data include: a first voiceprint feature category, a second voiceprint feature category, and a third voiceprint feature category.
- the training process of the preset voice classification model includes:
- the voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample are randomly divided into a training set with a first preset ratio and a verification set with a second preset ratio, and the training set is used to train the Preset a speech classification model, and use the verification set to verify the accuracy of the preset speech classification model after training.
- the training samples in the training set with different voiceprint characteristics are distributed to different folders.
- the training samples of the first voiceprint feature category are distributed to the first folder
- the training samples of the second voiceprint feature category are distributed to the second folder
- the training samples of the third voiceprint feature category are distributed to the third folder.
- Folder extract the training samples of the first preset ratio (for example, 70%) from different folders as the total training samples to train the preset voice classification model, and take the remaining second samples from different folders.
- a preset proportion (for example, 30%) of training samples is used as a total test sample to verify the accuracy of the preset voice classification model after training.
- the training ends, and the trained preset voice classification model is used as a classifier to identify the category of the voiceprint feature data; if the accuracy rate is less than When the accuracy rate is preset, the number of positive samples and the number of negative samples are increased to retrain the preset voice classification model until the accuracy rate is greater than or equal to the preset accuracy rate.
- the terminal device 1 is also used to perform enhanced amplification processing on the voice corresponding to the same voiceprint feature data; and encode the voice after the enhanced amplification processing. That is, the terminal device 1 separates the voices with different voiceprint features from the voice according to the voiceprint features, respectively strengthens the amplification process of the separated voices, and encodes the voices corresponding to the same voiceprint features. Stored as a separate voice file, and stored separately as a voice file.
- Step S14 Recognizing the speech after the separation process to obtain the recognized text of the speech.
- the terminal device 1 converts the separated speech into text through speech recognition, and uses it as the initial speech recognition text; and matches the initial speech recognition text with a preset text database to obtain the matched Speech recognition text.
- the specific process of the terminal device 1 converting the separated speech into text through speech recognition includes:
- the grammar rule is the Viterbi algorithm.
- the voice to be recognized is "Hello", which is transformed into a 39-dimensional acoustic feature vector after feature extraction, and multiple corresponding sub-words /n//i//h//ao are obtained through multiple HMM phoneme models /, Splice multiple sub-words into characters according to the preset pronunciation dictionary, such as you, Nepal; good, number. Decode by Viterbi algorithm to obtain the optimal sequence "Hello" and output the text.
- At least two text databases may be preset, for example, a first text database and a second text database.
- the first text database can be dedicated to storing multiple modal particles, such as “um”, “ah”, “yeah”, etc. The modal particles have nothing to do with the content of the meeting and easily affect the readability of the speech converted into text.
- the second text database can be dedicated to storing multiple professional words and their corresponding pinyin, such as "feature vector”, “feature matrix”, “tensor analysis”, etc. The professional words are more complex, so they tend to appear in batches during the process of speech recognition error.
- a third text database can also be preset according to the actual situation, specifically for storing sentences such as names or place names. This article does not make specific restrictions on the number of pre-set databases and corresponding contents of this article.
- the terminal device 1 matching the initial voice recognition text with a preset text database specifically includes:
- the matching the initial speech recognition text with a preset first text database includes: determining whether there is a first word in the initial speech recognition text that matches a word in the preset first text database; When it is determined that there is a first word that matches a word in the preset first text database in the initial voice recognition text, the first word that matches in the initial voice recognition text is processed.
- the processing of the matching first word in the initial speech recognition text may further include: judging whether the matching first word is based on the pre-trained modal particle model based on the deep learning network Is the modal particle to be deleted; when it is determined that the first matching word is the modal particle to be deleted, the first matching word in the initial speech recognition text is eliminated; when the matching first word is determined When a word is not a modal particle to be deleted, the first matching word in the initial speech recognition text is retained.
- the initial speech recognition text is "this is very easy to use”
- the modal word "this” is stored in the preset first text database
- the initial speech recognition text is matched with the preset first text database to determine The matched word is "this”, and then judge whether the matched first word "this” is the modal particle to be deleted according to the pre-trained modal particle model based on the deep learning network.
- the network's modal particle model determines that the matched first word "this” does not belong to the modal particle to be deleted in "this is very useful”, then the first matching word in the initial speech recognition text is retained, The first matching result obtained is "This is pretty easy to use”.
- the initial speech recognition text is "this, we are going to have a meeting”
- the first text database is preset to store the modal word "this”
- the initial speech recognition text is matched with the preset first text database to determine The matched word is "this”, and then judge whether the matched first word "this” is the modal particle to be deleted according to the pre-trained modal particle model based on the deep learning network.
- the network's modal particle model determines that the first matching word "this” belongs to the modal particle to be deleted in "this, we are going to have a meeting”, and then the first matching word in the initial speech recognition text is eliminated.
- the first matching result obtained was "We are going to have a meeting.”
- the matching the first matching result with a preset second text database includes:
- the first matching result is "this is an original giant earthquake", and the words in the first matching result are converted to the first pinyin as "zhe shi yige yuanshi Juzhen";
- the second text database is preset to store professional words "Matrix” and the corresponding second pinyin "juzheng”, when it is determined that there is a second pinyin identical to the first pinyin in the preset second text database, the word “juzheng” corresponding to the second pinyin " "Matrix” is extracted as the word corresponding to the first pinyin "juzheng", and the second matching result obtained is "This is an original matrix".
- This application converts the separated speech into text through speech recognition technology, as the initial speech recognition text; and matches the initial speech recognition text with a preset text database to obtain the matched speech recognition text, which can be recognized
- the voice text of the words spoken by different people in the voice is convenient for the recorder to gather information.
- FIG. 3 is a schematic diagram of an audio separation device 40 in an embodiment of the application.
- the audio separation device 40 runs in an electronic device.
- the audio separation device 40 may include multiple functional modules composed of program code segments.
- the program code of each program segment in the audio separation device 40 can be stored in a memory and executed by at least one processor to perform the audio separation function.
- the audio separation device 40 can be divided into multiple functional modules according to the functions it performs.
- the audio separation device 40 may include an acquisition module 401, a noise filtering module 402, a speech separation module 403, and a text recognition module 404.
- the module referred to in this application refers to a series of computer program segments that can be executed by at least one processor and can complete fixed functions, and are stored in a memory. In some embodiments, the functions of each module will be detailed in subsequent embodiments.
- the acquiring module 401 is used for acquiring voice.
- the acquisition module 401 acquires voice through the voice acquisition unit 11.
- the voice acquisition unit 11 may include, but is not limited to, an electric, capacitive, piezoelectric, electromagnetic, and semiconductor microphone.
- the acquisition module 401 can receive the voice sent by the external device 2 communicatively connected with the terminal device 1.
- the acquiring module 401 acquires the voice from the storage device of the terminal device 1.
- the noise filtering module 402 is configured to perform noise filtering on the speech.
- the noise filtering module 402 filters the environmental noise in the speech.
- the noise filtering module 402 detects whether the voice decibel of the acquired voice is within a preset decibel threshold range, and when the voice decibel of the voice is not within the preset decibel threshold range, the noise filtering module 402 Perform noise filtering on the voice.
- the preset decibel threshold can be set as required. In this embodiment, the preset decibel threshold range can be set to 70-80db.
- the noise filtering module 402 selects the voice information whose voice decibel exceeds the first decibel threshold as the environmental noise from the voice, and deletes the environmental noise whose voice decibel exceeds the first decibel threshold, so as to realize the The noise filtering of the speech.
- the first decibel threshold can be set as required, for example, the first decibel threshold can be set to 80db.
- the noise filtering module 402 filters the environmental noise in the speech by using a deep learning voiceprint noise reduction method.
- the voiceprint noise reduction method through deep learning to filter the environmental noise in the voice includes: establishing a machine learning and deep learning model; establishing a voiceprint recognition model; and passing the acquired voice through Machine learning and deep learning models are used for learning, and the environmental noise in the speech is recognized and distinguished; the speech after recognition by the machine learning and deep learning models is filtered to remove the speech that does not belong to human speech audio Ambient noise, get the voice that has undergone preliminary screening; determine whether the voice that has undergone the preliminary screening reaches the preset threshold; when it is determined that the voice that has undergone the preliminary screening reaches the preset threshold, the voice that has reached the preset threshold and the voiceprint
- the recognition model is compared and extracted, the voice frequency and the spectral image that are consistent with the voiceprint recognition model are retained, the voices that are not consistent with the voiceprint recognition model are eliminated, and the voice processed by voiceprint noise reduction is obtained.
- the noise filtering module 402 uses a large amount of obtained speech environment audio and a large amount of specific person's speech audio to build a machine learning and deep learning model; converts all the speech environment audio and specific person's speech audio into a pop chart.
- the format is imported into the terminal device 1, and through a lot of repeated training, the environmental noise (environmental sound) and the voice pop chart of a specific person's speech are distinguished through machine learning and deep learning.
- each individual’s unique voiceprint can be observed using the grammar.
- Acquire the voiceprint of a specific speaker perform feature extraction operation on the voiceprint of the person, use the existing voiceprint of the specific speaker to build a voiceprint spectrogram, and perform the features on the voiceprint spectrogram
- a voiceprint recognition model that only belongs to the person can be established.
- the modeling methods of voiceprint recognition models are divided into three types: text-related, text-independent and text prompt. Since the input voice content cannot be determined, the text-independent type is selected for voiceprint modeling, thereby obtaining the voiceprint recognition model.
- the text irrelevant includes: GMM-UBM, GMM-SVM, GMM-UBM-LF, i-vector/PLDA).
- GMM-UBM is selected to build the voiceprint modeling of the speaker confirmation system.
- the MFCC feature vector is extracted, and after a large amount of human voiceprint data is repeatedly trained and MAP adaptive Process and confirm the decision to obtain a human voiceprint recognition model with a high voiceprint recognition rate.
- the MFCC feature vector extraction process includes input sample speech, the sample speech pre-emphasis, framing, and windowing, the processed sample speech is subjected to Fourier transform, Mel frequency filtering, and Log logarithmic energy , Calculate the cepstrum of the sample and output the MFCC image.
- the noise filtering module 402 filters white noise in the speech.
- white noise refers to noise with equal noise energy contained in frequency bands of equal bandwidth within a wider frequency range.
- the white noise in the speech can be removed by the wavelet transform algorithm or the Kalman filter algorithm.
- the voice separation module 403 is configured to use a preset voice classification model to perform separation processing on the filtered voice according to the voiceprint features of the voice.
- the voice separation module 403 uses a preset voice classification model to separate the filtered voice according to the voiceprint features of the voice, including: extracting voiceprint feature data from the filtered voice, and dividing the voiceprint
- the feature data is input into the preset voice classification model for classification to obtain the classification result.
- the voice corresponding to the same voiceprint feature data is coded and stored as a separate voice file, thus realizing the separation processing of the voice .
- the voiceprint feature can be used to verify the speaker's identity and distinguish the speaker's voice.
- the voiceprint feature data includes, but is not limited to, Mel cepstrum coefficient MFCC, perceptual linear prediction coefficient PLP, depth feature Deep Feature, and energy regularization spectrum coefficient PNCC.
- the voice separation module 403 uses wavelet transform technology to extract the Mel cepstrum coefficient MFCC, the perceptual linear prediction coefficient PLP, the depth feature Deep Feature or the energy normalized spectral coefficient PNCC, etc.
- Voiceprint feature data and input the voiceprint feature data into the preset voice classification model to obtain the classification result according to the Mel cepstrum coefficient MFCC, the perceptual linear prediction coefficient PLP, the depth feature Deep Feature or the energy normalized spectrum coefficient PNCC.
- the voice corresponding to the same voiceprint feature data is encoded and stored as a separate voice file.
- the preset speech classification model includes at least one of the following: a vector machine model, a random model, and a neural network model.
- the terminal device uses a pre-trained preset voice classification model to determine the category of the voiceprint feature data according to the extracted voiceprint feature data.
- the categories of the voiceprint feature data include: a first voiceprint feature category, a second voiceprint feature category, and a third voiceprint feature category.
- the training process of inputting the voiceprint feature data into the preset voice classification model for classification to obtain the classification result includes:
- the voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample are randomly divided into a training set with a first preset ratio and a verification set with a second preset ratio, and the training set is used to train the Preset a speech classification model, and use the verification set to verify the accuracy of the preset speech classification model after training.
- the training samples in the training set with different voiceprint characteristics are distributed to different folders.
- the training samples of the first voiceprint feature category are distributed to the first folder
- the training samples of the second voiceprint feature category are distributed to the second folder
- the training samples of the third voiceprint feature category are distributed to the third folder.
- Folder extract the training samples of the first preset ratio (for example, 70%) from different folders as the total training samples to train the preset voice classification model, and take the remaining second samples from different folders.
- a preset proportion (for example, 30%) of training samples is used as a total test sample to verify the accuracy of the preset voice classification model after training.
- the training ends, and the trained preset voice classification model is used as a classifier to identify the category of the voiceprint feature data; if the accuracy rate is less than When the accuracy rate is preset, the number of positive samples and the number of negative samples are increased to retrain the preset voice classification model until the accuracy rate is greater than or equal to the preset accuracy rate.
- the voice separation module 403 is also used to perform enhanced amplification processing on the voice corresponding to the same voiceprint feature data; and encode the voice after the enhanced amplification processing. That is, the terminal device 1 separates the voices with different voiceprint features from the voice according to the voiceprint features, respectively strengthens the amplification process of the separated voices, and encodes the voices corresponding to the same voiceprint features. Stored as a separate voice file, and stored separately as a voice file.
- the text recognition module 404 is configured to recognize the speech after the separation process to obtain the recognized text of the speech.
- the text recognition module 404 converts the separated speech into text through speech recognition, as the initial speech recognition text; and matches the initial speech recognition text with a preset text database to obtain the matching Speech recognition text.
- the specific process for the text recognition module 404 to convert the separated speech into text through speech recognition includes:
- the grammar rule is the Viterbi algorithm.
- the voice to be recognized is "Hello", which is transformed into a 39-dimensional acoustic feature vector after feature extraction, and multiple corresponding sub-words /n//i//h//ao are obtained through multiple HMM phoneme models /, Splice multiple sub-words into characters according to the preset pronunciation dictionary, such as you, Nepal; good, number. Decode by Viterbi algorithm to obtain the optimal sequence "Hello" and output the text.
- At least two text databases may be preset, for example, a first text database and a second text database.
- the first text database can be dedicated to storing multiple modal particles, such as “um”, “ah”, “yeah”, etc. The modal particles have nothing to do with the content of the meeting and easily affect the readability of the speech converted into text.
- the second text database can be dedicated to storing multiple professional words and their corresponding pinyin, such as "feature vector”, “feature matrix”, “tensor analysis”, etc. The professional words are more complex, so they tend to appear in batches during the process of speech recognition error.
- a third text database can also be preset according to the actual situation, specifically for storing sentences such as names or place names. This article does not make specific restrictions on the number of pre-set text databases and corresponding content.
- the text recognition module 404 matching the initial speech recognition text with a preset text database specifically includes:
- the matching the initial speech recognition text with a preset first text database includes: determining whether there is a first word in the initial speech recognition text that matches a word in the preset first text database; When it is determined that there is a first word that matches a word in the preset first text database in the initial voice recognition text, the first word that matches in the initial voice recognition text is processed.
- the processing of the matching first word in the initial speech recognition text may further include: judging whether the matching first word is based on the pre-trained modal particle model based on the deep learning network Is the modal particle to be deleted; when it is determined that the first matching word is the modal particle to be deleted, the first matching word in the initial speech recognition text is eliminated; when the matching first word is determined When a word is not a modal particle to be deleted, the first matching word in the initial speech recognition text is retained.
- the initial speech recognition text is "this is very easy to use”
- the modal word "this” is stored in the preset first text database
- the initial speech recognition text is matched with the preset first text database to determine The matched word is "this”, and then judge whether the matched first word "this” is the modal particle to be deleted according to the pre-trained modal particle model based on the deep learning network.
- the network's modal particle model determines that the matched first word "this” does not belong to the modal particle to be deleted in "this is very useful”, then the first matching word in the initial speech recognition text is retained, The first matching result obtained is "This is pretty easy to use”.
- the initial speech recognition text is "this, we are going to have a meeting”
- the first text database is preset to store the modal word "this”
- the initial speech recognition text is matched with the preset first text database to determine The matched word is "this”, and then judge whether the matched first word "this” is the modal particle to be deleted according to the pre-trained modal particle model based on the deep learning network.
- the network's modal particle model determines that the first matching word "this” belongs to the modal particle to be deleted in "this, we are going to have a meeting”, and then the first matching word in the initial speech recognition text is eliminated.
- the first matching result obtained was "We are going to have a meeting.”
- the matching the first matching result with a preset second text database includes:
- the first matching result is "this is an original giant earthquake", and the words in the first matching result are converted to the first pinyin as "zhe shi yige yuanshi Juzhen";
- the second text database is preset to store professional words "Matrix” and the corresponding second pinyin "juzheng”, when it is determined that there is a second pinyin identical to the first pinyin in the preset second text database, the word “juzheng” corresponding to the second pinyin " "Matrix” is extracted as the word corresponding to the first pinyin "juzheng", and the second matching result obtained is "This is an original matrix".
- This application converts the separated speech into text through speech recognition technology, as the initial speech recognition text; and matches the initial speech recognition text with a preset text database to obtain the matched speech recognition text, which can be recognized
- the voice text of the words spoken by different people in the voice is convenient for the recorder to gather information.
- FIG. 4 is a schematic diagram of a preferred embodiment of the electronic device 7 of this application.
- the electronic device 7 includes a memory 71, a processor 72, and a computer program 73 that is stored in the memory 71 and can run on the processor 72.
- the steps in the above audio separation method embodiment are implemented, such as steps S11 to S14 shown in FIG. 1. That is, the audio separation method includes: acquiring speech; performing noise filtering on the speech; extracting voiceprint feature data from the filtered voice, and inputting the voiceprint feature data into a preset voice classification model for classification to obtain a classification result According to the classification result, the voice corresponding to the same voiceprint feature data is encoded and stored as a separate voice file to separate the voice; and the voice after the separation is recognized to obtain the information of the voice Recognize the text.
- the functions of the modules/units in the foregoing audio separation device embodiment are implemented, for example, the modules 401 to 404 in FIG. 3.
- the computer program 73 may be divided into one or more modules/units, and the one or more modules/units are stored in the memory 71 and executed by the processor 72 to complete This application.
- the one or more modules/units may be a series of computer program instruction segments capable of completing specific functions, and the instruction segments are used to describe the execution process of the computer program 73 in the electronic device 7.
- the computer program 73 can be divided into an acquisition module 401, a noise filtering module 402, a speech separation module 403, and a text recognition module 404 in FIG. 3.
- the computer program 73 can be divided into an acquisition module 401, a noise filtering module 402, a speech separation module 403, and a text recognition module 404 in FIG. 3.
- the second embodiment For specific functions of each module, refer to the second embodiment.
- the electronic device 7 and the terminal device 1 are the same device.
- the electronic device 7 may be a computing device such as a desktop computer, a notebook, a palmtop computer, and a cloud server.
- the schematic diagram is only an example of the electronic device 7 and does not constitute a limitation on the electronic device 7. It may include more or less components than those shown in the figure, or a combination of certain components, or different components. Components, for example, the electronic device 7 may also include input and output devices, network access devices, buses, and the like.
- the so-called processor 72 may be a central processing module (Central Processing Unit, CPU), other general-purpose processors, digital signal processors (Digital Signal Processor, DSP), application specific integrated circuits (Application Specific Integrated Circuit, ASIC), Ready-made programmable gate array (Field-Programmable Gate Array, FPGA) or other programmable logic devices, discrete gate or transistor logic devices, discrete hardware components, etc.
- the general-purpose processor can be a microprocessor or the processor 72 can also be any conventional processor, etc.
- the processor 72 is the control center of the electronic device 7 and connects the entire electronic device 7 with various interfaces and lines. Parts.
- the memory 71 may be used to store the computer program 73 and/or modules/units.
- the processor 72 runs or executes the computer programs and/or modules/units stored in the memory 71 and calls the computer programs and/or modules/units stored in the memory 71.
- the data in 71 realizes various functions of the electronic device 7 described above.
- the memory 71 may mainly include a storage program area and a storage data area.
- the storage program area may store an operating system, an application program required by at least one function (such as a sound playback function, an image playback function, etc.), etc.; the storage data area may The data (such as audio data, phone book, etc.) created according to the use of the electronic device 7 is stored.
- the memory 71 may include a high-speed random access memory, and may also include a non-volatile memory, such as a hard disk, a memory, a plug-in hard disk, a Smart Media Card (SMC), and a Secure Digital (SD) Card, Flash Card, at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
- a non-volatile memory such as a hard disk, a memory, a plug-in hard disk, a Smart Media Card (SMC), and a Secure Digital (SD) Card, Flash Card, at least one magnetic disk storage device, flash memory device, or other volatile solid-state storage device.
- the integrated module/unit of the electronic device 7 may be stored in a computer-readable storage medium, which may be non-easy. Loss of sex can also be volatile. Based on this understanding, this application implements all or part of the processes in the above-mentioned embodiments and methods, and can also be completed by instructing relevant hardware through a computer program.
- the computer program can be stored in a computer-readable storage medium.
- the above audio separation method can be realized, which includes: acquiring speech; filtering the speech noise; extracting voiceprint feature data from the filtered voice, and inputting the voiceprint feature data Perform classification to a preset voice classification model to obtain a classification result, encode and store the voice corresponding to the same voiceprint feature data as a separate voice file according to the classification result, and perform separation processing on the voice; and The subsequent voice is recognized to obtain the recognized text of the voice.
- the computer program includes computer program code
- the computer program code may be in the form of source code, object code, executable file, or some intermediate forms.
- the computer-readable medium may include: any entity or device capable of carrying the computer program code, recording medium, U disk, mobile hard disk, magnetic disk, optical disk, computer memory, read-only memory (ROM, Read-Only Memory) , Random Access Memory (RAM, Random Access Memory), electrical carrier signal, telecommunications signal, and software distribution media, etc.
- ROM Read-Only Memory
- RAM Random Access Memory
- electrical carrier signal telecommunications signal
- software distribution media etc.
- the content contained in the computer-readable medium can be appropriately added or deleted in accordance with the requirements of the legislation and patent practice in the jurisdiction.
- the computer-readable medium Does not include electrical carrier signals and telecommunication signals.
- the functional modules in the various embodiments of the present application may be integrated in the same processing module, or each module may exist alone physically, or two or more modules may be integrated in the same module.
- the above-mentioned integrated modules can be implemented in the form of hardware, or in the form of hardware plus software functional modules.
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Computational Linguistics (AREA)
- Signal Processing (AREA)
- Quality & Reliability (AREA)
- Software Systems (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Theoretical Computer Science (AREA)
- Artificial Intelligence (AREA)
- Medical Informatics (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Evolutionary Computation (AREA)
- Data Mining & Analysis (AREA)
- Probability & Statistics with Applications (AREA)
- Machine Translation (AREA)
Abstract
Description
Claims (20)
- 一种音频分离方法,其中,所述方法包括:An audio separation method, wherein the method includes:获取语音;Get voice对所述语音进行噪声过滤;Noise filtering the voice;从过滤后的语音中提取声纹特征数据,将所述声纹特征数据输入到预设语音分类模型进行分类得到分类结果,根据所述分类结果将相同的声纹特征数据对应的语音进行编码并存储为单独的语音文件而将所述语音进行分离处理;及Extract voiceprint feature data from the filtered voice, input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result. Store the voice as a separate voice file and separate the voice; and对经过分离处理后的语音进行识别以获取所述语音的识别文本。Recognizing the speech after the separation process to obtain the recognized text of the speech.
- 如权利要求1所述的音频分离方法,其中,所述预设语音分类模型的训练过程包括:The audio separation method according to claim 1, wherein the training process of the preset voice classification model comprises:获取正样本的声纹特征数据及负样本的声纹特征数据,并将正样本的声纹特征数据标注声纹特征类别,以使正样本的声纹特征数据携带声纹特征类别标签;Acquire the voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample, and label the voiceprint feature category of the voiceprint feature data of the positive sample, so that the voiceprint feature data of the positive sample carry the voiceprint feature category label;将所述正样本的声纹特征数据及所述负样本的声纹特征数据随机分成第一预设比例的训练集和第二预设比例的验证集,利用所述训练集训练所述预设语音分类模型,并利用所述验证集验证训练后的所述预设语音分类模型的准确率;The voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample are randomly divided into a training set of a first preset ratio and a verification set of a second preset ratio, and the training set is used to train the preset A voice classification model, and use the verification set to verify the accuracy of the preset voice classification model after training;若所述准确率大于或者等于预设准确率时,则结束训练,并以训练后的所述预设语音分类模型作为分类器识别所述声纹特征数据的类别;及If the accuracy rate is greater than or equal to the preset accuracy rate, the training is ended, and the preset voice classification model after training is used as a classifier to identify the category of the voiceprint feature data; and若所述准确率小于预设准确率时,则增加正样本数量及负样本数量以重新训练所述预设语音分类模型直至所述准确率大于或者等于预设准确率。If the accuracy rate is less than the preset accuracy rate, the number of positive samples and the number of negative samples are increased to retrain the preset voice classification model until the accuracy rate is greater than or equal to the preset accuracy rate.
- 如权利要求1所述的音频分离方法,其中,所述对所述语音进行噪声过滤包括:The audio separation method according to claim 1, wherein said performing noise filtering on said voice comprises:从所述语音中选择语音分贝超过第一分贝阈值的语音信息作为环境噪声,并将语音分贝超过第一分贝阈值的所述环境噪声进行删除。The voice information whose voice decibel exceeds the first decibel threshold is selected from the voice as environmental noise, and the environmental noise whose voice decibel exceeds the first decibel threshold is deleted.
- 如权利要求1所述的音频分离方法,其中,所述对所述语音进行噪声过滤包括:The audio separation method according to claim 1, wherein said performing noise filtering on said voice comprises:建立机器学习及深度学习模型;建立声纹识别模型;将获取的所述语音通过所述机器学习及深度学习模型进行学习,对所述语音中的环境噪声进行识别区分;将经过所述机器学习及深度学习模型识别后的语音进行过滤,剔除掉所述语音中不属于人说话音频的环境噪声,得到经过初步筛查的语音;判断经过初步筛查的语音是否达到预设阈值;当确定经过初步筛查的语音达到预设阈值时,将达到预设阈值的语音与所述声纹识别模型进行对比提取, 保留与所述声纹识别模型相符合的语音频率及语谱图像,剔除掉与所述声纹识别模型不符合的语音,得到声纹降噪处理的语音。Establish a machine learning and deep learning model; establish a voiceprint recognition model; learn the acquired voice through the machine learning and deep learning model, and identify and distinguish the environmental noise in the voice; pass the machine learning And the speech after recognition by the deep learning model is filtered, and the environmental noise in the speech that is not human speech audio is removed, and the speech that has undergone preliminary screening is obtained; it is determined whether the speech after the preliminary screening reaches the preset threshold; When the preliminarily screened voice reaches the preset threshold, the voice that reaches the preset threshold is compared and extracted with the voiceprint recognition model, and the voice frequency and spectral image that are consistent with the voiceprint recognition model are retained, and those with The voiceprint recognition model does not conform to the voice, and the voiceprint noise reduction processed voice is obtained.
- 如权利要求1所述的音频分离方法,其中,所述对经过分离处理后的语音进行识别以获取所述语音的识别文本包括:The audio separation method according to claim 1, wherein said recognizing the speech after the separation process to obtain the recognized text of the speech comprises:通过语音识别将经过分离处理后的语音转化为文本,作为初始语音识别文本;及Convert the separated speech into text through speech recognition as the initial speech recognition text; and将所述初始语音识别文本与预设文本数据库进行匹配,得到匹配后的语音识别文本。The initial speech recognition text is matched with the preset text database to obtain the matched speech recognition text.
- 如权利要求5所述的音频分离方法,其中,所述通过语音识别将经过分离处理后的语音转化为文本包括:5. The audio separation method according to claim 5, wherein said converting the separated speech into text through speech recognition comprises:提取所述语音的音频特征,转换为预设长度的声学特征向量;Extracting audio features of the voice and converting them into acoustic feature vectors of preset length;根据解码算法将所述特征向量解码成词序;Decoding the feature vector into word order according to a decoding algorithm;通过HMM音素模型得到对应词序的子词,所述子词为声母和韵母;Obtain the subwords corresponding to the word order through the HMM phoneme model, where the subwords are initials and vowels;根据预设的发音字典将多个子词拼接成文字;及Combine multiple sub-words into text according to the preset pronunciation dictionary; and使用维特比算法解码得到最优序列,得到文本。Use the Viterbi algorithm to decode the optimal sequence and get the text.
- 如权利要求1所述的音频分离方法,其中,The audio separation method according to claim 1, wherein:所述声纹特征包括梅尔倒谱系数MFCC、感知线性预测系数PLP、深度特征Deep Feature以及能量规整谱系数PNCC。The voiceprint features include Mel cepstrum coefficient MFCC, perceptual linear prediction coefficient PLP, depth feature Deep Feature, and energy regularization spectrum coefficient PNCC.
- 一种音频分离装置,其中,所述装置包括:An audio separation device, wherein the device includes:获取模块,用于获取语音;Acquisition module, used to acquire voice;噪声过滤模块,用于对所述语音进行噪声过滤;A noise filtering module for performing noise filtering on the voice;语音分离模块,用于从过滤后的语音中提取声纹特征数据,将所述声纹特征数据输入到预设语音分类模型进行分类得到分类结果,根据所述分类结果将相同的声纹特征数据对应的语音进行编码并存储为单独的语音文件而将所述语音进行分离处理;及The voice separation module is used to extract voiceprint feature data from the filtered voice, input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and classify the same voiceprint feature data according to the classification result The corresponding speech is encoded and stored as a separate speech file to separate the speech; and文本识别模块,用于对经过分离处理后的语音进行识别以获取所述语音的识别文本。The text recognition module is used to recognize the speech after the separation process to obtain the recognized text of the speech.
- 一种电子设备,其中,所述电子设备包括处理器,所述处理器用于执行存储器中存储的计算机程序时实现一种音频分离方法,所述音频分离方法包括:An electronic device, wherein the electronic device includes a processor configured to implement an audio separation method when executing a computer program stored in a memory, and the audio separation method includes:获取语音;Get voice对所述语音进行噪声过滤;Noise filtering the voice;从过滤后的语音中提取声纹特征数据,将所述声纹特征数据输入到预设语音分类模型进行分类得到分类结果,根据所述分类结果将相同的声纹特征数据对应的语音进行编码并存储为单独的语音文件而将所述语音进行分离处理;及对经过分离处理后的语音进行识别以获取所述语音的识别文本。Extract voiceprint feature data from the filtered voice, input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result. The voice is stored as a separate voice file to perform separation processing; and the voice after the separation processing is recognized to obtain the recognized text of the voice.
- 如权利要求9所述的电子设备,其中,所述预设语音分类模型的训练过 程包括:The electronic device according to claim 9, wherein the training process of the preset voice classification model comprises:获取正样本的声纹特征数据及负样本的声纹特征数据,并将正样本的声纹特征数据标注声纹特征类别,以使正样本的声纹特征数据携带声纹特征类别标签;Acquire the voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample, and label the voiceprint feature category of the voiceprint feature data of the positive sample, so that the voiceprint feature data of the positive sample carry the voiceprint feature category label;将所述正样本的声纹特征数据及所述负样本的声纹特征数据随机分成第一预设比例的训练集和第二预设比例的验证集,利用所述训练集训练所述预设语音分类模型,并利用所述验证集验证训练后的所述预设语音分类模型的准确率;The voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample are randomly divided into a training set of a first preset ratio and a verification set of a second preset ratio, and the training set is used to train the preset A voice classification model, and use the verification set to verify the accuracy of the preset voice classification model after training;若所述准确率大于或者等于预设准确率时,则结束训练,并以训练后的所述预设语音分类模型作为分类器识别所述声纹特征数据的类别;及If the accuracy rate is greater than or equal to the preset accuracy rate, the training is ended, and the preset voice classification model after training is used as a classifier to identify the category of the voiceprint feature data; and若所述准确率小于预设准确率时,则增加正样本数量及负样本数量以重新训练所述预设语音分类模型直至所述准确率大于或者等于预设准确率。If the accuracy rate is less than the preset accuracy rate, the number of positive samples and the number of negative samples are increased to retrain the preset voice classification model until the accuracy rate is greater than or equal to the preset accuracy rate.
- 如权利要求9所述的电子设备,其中,所述对所述语音进行噪声过滤包括:9. The electronic device of claim 9, wherein said performing noise filtering on said voice comprises:从所述语音中选择语音分贝超过第一分贝阈值的语音信息作为环境噪声,并将语音分贝超过第一分贝阈值的所述环境噪声进行删除。The voice information whose voice decibel exceeds the first decibel threshold is selected from the voice as environmental noise, and the environmental noise whose voice decibel exceeds the first decibel threshold is deleted.
- 如权利要求9所述的电子设备,其中,所述对所述语音进行噪声过滤包括:9. The electronic device of claim 9, wherein said performing noise filtering on said voice comprises:建立机器学习及深度学习模型;建立声纹识别模型;将获取的所述语音通过所述机器学习及深度学习模型进行学习,对所述语音中的环境噪声进行识别区分;将经过所述机器学习及深度学习模型识别后的语音进行过滤,剔除掉所述语音中不属于人说话音频的环境噪声,得到经过初步筛查的语音;判断经过初步筛查的语音是否达到预设阈值;当确定经过初步筛查的语音达到预设阈值时,将达到预设阈值的语音与所述声纹识别模型进行对比提取,保留与所述声纹识别模型相符合的语音频率及语谱图像,剔除掉与所述声纹识别模型不符合的语音,得到声纹降噪处理的语音。Establish a machine learning and deep learning model; establish a voiceprint recognition model; learn the acquired voice through the machine learning and deep learning model, and identify and distinguish the environmental noise in the voice; pass the machine learning And the speech after recognition by the deep learning model is filtered, and the environmental noise in the speech that is not human speech audio is removed, and the speech that has undergone preliminary screening is obtained; it is determined whether the speech after the preliminary screening reaches the preset threshold; When the preliminarily screened voice reaches the preset threshold, the voice that reaches the preset threshold is compared and extracted with the voiceprint recognition model, and the voice frequency and spectral image that are consistent with the voiceprint recognition model are retained, and those with The voiceprint recognition model does not conform to the voice, and the voiceprint noise reduction processed voice is obtained.
- 如权利要求9所述的电子设备,其中,所述对经过分离处理后的语音进行识别以获取所述语音的识别文本包括:9. The electronic device of claim 9, wherein the recognizing the speech after the separation process to obtain the recognized text of the speech comprises:通过语音识别将经过分离处理后的语音转化为文本,作为初始语音识别文本;及Convert the separated speech into text through speech recognition as the initial speech recognition text; and将所述初始语音识别文本与预设文本数据库进行匹配,得到匹配后的语音识别文本。The initial speech recognition text is matched with the preset text database to obtain the matched speech recognition text.
- 如权利要求13所述的电子设备,其中,所述通过语音识别将经过分离处理后的语音转化为文本包括:The electronic device according to claim 13, wherein said converting the speech after the separation process into text through speech recognition comprises:提取所述语音的音频特征,转换为预设长度的声学特征向量;Extracting audio features of the voice and converting them into acoustic feature vectors of preset length;根据解码算法将所述特征向量解码成词序;Decoding the feature vector into word order according to a decoding algorithm;通过HMM音素模型得到对应词序的子词,所述子词为声母和韵母;Obtain the subwords corresponding to the word order through the HMM phoneme model, where the subwords are initials and vowels;根据预设的发音字典将多个子词拼接成文字;及Combine multiple sub-words into text according to the preset pronunciation dictionary; and使用维特比算法解码得到最优序列,得到文本。Use the Viterbi algorithm to decode the optimal sequence and get the text.
- 一种计算机可读存储介质,其上存储有计算机程序,其中,所述计算机程序被处理器执行时实现所述音频分离方法,所述音频分离方法包括:A computer-readable storage medium having a computer program stored thereon, wherein the computer program is executed by a processor to implement the audio separation method, and the audio separation method includes:获取语音;Get voice对所述语音进行噪声过滤;Noise filtering the voice;从过滤后的语音中提取声纹特征数据,将所述声纹特征数据输入到预设语音分类模型进行分类得到分类结果,根据所述分类结果将相同的声纹特征数据对应的语音进行编码并存储为单独的语音文件而将所述语音进行分离处理;及对经过分离处理后的语音进行识别以获取所述语音的识别文本。Extract voiceprint feature data from the filtered voice, input the voiceprint feature data into a preset voice classification model for classification to obtain a classification result, and encode the voice corresponding to the same voiceprint feature data according to the classification result. The voice is stored as a separate voice file to perform separation processing; and the voice after the separation processing is recognized to obtain the recognized text of the voice.
- 如权利要求15所述的计算机可读存储介质,其中,所述预设语音分类模型的训练过程包括:15. The computer-readable storage medium of claim 15, wherein the training process of the preset speech classification model comprises:获取正样本的声纹特征数据及负样本的声纹特征数据,并将正样本的声纹特征数据标注声纹特征类别,以使正样本的声纹特征数据携带声纹特征类别标签;Acquire the voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample, and label the voiceprint feature category of the voiceprint feature data of the positive sample, so that the voiceprint feature data of the positive sample carry the voiceprint feature category label;将所述正样本的声纹特征数据及所述负样本的声纹特征数据随机分成第一预设比例的训练集和第二预设比例的验证集,利用所述训练集训练所述预设语音分类模型,并利用所述验证集验证训练后的所述预设语音分类模型的准确率;The voiceprint feature data of the positive sample and the voiceprint feature data of the negative sample are randomly divided into a training set of a first preset ratio and a verification set of a second preset ratio, and the training set is used to train the preset A voice classification model, and use the verification set to verify the accuracy of the preset voice classification model after training;若所述准确率大于或者等于预设准确率时,则结束训练,并以训练后的所述预设语音分类模型作为分类器识别所述声纹特征数据的类别;及If the accuracy rate is greater than or equal to the preset accuracy rate, the training is ended, and the preset voice classification model after training is used as a classifier to identify the category of the voiceprint feature data; and若所述准确率小于预设准确率时,则增加正样本数量及负样本数量以重新训练所述预设语音分类模型直至所述准确率大于或者等于预设准确率。If the accuracy rate is less than the preset accuracy rate, the number of positive samples and the number of negative samples are increased to retrain the preset voice classification model until the accuracy rate is greater than or equal to the preset accuracy rate.
- 如权利要求15所述的计算机可读存储介质,其中,所述对所述语音进行噪声过滤包括:15. The computer-readable storage medium of claim 15, wherein said performing noise filtering on said speech comprises:从所述语音中选择语音分贝超过第一分贝阈值的语音信息作为环境噪声,并将语音分贝超过第一分贝阈值的所述环境噪声进行删除。The voice information whose voice decibel exceeds the first decibel threshold is selected from the voice as environmental noise, and the environmental noise whose voice decibel exceeds the first decibel threshold is deleted.
- 如权利要求15所述的计算机可读存储介质,其中,所述对所述语音进行噪声过滤包括:15. The computer-readable storage medium of claim 15, wherein said performing noise filtering on said speech comprises:建立机器学习及深度学习模型;建立声纹识别模型;将获取的所述语音通过所述机器学习及深度学习模型进行学习,对所述语音中的环境噪声进行识别区分;将经过所述机器学习及深度学习模型识别后的语音进行过滤,剔除掉所述语音中不属于人说话音频的环境噪声,得到经过初步筛查的语音;判断经过初步筛查的语音是否达到预设阈值;当确定经过初步筛查的语音达到预设阈值时,将达到预设阈值的语音与所述声纹识别模型进行对比提取,保留与所述声纹识别模型相符合的语音频率及语谱图像,剔除掉与所述声纹识别模型不符合的语音,得到声纹降噪处理的语音。Establish a machine learning and deep learning model; establish a voiceprint recognition model; learn the acquired voice through the machine learning and deep learning model, and identify and distinguish the environmental noise in the voice; pass the machine learning And the speech after recognition by the deep learning model is filtered, and the environmental noise in the speech that is not human speech audio is removed, and the speech that has undergone preliminary screening is obtained; it is determined whether the speech after the preliminary screening reaches the preset threshold; When the preliminarily screened voice reaches the preset threshold, the voice that reaches the preset threshold is compared and extracted with the voiceprint recognition model, and the voice frequency and spectral image that are consistent with the voiceprint recognition model are retained, and those with The voiceprint recognition model does not conform to the voice, and the voiceprint noise reduction processed voice is obtained.
- 如权利要求15所述的计算机可读存储介质,其中,所述对经过分离处理后的语音进行识别以获取所述语音的识别文本包括:15. The computer-readable storage medium according to claim 15, wherein the recognizing the speech after the separation process to obtain the recognized text of the speech comprises:通过语音识别将经过分离处理后的语音转化为文本,作为初始语音识别文本;及Convert the separated speech into text through speech recognition as the initial speech recognition text; and将所述初始语音识别文本与预设文本数据库进行匹配,得到匹配后的语音识别文本。The initial speech recognition text is matched with the preset text database to obtain the matched speech recognition text.
- 如权利要求19所述的计算机可读存储介质,其中,所述通过语音识别将经过分离处理后的语音转化为文本包括:19. The computer-readable storage medium of claim 19, wherein said converting the separated speech into text through speech recognition comprises:提取所述语音的音频特征,转换为预设长度的声学特征向量;Extracting audio features of the voice and converting them into acoustic feature vectors of preset length;根据解码算法将所述特征向量解码成词序;Decoding the feature vector into word order according to a decoding algorithm;通过HMM音素模型得到对应词序的子词,所述子词为声母和韵母;Obtain the subwords corresponding to the word order through the HMM phoneme model, where the subwords are initials and vowels;根据预设的发音字典将多个子词拼接成文字;及Combine multiple sub-words into text according to the preset pronunciation dictionary; and使用维特比算法解码得到最优序列,得到文本。Use the Viterbi algorithm to decode the optimal sequence and get the text.
Applications Claiming Priority (2)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910678465.5A CN110473566A (en) | 2019-07-25 | 2019-07-25 | Audio separation method, device, electronic equipment and computer readable storage medium |
CN201910678465.5 | 2019-07-25 |
Publications (1)
Publication Number | Publication Date |
---|---|
WO2021012734A1 true WO2021012734A1 (en) | 2021-01-28 |
Family
ID=68508340
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
PCT/CN2020/086757 WO2021012734A1 (en) | 2019-07-25 | 2020-04-24 | Audio separation method and apparatus, electronic device and computer-readable storage medium |
Country Status (2)
Country | Link |
---|---|
CN (1) | CN110473566A (en) |
WO (1) | WO2021012734A1 (en) |
Families Citing this family (27)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110473566A (en) * | 2019-07-25 | 2019-11-19 | 深圳壹账通智能科技有限公司 | Audio separation method, device, electronic equipment and computer readable storage medium |
CN111105801B (en) * | 2019-12-03 | 2022-04-01 | 云知声智能科技股份有限公司 | Role voice separation method and device |
US11303464B2 (en) * | 2019-12-05 | 2022-04-12 | Microsoft Technology Licensing, Llc | Associating content items with images captured of meeting content |
CN113035225B (en) * | 2019-12-09 | 2023-02-28 | 中国科学院自动化研究所 | Visual voiceprint assisted voice separation method and device |
CN111081275B (en) * | 2019-12-20 | 2023-05-26 | 惠州Tcl移动通信有限公司 | Terminal processing method and device based on sound analysis, storage medium and terminal |
CN110970036B (en) * | 2019-12-24 | 2022-07-12 | 网易(杭州)网络有限公司 | Voiceprint recognition method and device, computer storage medium and electronic equipment |
CN111243620B (en) * | 2020-01-07 | 2022-07-19 | 腾讯科技(深圳)有限公司 | Voice separation model training method and device, storage medium and computer equipment |
CN111489756B (en) * | 2020-03-31 | 2024-03-01 | 中国工商银行股份有限公司 | Voiceprint recognition method and device |
CN111462754B (en) * | 2020-04-16 | 2022-08-09 | 深圳航天科创实业有限公司 | Method for establishing dispatching control voice recognition model of power system |
CN111552777B (en) * | 2020-04-24 | 2023-09-26 | 北京达佳互联信息技术有限公司 | Audio identification method and device, electronic equipment and storage medium |
CN111627457A (en) * | 2020-05-13 | 2020-09-04 | 广州国音智能科技有限公司 | Voice separation method, system and computer readable storage medium |
CN111768801A (en) * | 2020-06-12 | 2020-10-13 | 瑞声科技(新加坡)有限公司 | Airflow noise eliminating method and device, computer equipment and storage medium |
CN111785291A (en) * | 2020-07-02 | 2020-10-16 | 北京捷通华声科技股份有限公司 | Voice separation method and voice separation device |
CN111968657B (en) * | 2020-08-17 | 2022-08-16 | 北京字节跳动网络技术有限公司 | Voice processing method and device, electronic equipment and computer readable medium |
CN112084746A (en) * | 2020-09-11 | 2020-12-15 | 广东电网有限责任公司 | Entity identification method, system, storage medium and equipment |
CN112102854A (en) * | 2020-09-22 | 2020-12-18 | 福建鸿兴福食品有限公司 | Recording filtering method and device and computer readable storage medium |
CN112233694B (en) * | 2020-10-10 | 2024-03-05 | 中国电子科技集团公司第三研究所 | Target identification method and device, storage medium and electronic equipment |
CN112242137B (en) * | 2020-10-15 | 2024-05-17 | 上海依图网络科技有限公司 | Training of human voice separation model and human voice separation method and device |
CN112792849B (en) * | 2021-01-06 | 2022-07-26 | 厦门攸信信息技术有限公司 | Collision detection method, robot, mobile terminal and storage medium |
CN112634875B (en) * | 2021-03-04 | 2021-06-08 | 北京远鉴信息技术有限公司 | Voice separation method, voice separation device, electronic device and storage medium |
CN112992153B (en) * | 2021-04-27 | 2021-08-17 | 太平金融科技服务(上海)有限公司 | Audio processing method, voiceprint recognition device and computer equipment |
CN112989107B (en) * | 2021-05-18 | 2021-07-30 | 北京世纪好未来教育科技有限公司 | Audio classification and separation method and device, electronic equipment and storage medium |
CN113314144A (en) * | 2021-05-19 | 2021-08-27 | 中国南方电网有限责任公司超高压输电公司广州局 | Voice recognition and power equipment fault early warning method, system, terminal and medium |
CN113314108B (en) * | 2021-06-16 | 2024-02-13 | 深圳前海微众银行股份有限公司 | Method, apparatus, device, storage medium and program product for processing voice data |
CN113505612A (en) * | 2021-07-23 | 2021-10-15 | 平安科技(深圳)有限公司 | Multi-person conversation voice real-time translation method, device, equipment and storage medium |
CN113539292A (en) * | 2021-07-28 | 2021-10-22 | 联想(北京)有限公司 | Voice separation method and device |
CN116504246B (en) * | 2023-06-26 | 2023-11-24 | 深圳市矽昊智能科技有限公司 | Voice remote control method, device, storage medium and device based on Bluetooth device |
Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103811020A (en) * | 2014-03-05 | 2014-05-21 | 东北大学 | Smart voice processing method |
CN105719659A (en) * | 2016-02-03 | 2016-06-29 | 努比亚技术有限公司 | Recording file separation method and device based on voiceprint identification |
CN108198569A (en) * | 2017-12-28 | 2018-06-22 | 北京搜狗科技发展有限公司 | A kind of audio-frequency processing method, device, equipment and readable storage medium storing program for executing |
CN108831440A (en) * | 2018-04-24 | 2018-11-16 | 中国地质大学(武汉) | A kind of vocal print noise-reduction method and system based on machine learning and deep learning |
CN108922557A (en) * | 2018-06-14 | 2018-11-30 | 北京联合大学 | A kind of the multi-person speech separation method and system of chat robots |
CN109065051A (en) * | 2018-09-30 | 2018-12-21 | 珠海格力电器股份有限公司 | A kind of voice recognition processing method and device |
CN109545228A (en) * | 2018-12-14 | 2019-03-29 | 厦门快商通信息技术有限公司 | A kind of end-to-end speaker's dividing method and system |
CN110473566A (en) * | 2019-07-25 | 2019-11-19 | 深圳壹账通智能科技有限公司 | Audio separation method, device, electronic equipment and computer readable storage medium |
Family Cites Families (12)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN101404160B (en) * | 2008-11-21 | 2011-05-04 | 北京科技大学 | Voice denoising method based on audio recognition |
CN103871413A (en) * | 2012-12-13 | 2014-06-18 | 上海八方视界网络科技有限公司 | Men and women speaking voice classification method based on SVM and HMM mixing model |
US20170061978A1 (en) * | 2014-11-07 | 2017-03-02 | Shannon Campbell | Real-time method for implementing deep neural network based speech separation |
CN105427858B (en) * | 2015-11-06 | 2019-09-03 | 科大讯飞股份有限公司 | Realize the method and system that voice is classified automatically |
CN106971737A (en) * | 2016-01-14 | 2017-07-21 | 芋头科技(杭州)有限公司 | A kind of method for recognizing sound-groove spoken based on many people |
CN106782565A (en) * | 2016-11-29 | 2017-05-31 | 重庆重智机器人研究院有限公司 | A kind of vocal print feature recognition methods and system |
FR3067511A1 (en) * | 2017-06-09 | 2018-12-14 | Orange | SOUND DATA PROCESSING FOR SEPARATION OF SOUND SOURCES IN A MULTI-CHANNEL SIGNAL |
CN108564952B (en) * | 2018-03-12 | 2019-06-07 | 新华智云科技有限公司 | The method and apparatus of speech roles separation |
CN109272993A (en) * | 2018-08-21 | 2019-01-25 | 中国平安人寿保险股份有限公司 | Recognition methods, device, computer equipment and the storage medium of voice class |
CN109065075A (en) * | 2018-09-26 | 2018-12-21 | 广州势必可赢网络科技有限公司 | A kind of method of speech processing, device, system and computer readable storage medium |
CN109256150B (en) * | 2018-10-12 | 2021-11-30 | 北京创景咨询有限公司 | Speech emotion recognition system and method based on machine learning |
CN109920435B (en) * | 2019-04-09 | 2021-04-06 | 厦门快商通信息咨询有限公司 | Voiceprint recognition method and voiceprint recognition device |
-
2019
- 2019-07-25 CN CN201910678465.5A patent/CN110473566A/en active Pending
-
2020
- 2020-04-24 WO PCT/CN2020/086757 patent/WO2021012734A1/en active Application Filing
Patent Citations (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103811020A (en) * | 2014-03-05 | 2014-05-21 | 东北大学 | Smart voice processing method |
CN105719659A (en) * | 2016-02-03 | 2016-06-29 | 努比亚技术有限公司 | Recording file separation method and device based on voiceprint identification |
CN108198569A (en) * | 2017-12-28 | 2018-06-22 | 北京搜狗科技发展有限公司 | A kind of audio-frequency processing method, device, equipment and readable storage medium storing program for executing |
CN108831440A (en) * | 2018-04-24 | 2018-11-16 | 中国地质大学(武汉) | A kind of vocal print noise-reduction method and system based on machine learning and deep learning |
CN108922557A (en) * | 2018-06-14 | 2018-11-30 | 北京联合大学 | A kind of the multi-person speech separation method and system of chat robots |
CN109065051A (en) * | 2018-09-30 | 2018-12-21 | 珠海格力电器股份有限公司 | A kind of voice recognition processing method and device |
CN109545228A (en) * | 2018-12-14 | 2019-03-29 | 厦门快商通信息技术有限公司 | A kind of end-to-end speaker's dividing method and system |
CN110473566A (en) * | 2019-07-25 | 2019-11-19 | 深圳壹账通智能科技有限公司 | Audio separation method, device, electronic equipment and computer readable storage medium |
Also Published As
Publication number | Publication date |
---|---|
CN110473566A (en) | 2019-11-19 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
WO2021012734A1 (en) | Audio separation method and apparatus, electronic device and computer-readable storage medium | |
WO2021208287A1 (en) | Voice activity detection method and apparatus for emotion recognition, electronic device, and storage medium | |
CN107680582B (en) | Acoustic model training method, voice recognition method, device, equipment and medium | |
CN110211565B (en) | Dialect identification method and device and computer readable storage medium | |
WO2021128741A1 (en) | Voice emotion fluctuation analysis method and apparatus, and computer device and storage medium | |
WO2021047319A1 (en) | Voice-based personal credit assessment method and apparatus, terminal and storage medium | |
CN112233680B (en) | Speaker character recognition method, speaker character recognition device, electronic equipment and storage medium | |
KR101068122B1 (en) | Apparatus and method for rejection based garbage and anti-word model in a speech recognition | |
Chauhan et al. | Speaker recognition and verification using artificial neural network | |
CN106297769B (en) | A kind of distinctive feature extracting method applied to languages identification | |
Biagetti et al. | Speaker identification in noisy conditions using short sequences of speech frames | |
Shareef et al. | Gender voice classification with huge accuracy rate | |
CN108665901B (en) | Phoneme/syllable extraction method and device | |
CN114360514A (en) | Speech recognition method, apparatus, device, medium, and product | |
CN113112992B (en) | Voice recognition method and device, storage medium and server | |
Nirjon et al. | sMFCC: exploiting sparseness in speech for fast acoustic feature extraction on mobile devices--a feasibility study | |
Obaid et al. | Small vocabulary isolated-word automatic speech recognition for single-word commands in Arabic spoken | |
CN113658596A (en) | Semantic identification method and semantic identification device | |
Hizlisoy et al. | Text independent speaker recognition based on MFCC and machine learning | |
CN116110370A (en) | Speech synthesis system and related equipment based on man-machine speech interaction | |
CN115547345A (en) | Voiceprint recognition model training and related recognition method, electronic device and storage medium | |
CN114203160A (en) | Method, device and equipment for generating sample data set | |
CN114420136A (en) | Method and device for training voiceprint recognition model and storage medium | |
Balpande et al. | Speaker recognition based on mel-frequency cepstral coefficients and vector quantization | |
Tzudir et al. | Low-resource dialect identification in Ao using noise robust mean Hilbert envelope coefficients |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
121 | Ep: the epo has been informed by wipo that ep was designated in this application |
Ref document number: 20845047 Country of ref document: EP Kind code of ref document: A1 |
|
NENP | Non-entry into the national phase |
Ref country code: DE |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20845047 Country of ref document: EP Kind code of ref document: A1 |
|
32PN | Ep: public notification in the ep bulletin as address of the adressee cannot be established |
Free format text: NOTING OF LOSS OF RIGHTS PURSUANT TO RULE 112(1) EPC (EPO FORM 1205A DATED 08.08.2022) |
|
122 | Ep: pct application non-entry in european phase |
Ref document number: 20845047 Country of ref document: EP Kind code of ref document: A1 |