CN113888777A - Voiceprint unlocking method and device based on cloud machine learning - Google Patents

Voiceprint unlocking method and device based on cloud machine learning Download PDF

Info

Publication number
CN113888777A
CN113888777A CN202111051063.6A CN202111051063A CN113888777A CN 113888777 A CN113888777 A CN 113888777A CN 202111051063 A CN202111051063 A CN 202111051063A CN 113888777 A CN113888777 A CN 113888777A
Authority
CN
China
Prior art keywords
recording information
voiceprint
background model
voice representation
voice
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202111051063.6A
Other languages
Chinese (zh)
Other versions
CN113888777B (en
Inventor
吴伟
张嵘
陈鑫
孙嘉鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Institute Of Jindun Public Security Technology Co ltd
Original Assignee
Nanjing Institute Of Jindun Public Security Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Institute Of Jindun Public Security Technology Co ltd filed Critical Nanjing Institute Of Jindun Public Security Technology Co ltd
Priority to CN202111051063.6A priority Critical patent/CN113888777B/en
Publication of CN113888777A publication Critical patent/CN113888777A/en
Application granted granted Critical
Publication of CN113888777B publication Critical patent/CN113888777B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C9/00Individual registration on entry or exit
    • G07C9/00174Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys
    • G07C9/00563Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys using personal physical data of the operator, e.g. finger prints, retinal images, voicepatterns
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C9/00Individual registration on entry or exit
    • G07C9/00174Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys
    • G07C9/00309Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys operated with bidirectional data transmission between data carrier and locks
    • GPHYSICS
    • G07CHECKING-DEVICES
    • G07CTIME OR ATTENDANCE REGISTERS; REGISTERING OR INDICATING THE WORKING OF MACHINES; GENERATING RANDOM NUMBERS; VOTING OR LOTTERY APPARATUS; ARRANGEMENTS, SYSTEMS OR APPARATUS FOR CHECKING NOT PROVIDED FOR ELSEWHERE
    • G07C9/00Individual registration on entry or exit
    • G07C9/00174Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys
    • G07C9/00571Electronically operated locks; Circuits therefor; Nonmechanical keys therefor, e.g. passive or active electrical keys or other data carriers without mechanical keys operated by interacting with a central unit
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02DCLIMATE CHANGE MITIGATION TECHNOLOGIES IN INFORMATION AND COMMUNICATION TECHNOLOGIES [ICT], I.E. INFORMATION AND COMMUNICATION TECHNOLOGIES AIMING AT THE REDUCTION OF THEIR OWN ENERGY USE
    • Y02D30/00Reducing energy consumption in communication networks
    • Y02D30/70Reducing energy consumption in communication networks in wireless communication networks

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Software Systems (AREA)
  • Medical Informatics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Artificial Intelligence (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Computer Networks & Wireless Communication (AREA)
  • Complex Calculations (AREA)
  • Lock And Its Accessories (AREA)

Abstract

The invention discloses a voiceprint unlocking method based on cloud machine learning, which belongs to the technical field of voiceprint recognition and comprises the following steps: acquiring first recording information, and extracting through a probability density function to obtain a first voice representation; obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization; acquiring second recording information, and extracting through a probability density function to obtain a second voice representation; comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model; and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lock if the first recording information is matched with the second recording information.

Description

Voiceprint unlocking method and device based on cloud machine learning
Technical Field
The invention relates to a voiceprint unlocking method and device based on cloud machine learning, and belongs to the technical field of voiceprint recognition.
Background
The lockset is the most important link of security protection, is a product which is just needed and is needed by every family, and the traditional mechanical lock is a large-key-volume high-reliability full-mechanical coded lock without electronic devices. It operates in a unique manner, similar to the dialing of an old telephone set-starting from the start of the dial, turning the dial clockwise to a certain number and then back to the start, a password is entered. The operation is repeated until the last password is input, and the dial is rotated anticlockwise from the starting point to unlock the lock. When the lock is unlocked, the inside is reset, so that the dial is retreated to the starting point to close the lock, the lock can be unlocked only by inputting the password again, and the problem of the reset inside does not need to be considered. If the password is wrongly input, the dial can be reset by rotating anticlockwise (virtual unlocking), and then the password is input again. However, the traditional mechanical lock is low in unlocking capability of the anti-theft technology, a key is easy to lose or even be copied, inconvenience is caused by forgetting the key in daily life, an unlocking method using voiceprint information as an unlocking secret key is urgently needed, and the problem that the traditional local voiceprint recognition training library is low in recognition rate due to the fact that training data are few exists.
Disclosure of Invention
The invention aims to overcome the defects in the prior art, provides a voiceprint unlocking method and device based on cloud machine learning, and solves the problem that a traditional local voiceprint recognition training library is low in recognition rate due to the fact that training data are few.
In order to achieve the purpose, the invention is realized by adopting the following technical scheme:
on one hand, the invention provides a voiceprint unlocking method based on cloud machine learning, which is applied to a main control board embedded in a lockset and comprises the following steps:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
uploading the first voice representation to a cloud server, wherein the cloud server is used for obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm for the first voice representation and synchronizing the general voiceprint background model to a main control board;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the universal voiceprint background model downloaded from the cloud server, and calculating the approximation degree of the second voice representation and the universal voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
Furthermore, the first recording information and the second recording information are recorded and acquired through a microphone arranged on the lock, and the microphone is connected with the main control panel.
Further, the audio content of the first recording information is a predetermined section of specific text or sound, and is stored in the main control board.
In a second aspect, the present invention provides a voiceprint unlocking method based on cloud machine learning, which is applied to a lock, and includes:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
Further, the second speech characterization is compared with the universal voiceprint background model by using a maximum posterior probability algorithm.
In a third aspect, the present invention provides a voiceprint unlocking device based on cloud machine learning, including:
the first extraction unit is used for acquiring first recording information and extracting a first voice representation through a probability density function;
the background library establishing unit is used for obtaining a general voiceprint background model as a background library through a first voice representation by using an expectation maximization algorithm;
the second extraction unit is used for acquiring second recording information and extracting a second voice representation through a probability density function;
a comparison calculation unit; comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and the comparison unit is used for comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
In a fourth aspect, the present invention provides a voiceprint unlocking system based on cloud machine learning, which includes a lock, and further includes:
the microphone is arranged on the lockset and used for recording the speaking recording information and sending the recording information to the main control board;
the master control board is embedded in the lockset and used as an algorithm program carrier, is connected with an external microphone, acquires recording information, obtains voice representation through probability density function extraction, uploads the voice representation to the cloud server, and accesses the Internet to download a universal voiceprint background model from the cloud server at regular intervals; comparing the extracted voice representation with the downloaded general voiceprint background model, calculating the approximation degree of the voice representation and the downloaded general voiceprint background model, comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to the comparison result, and opening a lock if the first recording information is matched with the second recording information;
and the cloud server is used for obtaining a general voiceprint background model as a background library through the received voice representation as training data and an expectation maximization algorithm, and synchronizing the general voiceprint background model to the main control board.
Furthermore, a storage module is arranged on the main control board and used for storing the received audio information.
In a fifth aspect, the invention provides a voiceprint unlocking device based on cloud machine learning, which comprises a processor and a storage medium;
the storage medium is used for storing instructions;
the processor is configured to operate in accordance with the instructions to perform the steps of the method according to any of the above.
In a sixth aspect, the invention provides a computer-readable storage medium having stored thereon a computer program which, when executed by a processor, performs the steps of any of the methods described above.
Compared with the prior art, the invention has the following beneficial effects: compared with the traditional local voiceprint recognition training library, the expectation maximization algorithm and the general voiceprint background model are deployed at the cloud end of the server, training is carried out through data uploaded by each terminal, the data are asynchronously issued to the main control boards of each access network, and data of other speakers are used for each speaker to carry out pre-training, so that the problem of low recognition rate caused by the fact that the traditional local voiceprint recognition training library is few in training data is solved.
Drawings
FIG. 1 is a flow chart of the method of the present invention;
FIG. 2 is a functional block diagram of the method of the present invention;
FIG. 3 is a speech characterization training block diagram.
Detailed Description
The invention is further described below with reference to the accompanying drawings. The following examples are only for illustrating the technical solutions of the present invention more clearly, and the protection scope of the present invention is not limited thereby.
Example 1
This embodiment introduces a voiceprint unlocking method based on cloud machine learning, including: be applied to and inlay in inside main control board of tool to lock, include:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
uploading the first voice representation to a cloud server, wherein the cloud server is used for obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm for the first voice representation and synchronizing the general voiceprint background model to a main control board;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the universal voiceprint background model downloaded from the cloud server, and calculating the approximation degree of the second voice representation and the universal voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
The first recording information and the second recording information are recorded and obtained through a microphone arranged on the lock.
As shown in fig. 1 to fig. 3, the application process of the voiceprint unlocking method based on cloud machine learning provided in this embodiment specifically involves the following steps:
firstly, a user inputs voice through a microphone, the voice is transmitted to a main control board, and voice representation is extracted through a probability density function;
a speaker wakes up the main control panel to record audio through a microphone by operating a touch, a switch or a button and the like, the audio content is a pre-appointed section of specific characters or sound, the main control panel stores the audio in a cache, and the audio is voiceprint information.
The probability density function is a linear combination of weighted summation of M component densities and a plurality of Gaussian distribution functions, and can theoretically fit any type of distribution of voice characterization, and the formula is as follows:
Figure BDA0003252770680000061
where x is a random vector of dimension d and λ is the parameter set { λ ] of the Gaussian mixture model1,...,λ2,...,λM},λi=(wiii),i∈[1,…,M],wiIs the weight component of the mixture of the weight components,
Figure BDA0003252770680000062
is a probability density function of the ith d-dimensional Gaussian component, muiiTheir mean and variance, respectively; the probability density function of the gaussian component is:
Figure BDA0003252770680000063
and the data obtained after the function operation is used for describing the distribution of the data points/characteristic points, namely the voice characterization.
Secondly, the main control board is used as a terminal device to upload the recognized voice representation to a cloud server through mobile data or WiFi, the cloud server is used as training data through the voice representation, and a general voiceprint background model is obtained through an expectation maximization algorithm and used as a background library;
the expectation-maximization algorithm is an algorithm that finds the parameter maximum likelihood estimate or the maximum a posteriori probability in a probabilistic model that relies on hidden variables that cannot be observed.
The expectation maximization algorithm is divided into two steps:
first, solving the posterior probability by fixed parameters, fixing mu and sigma, and calculating the posterior probability distribution p (z)(n)|x(n)) The formula is as follows:
Figure BDA0003252770680000064
wherein mu is mean value of Gaussian distribution, sigma is variance of Gaussian distribution, k represents kth Gaussian distribution, and pikRepresents the weight coefficient of the k-th Gaussian distribution and satisfies pik≥0,
Figure BDA0003252770680000071
I.e. the prior probability that sample x is generated by the kth gaussian distribution; n represents the nth sample; gamma is posterior distribution; x is the number of(n)Representing n training samples generated by a Gaussian mixture model; z is a radical of(n)Indicating from which gaussian it came.
Using the above formula gammankAnd expressing the posterior probability of the nth sample to the kth Gaussian distribution, wherein the posterior probability is expressed by an NxK matrix, N is a sample set, and K is a Gaussian distribution set.
The second step is that: fixing the posterior probability, optimizing parameters corresponding to the lower bound of the solution evidence, and maximizing marginal likelihood p (x | mu, sigma) under the condition of known posterior probability, namely maximizing log likelihood (probability), wherein the formula is as follows:
Figure BDA0003252770680000072
wherein q (z) is p (z)(n)K), ELBO is defined as the lower bound of the function by an inequality, so the lower bound of its log-likelihood is further solved as follows:
Figure BDA0003252770680000073
Figure BDA0003252770680000074
wherein D is a training set, gamma is posterior distribution, and the Lagrangian method is utilized to solve the following updated conclusion:
Figure BDA0003252770680000075
Figure BDA0003252770680000081
Figure BDA0003252770680000082
wherein
Figure BDA0003252770680000083
And obtaining a universal voiceprint background model consisting of a plurality of voice representations through the expectation maximization algorithm, asynchronously sending the model to a main control board which is communicated with the cloud server at present, and updating the voiceprint background model on the main control board so as to reduce the recognition error value.
Step three, the main control board uses the current voice representation to compare with the universal voiceprint background model downloaded from the cloud server by using a maximum posterior probability algorithm,
the self-adaptive process of the maximum posterior probability algorithm mainly comprises three steps:
first, estimating speaker voice data based on sufficient statistics of each Gaussian component in the universal voiceprint background model, wherein for speaker voice characterization, sufficient statistics comprise frequency numbers (N) of observation sequences from each component ii) First order (E)i(X), mean expectation) and second order (S)i(X), mean expectation) moment to compute the weight, mean and variance of the gaussian mixture model.
Second, using data-dependent mixing parameters, i.e. hereinafter
Figure BDA0003252770680000084
And combining the sufficient statistics of the new estimation and the sufficient statistics of the universal voiceprint background model to obtain the final voice characteristic parameter estimation.
The specific process is as follows:
given a generic voiceprint background model and a speaker-specific observation sequence X ═ X1,…,xTThe division model is used as the division model;
where component i is paired with observation xtResponse speed of (2), observed data xtProbability of the ith component from the generic voiceprint background model:
Figure BDA0003252770680000091
wherein, wiIs a mixing weight component; mu.st、ΣtTheir mean and variance, respectively; g is the probability density of the gaussian component; m is the number of component densities.
Using Pr (i | x)ti) And (3) calculating sufficient statistics, wherein the sum of the probabilities of the T observation sequence vectors from the component i is as follows:
Figure BDA0003252770680000092
the mean of the T observation sequence vectors from component i is expected to be:
Figure BDA0003252770680000093
the variance of the T observation sequence vectors from component i is expected to be:
Figure BDA0003252770680000094
and thirdly, updating parameters (weight, mean value and variance) of the mixed components by using the sufficient statistics obtained in the second step:
Figure BDA0003252770680000095
Figure BDA0003252770680000096
Figure BDA0003252770680000097
for each parameter of the Gaussian mixture component, the data depends on the mixture parameter
Figure BDA0003252770680000098
Is defined as follows:
Figure BDA0003252770680000099
rρis a fixed correlation factor based on p, typically using the same a update parameter, i.e.
Figure BDA0003252770680000101
Figure BDA0003252770680000102
Experiments show that the value range of r is (8-20) effective, the self-adaptive process only updates the mean value with the best effect, and the actual system has the best effect
Figure BDA0003252770680000103
And γ is only a normalization factor to ensure that the sum of the updated weight parameters is 1, so it traverses according to the component i
Figure BDA0003252770680000104
Comparing the similarity degree value with the preset threshold value with the similarity degree represented by the voice to obtain whether the speaker is matched, and opening the lockset if the speaker is matched; the speech characterization and the absolute value of the difference are the approximation degree, and the smaller the difference is, the more similar the speech characterization and the absolute value are.
Example 2
The embodiment provides a voiceprint unlocking method based on cloud machine learning, which is applied to a lockset and comprises the following steps:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
Further, the second speech characterization is compared with the universal voiceprint background model by using a maximum posterior probability algorithm.
Example 3
This embodiment provides a voiceprint unlocking means based on high in the clouds machine learning, includes:
the first extraction unit is used for acquiring first recording information and extracting a first voice representation through a probability density function;
the background library establishing unit is used for obtaining a general voiceprint background model as a background library through a first voice representation by using an expectation maximization algorithm;
the second extraction unit is used for acquiring second recording information and extracting a second voice representation through a probability density function;
a comparison calculation unit; comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and the comparison unit is used for comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
Example 4
This embodiment provides a voiceprint system of unblanking based on high in clouds machine learning, including the tool to lock, still include:
the microphone is arranged on the lockset and used for recording the speaking recording information and sending the recording information to the main control board;
the master control board is embedded in the lockset and used as an algorithm program carrier, is connected with an external microphone, acquires recording information, obtains voice representation through probability density function extraction, uploads the voice representation to the cloud server, and accesses the Internet to download a universal voiceprint background model from the cloud server at regular intervals; comparing the extracted voice representation with the downloaded general voiceprint background model, calculating the approximation degree of the voice representation and the downloaded general voiceprint background model, comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to the comparison result, and opening a lock if the first recording information is matched with the second recording information;
and the cloud server is used for obtaining a general voiceprint background model as a background library through the received voice representation as training data and an expectation maximization algorithm, and synchronizing the general voiceprint background model to the main control board.
Furthermore, a storage module is arranged on the main control board and used for storing the received audio information.
Example 5
The embodiment provides a voiceprint unlocking device based on cloud machine learning, which comprises a processor and a storage medium;
the storage medium is used for storing instructions;
the processor is configured to operate in accordance with the instructions to perform the steps of the method according to any one of:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
Example 6
The present embodiments provide a computer-readable storage medium, on which a computer program is stored, which when executed by a processor implements the steps of any of the methods:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
The above description is only a preferred embodiment of the present invention, and it should be noted that, for those skilled in the art, several modifications and variations can be made without departing from the technical principle of the present invention, and these modifications and variations should also be regarded as the protection scope of the present invention.

Claims (10)

1. The utility model provides a voiceprint unlocking method based on cloud machine learning, its characterized in that is applied to the main control board that inlays in the inside tool to lock, includes:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
uploading the first voice representation to a cloud server, wherein the cloud server is used for obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm for the first voice representation and synchronizing the general voiceprint background model to a main control board;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the universal voiceprint background model downloaded from the cloud server, and calculating the approximation degree of the second voice representation and the universal voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
2. The cloud machine learning-based voiceprint unlocking method according to claim 1, wherein the voiceprint unlocking method comprises the following steps: the first recording information and the second recording information are recorded and obtained through a microphone arranged on the lock, and the microphone is connected with the main control panel.
3. The cloud machine learning-based voiceprint unlocking method according to claim 1, wherein the voiceprint unlocking method comprises the following steps: the audio content of the first recording information is a predetermined section of specific characters or sound, and is stored in the main control board.
4. The utility model provides a voiceprint unlocking method based on cloud machine learning, which is characterized in that, is applied to the tool to lock, includes:
acquiring first recording information, and extracting through a probability density function to obtain a first voice representation;
obtaining a general voiceprint background model as a background library by using an expectation maximization algorithm through a first voice characterization;
acquiring second recording information, and extracting through a probability density function to obtain a second voice representation;
comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
5. The cloud machine learning-based voiceprint unlocking method according to claim 4, wherein the voiceprint unlocking method comprises the following steps: the second speech characterization is compared with the generic voiceprint background model using a maximum a posteriori probability algorithm.
6. The utility model provides a voiceprint unlocking means based on high in the clouds machine learning which characterized in that includes:
the first extraction unit is used for acquiring first recording information and extracting a first voice representation through a probability density function;
the background library establishing unit is used for obtaining a general voiceprint background model as a background library through a first voice representation by using an expectation maximization algorithm;
the second extraction unit is used for acquiring second recording information and extracting a second voice representation through a probability density function;
a comparison calculation unit; comparing the second voice representation with the general voiceprint background model, and calculating the approximation degree of the second voice representation and the general voiceprint background model;
and the comparison unit is used for comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to a comparison result, and opening the lockset if the first recording information is matched with the second recording information.
7. The utility model provides a voiceprint system of unblanking based on high in clouds machine learning, includes the tool to lock, its characterized in that still includes:
the microphone is arranged on the lockset and used for recording the speaking recording information and sending the recording information to the main control board;
the master control board is embedded in the lockset and used as an algorithm program carrier, is connected with an external microphone, acquires recording information, obtains voice representation through probability density function extraction, uploads the voice representation to the cloud server, and accesses the Internet to download a universal voiceprint background model from the cloud server at regular intervals; comparing the extracted voice representation with the downloaded general voiceprint background model, calculating the approximation degree of the voice representation and the downloaded general voiceprint background model, comparing the approximation degree value with a preset threshold value, judging whether the first recording information is matched with the second recording information according to the comparison result, and opening a lock if the first recording information is matched with the second recording information;
and the cloud server is used for obtaining a general voiceprint background model as a background library through the received voice representation as training data and an expectation maximization algorithm, and synchronizing the general voiceprint background model to the main control board.
8. The cloud machine learning-based voiceprint unlocking method according to claim 6, wherein the voiceprint unlocking method comprises the following steps: and the main control board is provided with a storage module for storing the received audio information.
9. The utility model provides a voiceprint unlocking means based on high in the clouds machine learning which characterized in that: comprising a processor and a storage medium;
the storage medium is used for storing instructions;
the processor is configured to operate in accordance with the instructions to perform the steps of the method according to any one of claims 4 to 5.
10. A computer-readable storage medium having stored thereon a computer program, characterized in that: the program when executed by a processor implements the steps of the method of any one of claims 4 to 5.
CN202111051063.6A 2021-09-08 2021-09-08 Voiceprint unlocking method and device based on cloud machine learning Active CN113888777B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111051063.6A CN113888777B (en) 2021-09-08 2021-09-08 Voiceprint unlocking method and device based on cloud machine learning

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111051063.6A CN113888777B (en) 2021-09-08 2021-09-08 Voiceprint unlocking method and device based on cloud machine learning

Publications (2)

Publication Number Publication Date
CN113888777A true CN113888777A (en) 2022-01-04
CN113888777B CN113888777B (en) 2023-08-18

Family

ID=79008734

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111051063.6A Active CN113888777B (en) 2021-09-08 2021-09-08 Voiceprint unlocking method and device based on cloud machine learning

Country Status (1)

Country Link
CN (1) CN113888777B (en)

Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102982799A (en) * 2012-12-20 2013-03-20 中国科学院自动化研究所 Speech recognition optimization decoding method integrating guide probability
CN104143326A (en) * 2013-12-03 2014-11-12 腾讯科技(深圳)有限公司 Voice command recognition method and device
CN104992708A (en) * 2015-05-11 2015-10-21 国家计算机网络与信息安全管理中心 Short-time specific audio detection model generating method and short-time specific audio detection method
CN107680600A (en) * 2017-09-11 2018-02-09 平安科技(深圳)有限公司 Sound-groove model training method, audio recognition method, device, equipment and medium
CN108831440A (en) * 2018-04-24 2018-11-16 中国地质大学(武汉) A kind of vocal print noise-reduction method and system based on machine learning and deep learning
CN109273002A (en) * 2018-10-26 2019-01-25 蔚来汽车有限公司 Vehicle configuration method, system, vehicle device and vehicle
CN109872721A (en) * 2017-12-05 2019-06-11 富士通株式会社 Voice authentication method, information processing equipment and storage medium
CN110634507A (en) * 2018-06-06 2019-12-31 英特尔公司 Speech classification of audio for voice wakeup

Patent Citations (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102982799A (en) * 2012-12-20 2013-03-20 中国科学院自动化研究所 Speech recognition optimization decoding method integrating guide probability
CN104143326A (en) * 2013-12-03 2014-11-12 腾讯科技(深圳)有限公司 Voice command recognition method and device
CN104992708A (en) * 2015-05-11 2015-10-21 国家计算机网络与信息安全管理中心 Short-time specific audio detection model generating method and short-time specific audio detection method
CN107680600A (en) * 2017-09-11 2018-02-09 平安科技(深圳)有限公司 Sound-groove model training method, audio recognition method, device, equipment and medium
CN109872721A (en) * 2017-12-05 2019-06-11 富士通株式会社 Voice authentication method, information processing equipment and storage medium
CN108831440A (en) * 2018-04-24 2018-11-16 中国地质大学(武汉) A kind of vocal print noise-reduction method and system based on machine learning and deep learning
CN110634507A (en) * 2018-06-06 2019-12-31 英特尔公司 Speech classification of audio for voice wakeup
CN109273002A (en) * 2018-10-26 2019-01-25 蔚来汽车有限公司 Vehicle configuration method, system, vehicle device and vehicle

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
曾春艳;马超峰;王志锋;孔祥斌;: "基于卷积神经网络的鲁棒性说话人识别方法", 华中科技大学学报(自然科学版), no. 06, pages 44 - 49 *

Also Published As

Publication number Publication date
CN113888777B (en) 2023-08-18

Similar Documents

Publication Publication Date Title
EP2763134B1 (en) Method and apparatus for voice recognition
Campisi et al. User authentication using keystroke dynamics for cellular phones
US20220238117A1 (en) Voice identity feature extractor and classifier training
CN103065631B (en) A kind of method of speech recognition, device
US20080118082A1 (en) Removal of noise, corresponding to user input devices from an audio signal
CN110211575A (en) Voice for data enhancing adds method for de-noising and system
CN109119090A (en) Method of speech processing, device, storage medium and electronic equipment
WO2020098083A1 (en) Call separation method and apparatus, computer device and storage medium
WO2009020482A2 (en) Hidden markov model ('hmm')-based user authentication using keystroke dynamics
CN110570869A (en) Voiceprint recognition method, device, equipment and storage medium
CN112328994A (en) Voiceprint data processing method and device, electronic equipment and storage medium
CN107481736A (en) A kind of vocal print identification authentication system and its certification and optimization method and system
Westover et al. Achievable rates for pattern recognition
US20120069767A1 (en) Method and an arrangement for a mobile telecommunications network
CN113888777A (en) Voiceprint unlocking method and device based on cloud machine learning
CN109040466A (en) voice-based mobile terminal unlocking method and device
CN116383719A (en) MGF radio frequency fingerprint identification method for LFM radar
CN110364169A (en) Method for recognizing sound-groove, device, equipment and computer readable storage medium
CN114155880A (en) Illegal voice recognition method and system based on GBDT algorithm model
CN112131541A (en) Identity verification method and system based on vibration signal
CN113918941A (en) Abnormal behavior detection method and device, computing equipment and storage medium
CN111951791A (en) Voiceprint recognition model training method, recognition method, electronic device and storage medium
CN114400009B (en) Voiceprint recognition method and device and electronic equipment
CN107451437B (en) Locking method and device of mobile terminal
CN111400680B (en) Mobile phone unlocking password prediction method based on sensor and related device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant