CN113643692A - PLC voice recognition method based on machine learning - Google Patents
PLC voice recognition method based on machine learning Download PDFInfo
- Publication number
- CN113643692A CN113643692A CN202110319744.XA CN202110319744A CN113643692A CN 113643692 A CN113643692 A CN 113643692A CN 202110319744 A CN202110319744 A CN 202110319744A CN 113643692 A CN113643692 A CN 113643692A
- Authority
- CN
- China
- Prior art keywords
- voice
- plc
- speech
- model
- machine learning
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 20
- 238000010801 machine learning Methods 0.000 title claims abstract description 17
- 238000001514 detection method Methods 0.000 claims abstract description 18
- 238000013507 mapping Methods 0.000 claims abstract description 18
- 238000000605 extraction Methods 0.000 claims abstract description 16
- 238000012549 training Methods 0.000 claims abstract description 8
- 230000007704 transition Effects 0.000 claims description 9
- 238000001228 spectrum Methods 0.000 claims description 8
- 230000009466 transformation Effects 0.000 claims description 6
- 239000011159 matrix material Substances 0.000 claims description 3
- GNFTZDOKVXKIBK-UHFFFAOYSA-N 3-(2-methoxyethoxy)benzohydrazide Chemical compound COCCOC1=CC=CC(C(=O)NN)=C1 GNFTZDOKVXKIBK-UHFFFAOYSA-N 0.000 claims description 2
- 238000013500 data storage Methods 0.000 claims description 2
- 238000005315 distribution function Methods 0.000 claims description 2
- 238000009432 framing Methods 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims description 2
- 230000003595 spectral effect Effects 0.000 claims description 2
- 230000004048 modification Effects 0.000 abstract description 6
- 238000012986 modification Methods 0.000 abstract description 6
- 239000000203 mixture Substances 0.000 description 5
- 238000004891 communication Methods 0.000 description 3
- 230000008569 process Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- 238000010586 diagram Methods 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000004364 calculation method Methods 0.000 description 1
- 238000006243 chemical reaction Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000009977 dual effect Effects 0.000 description 1
- 230000006870 function Effects 0.000 description 1
- 230000003993 interaction Effects 0.000 description 1
- 238000010187 selection method Methods 0.000 description 1
- 230000035945 sensitivity Effects 0.000 description 1
- 238000007493 shaping process Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/06—Creation of reference templates; Training of speech recognition systems, e.g. adaptation to the characteristics of the speaker's voice
- G10L15/063—Training
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/08—Speech classification or search
- G10L15/14—Speech classification or search using statistical models, e.g. Hidden Markov Models [HMMs]
- G10L15/142—Hidden Markov Models [HMMs]
- G10L15/144—Training of HMMs
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
-
- G—PHYSICS
- G10—MUSICAL INSTRUMENTS; ACOUSTICS
- G10L—SPEECH ANALYSIS OR SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING; SPEECH OR AUDIO CODING OR DECODING
- G10L15/00—Speech recognition
- G10L15/22—Procedures used during a speech recognition process, e.g. man-machine dialogue
- G10L2015/223—Execution procedure of a spoken command
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Computational Linguistics (AREA)
- Health & Medical Sciences (AREA)
- Audiology, Speech & Language Pathology (AREA)
- Human Computer Interaction (AREA)
- Acoustics & Sound (AREA)
- Multimedia (AREA)
- Artificial Intelligence (AREA)
- Probability & Statistics with Applications (AREA)
- Electrically Operated Instructional Devices (AREA)
Abstract
The invention relates to a PLC voice recognition method based on machine learning, which comprises the following steps that a, voice signal samples are collected; b, voice signal endpoint detection and feature extraction; c, training an HMM-GMM model; d, establishing a mapping relation between the voice command and the PLC register data; e, collecting voice instructions; f, carrying out endpoint detection and feature extraction on the voice instruction; matching the characteristics of the voice instruction with the model; and h, modifying the register data according to the matching result and the mapping relation of the PLC register data. The method can realize the output of signals and the modification of parameters, can accurately identify voice instructions sent by operators, and also can replace industrial control means such as buttons, keys and the like with voice instructions which are more friendly to the operators, so that the operators do not need to face complex operation interfaces, and can realize remote operation equipment, thereby adding new modes and ideas for industrial control modes.
Description
Technical Field
The invention relates to the technical field of machine learning, in particular to a PLC voice recognition method based on machine learning.
Background
In the traditional industrial control, an operator inputs signals or modifies parameters to a PLC (programmable logic controller) by using equipment such as a button, a touch screen, a mouse and a keyboard, and outputs instructions to the outside after logic processing is carried out by the PLC, so that the equipment is controlled.
Based on the above premises, it is very important to provide a natural and convenient human-computer interaction mode.
Disclosure of Invention
Aiming at the situation and overcoming the defects of the prior art, the invention provides a PLC voice recognition method based on machine learning, which comprises the steps of acquiring voice data of a command required by equipment, establishing a training model after processing the voice data, matching the acquired command with the model after processing the voice data, writing a matching result into a PLC internal register to realize signal output and parameter modification, accurately recognizing a voice command sent by an operator, and carrying out corresponding operation according to command equipment.
The invention relates to a PLC voice recognition method based on machine learning, which comprises the following steps,
a, collecting a voice signal sample;
b, voice signal endpoint detection and feature extraction;
c, training an HMM-GMM model;
d, establishing a mapping relation between the voice command and the PLC register data;
e, collecting voice instructions;
f, carrying out endpoint detection and feature extraction on the voice instruction;
matching the characteristics of the voice instruction with the model;
and h, modifying the register data according to the matching result and the mapping relation of the PLC register data.
The invention has the beneficial effects that: based on machine learning, firstly, voice signal samples are collected, voice signal end point detection and feature extraction are carried out, then an HMM-GMM model is trained, then a mapping relation between a voice instruction and PLC register data is established, finally, the voice instruction is collected, end point detection and feature extraction are carried out on the voice instruction, the feature of the voice instruction is matched with the model, a matching result is written into a PLC internal register, signal output and parameter modification are achieved, the voice instruction sent by an operator can be accurately identified, namely, the fact that industrial control means such as buttons and keys are replaced by voice instructions which are friendly to the operator is achieved, the operator does not need to face a complex operation interface, meanwhile, remote operation equipment can be achieved, and new modes and ideas are added for industrial control modes.
Drawings
In order to more clearly illustrate the embodiments of the present invention or the technical solutions in the prior art, the drawings used in the description of the embodiments or the prior art will be briefly described below, and it is obvious that the drawings in the following description are only some embodiments of the present invention, and for those skilled in the art, other drawings can be obtained according to these drawings without creative efforts.
FIG. 1 is a flow chart of the general steps of the present invention.
FIG. 2 is a flow chart of step a of the present invention.
FIG. 3 is a waveform diagram of voice signal endpoint detection in step b according to the present invention.
FIG. 4 is a diagram corresponding to D1 in step D of the present invention.
FIG. 5 is a flow chart of step h of the present invention.
Detailed Description
The technical solutions in the embodiments of the present invention will be clearly and completely described below with reference to the drawings in the embodiments of the present invention, and it is obvious that the described embodiments are only a part of the embodiments of the present invention, and not all of the embodiments.
The embodiment one, a PLC voice recognition method based on machine learning, the concrete implementation steps are as follows,
a, collecting a voice signal sample;
b, voice signal endpoint detection and feature extraction;
c, training an HMM-GMM model;
d, establishing a mapping relation between the voice command and the PLC register data;
e, collecting voice instructions;
f, carrying out endpoint detection and feature extraction on the voice instruction (the voice instruction acquisition, endpoint detection and feature extraction modes are the same as those of a voice signal sample);
matching the characteristics of the voice instruction with the model;
and h, modifying the register data according to the matching result through the mapping relation between the matching result and the PLC register data, combining the mapping relation and the voice instruction prediction result, and connecting the destination PLC by using the Sanp7 to complete the modification of the register data of the corresponding address.
In the second embodiment, on the basis of the first embodiment, the step of collecting the voice signal sample in the step a is as follows,
a1, setting the collection times of each voice signal sample;
a2, setting a saving path of the voice signal sample;
a3, setting the format as pyaudio. paInt16, the number of sound channels as 1, the sampling rate as 16000, and the recording duration of a single voice signal as 2.5 s;
a4, collecting voice by using a pyaudio module;
a5, storing the collected voice signal sample by using a wave module;
a6, denoising the voice signal samples by using spectral subtraction;
and A7, executing a loop until the set collection times are reached.
In the third embodiment, on the basis of the first embodiment, the voice signal endpoint detection in the step b is to accurately determine the starting point and the ending point of the voice in a section of signal containing the voice, and distinguish the voice section from the non-voice section. The dual threshold method has three thresholds, the first two being thresholds for speech energy and the last being a threshold for speech zero crossing rate. The energy of voiced sound is higher than that of unvoiced sound, the zero crossing rate of unvoiced sound is higher than that of unvoiced sound, the voiced sound is distinguished by using the energy, then the unvoiced sound is extracted by using the zero crossing rate, the end point detection is completed, the specific steps are as follows,
bj1, taking a higher short-time energy as a threshold MH, and using the threshold to firstly separate out a voiced part in the voice, wherein the interval from A1 to A2 is an interval;
bj2, a lower energy threshold ML is selected, the threshold is utilized to search from A1 and A2 to two ends, the voice part of the lower energy section is also added into the voice section, the range of the voice section is further expanded, and the voice section is still between B1 and B2;
bj3, using short-time zero-crossing rate to distinguish consonants and silence, the threshold value of the short-time zero-crossing rate is Zs, continuously searching the speech segment which is distinguished by using short-time energy to both ends, if the short-time zero-crossing rate is greater than 3 times Zs, the speech segment is regarded as the unvoiced part of speech, and the unvoiced part is added into the speech segment, namely the obtained speech segment, and the speech segment is between C1 and C2.
In a fourth embodiment, on the basis of the first embodiment, the step of extracting the speech signal feature in the step b is as follows,
bt1, pre-emphasis, framing and windowing the speech;
bt2, obtaining a corresponding frequency spectrum for each short-time analysis window through FFT;
bt3, getting Mel Frequency spectrum from above Frequency spectrum through Mel filter bank (human auditory system is a special nonlinear system, which responds to different Frequency signal with different sensitivity. in speech feature extraction, human auditory system is very good, it can not only extract semantic information, but also extract personal feature of speaker, which are all the best that existing speech recognition system can look at;
bt4, performing cepstrum analysis on Mel frequency spectrum, taking logarithm, and performing inverse transformation, wherein the actual inverse transformation is generally realized by DCT discrete cosine transformation, taking the 2 nd to 13 th coefficients after DCT as MFCC coefficients, and obtaining Mel frequency cepstrum coefficient MFCC, which is the feature of the frame of voice.
Example five, based on the first example, the step c of training the HMM-GMM model is as follows,
c1, using HMM-GMM (Hidden Markov Model: a Markov process with Hidden node and visible node) for the phonemes of the speech signal, respectively, Gaussian Mixture Model: Gaussian Mixture Model can be regarded as a Model formed by combining K single Gaussian models, the K sub-models are Hidden variables of the Mixture Model, generally speaking, a Mixture Model can use any probability distribution, the Gaussian Mixture Model is used because the Gaussian distribution has good mathematical property and good calculation performance, the speech recognition is divided into three steps, namely, the first step is used for recognizing the frame into states and is completed by the GMM, the second step is used for combining the states into phonemes and is completed by the HMM, the third step is used for combining the phonemes into words and is completed by the HMM, the whole HMM-GMM network is understood to be mainly used for HMM network services, and the problems to be solved for the speech recognition, firstly, the probability of recognizing the feature of the current frame as the state, namely the Likelihood in the common HMM, is also the mean vector and the covariance matrix in the GMM, namely the GMM network is used for obtaining the probability of the current state; secondly, the probability of the last state being converted into the current state is the state transition probability, the process is Decoding in an HMM, one sequence is converted into the other sequence, and an exponential conversion mode is theoretically provided, so that each frame only takes the state with the highest probability, the route selection method is called as a Viterbi algorithm) for modeling, 3-state modeling is used, and the emission probability of the HMM is modeled by a Gaussian distribution function;
c2, initializing alignment, and corresponding the frame average of the voice signal to each state;
c3, updating model parameters, counting the times of transition of each state, dividing the times by the total times of transition to obtain the transition probability of each state, and calculating the mean vector and covariance matrix of MFCC characteristics of the states, namely the emission probability;
c4, using Viterbi algorithm, according to the transition probability and emission probability obtained in the last step, re-aligning the state level of the speech signal;
c5, repeating the step C2 and the step C3 until convergence;
and C6, saving the trained model.
Sixth, based on the first embodiment, the step of establishing the mapping relationship between the voice command and the PLC register data in step d is as follows,
d1, the data storage of PLC is related to the storage section by the form of Tag, and is divided into input (I), output (O), bit storage (M) and Data Block (DB), when the program accesses the corresponding (I/O) Tag, the program operates the corresponding address by accessing the Process Image Out of CPU, the concrete corresponding relation is as shown in the following FIG. 4;
d2, establishing a link between the PC and the PLC register by using Snap7, wherein Snap7 is an open source library based on Ethernet communication with Siemens PLC of S7 series, and supports Ethernet communication comprising S7-200 of S7 series, S7-200 Smart, S7-300, S7-400, S7-1200 and S7-1500, and the communication step is as follows: 1, instantiate snap 7; set link port number, 2, call API of snap 7: connect,3, the parameters need the IP address, rack number and slot number of the target PLC, 4, call API after the operation is completed: disconnect;
d3, mapping between voice command and PLC data register data, the principle of command operation executed by PLC is to modify data in corresponding register address, through API of snap 7: write _ area and read _ area can implement writing and reading of PLC register data. The parameters need operation type address, register address, start bit and data, and the operation can complete the input and output of I/O point;
for V-zone and M-zone, the API needs to be called: db _ write and client db _ read perform read-write operation on V and M variables, the parameters require register addresses, start bits and the number of bytes of read data (where byte data is 1, word and integer are 2, and double shaping and floating point are 4), and this operation can complete the write-in and read of variable data;
the voice signal replaces a physical button or a key on the touch screen, data capable of realizing the function is written into the designated register address, and mapping between the voice signal and the PLC register data is completed, for example, if the voice signal is "motor start No. 1", and if the motor start No. 1 is assumed when the output point Q0.1 is set, a client.write _ area (0X82,0,0, struck.unpack ('B',2)) statement is corresponded in the program.
Seventh embodiment, based on the first embodiment, the step g of matching the features of the voice command with the models, importing the model group created by each phoneme of the HMM-GMM voice signal sample, and matching the features of the voice command with each model of the model group to obtain the voice sample with the highest matching rate includes the following specific steps,
g1, importing the trained model group;
g2, creating a prediction score list;
g3, matching the input speech with each model of the model group;
g4, calculating a matching score and storing the matching score in a prediction score list;
g5, selecting the model with the highest score;
g6, outputting the voice signal mark corresponding to the model.
When the invention is used specifically, based on machine learning, a voice instruction model is trained, a PLC is connected through a PC end program, industrial control means such as buttons and keys are replaced by voice instructions which are more friendly to operators, so that the operators do not need to face a complex operation interface, meanwhile, remote operation equipment can be realized, a new mode and thought are added for an industrial control mode, and the method comprises the following specific implementation steps,
a, collecting a voice signal sample;
b, voice signal endpoint detection and feature extraction;
c, training an HMM-GMM model;
d, establishing a mapping relation between the voice command and the PLC register data;
e, collecting voice instructions;
f, carrying out endpoint detection and feature extraction on the voice instruction (the voice instruction acquisition, endpoint detection and feature extraction modes are the same as those of a voice signal sample);
matching the characteristics of the voice instruction with the model;
and h, modifying the register data according to the matching result through the mapping relation between the matching result and the PLC register data, combining the mapping relation and the voice instruction prediction result, and connecting the destination PLC by using the Sanp7 to complete the modification of the register data of the corresponding address.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.
Claims (7)
1. The PLC voice recognition method based on machine learning is characterized by comprising the following concrete implementation steps,
a, collecting a voice signal sample;
b, voice signal endpoint detection and feature extraction;
c, training an HMM-GMM model;
d, establishing a mapping relation between the voice command and the PLC register data;
e, collecting voice instructions;
f, carrying out endpoint detection and feature extraction on the voice instruction;
matching the characteristics of the voice instruction with the model;
and h, modifying the register data according to the matching result and the mapping relation of the PLC register data.
2. The PLC speech recognition method based on machine learning of claim 1, wherein the step of collecting speech signal samples in step a is as follows,
a1, setting the collection times of each voice signal sample;
a2, setting a saving path of the voice signal sample;
a3, setting the format as pyaudio. paInt16, the number of sound channels as 1, the sampling rate as 16000, and the recording duration of a single voice signal as 2.5 s;
a4, collecting voice by using a pyaudio module;
a5, storing the collected voice signal sample by using a wave module;
a6, denoising the voice signal samples by using spectral subtraction;
and A7, executing a loop until the set collection times are reached.
3. The PLC voice recognition method based on machine learning of claim 1, wherein the voice signal end point detection in step b is as follows,
bj1, taking a higher short-time energy as a threshold MH, and using the threshold to firstly separate out a voiced part in the voice, wherein the interval from A1 to A2 is an interval;
bj2, a lower energy threshold ML is selected, the threshold is utilized to search from A1 and A2 to two ends, the voice part of the lower energy section is also added into the voice section, the range of the voice section is further expanded, and the voice section is still between B1 and B2;
bj3, using short-time zero-crossing rate to distinguish consonants and silence, the threshold value of the short-time zero-crossing rate is Zs, continuously searching the speech segment which is distinguished by using short-time energy to both ends, if the short-time zero-crossing rate is greater than 3 times Zs, the speech segment is regarded as the unvoiced part of speech, and the unvoiced part is added into the speech segment, namely the obtained speech segment, and the speech segment is between C1 and C2.
4. The PLC speech recognition method based on machine learning of claim 1, wherein the step of speech signal feature extraction in the step b is as follows,
bt1, pre-emphasis, framing and windowing the speech;
bt2, obtaining a corresponding frequency spectrum for each short-time analysis window through FFT;
bt3, passing the spectrum through a Mel filter bank to obtain a Mel spectrum;
bt4, performing cepstrum analysis on Mel frequency spectrum, taking logarithm, and performing inverse transformation, wherein the actual inverse transformation is generally realized by DCT discrete cosine transformation, taking the 2 nd to 13 th coefficients after DCT as MFCC coefficients, and obtaining Mel frequency cepstrum coefficient MFCC, which is the feature of the frame of voice.
5. The PLC speech recognition method based on machine learning of claim 1, wherein the step of training the HMM-GMM model of step c is as follows,
c1, modeling phonemes of the speech signal by using HMM-GMM and by using 3-state modeling, respectively, wherein the emission probability of the HMM is modeled by using a Gaussian distribution function;
c2, initializing alignment, and corresponding the frame average of the voice signal to each state;
c3, updating model parameters, counting the times of transition of each state, dividing the times by the total times of transition to obtain the transition probability of each state, and calculating the mean vector and covariance matrix of MFCC characteristics of the states, namely the emission probability;
c4, using Viterbi algorithm, according to the transition probability and emission probability obtained in the last step, re-aligning the state level of the speech signal;
c5, repeating the step C2 and the step C3 until convergence;
and C6, saving the trained model.
6. The PLC voice recognition method based on machine learning of claim 1, wherein the step of establishing the mapping relationship between the voice command and the PLC register data in the step d is as follows,
d1, the data storage of the PLC is related to the storage interval in the form of Tag and is divided into an input (I), an output (O), a bit storage (M) and a Data Block (DB);
d2, establishing the link between the PC and the PLC register by using snap 7;
d3, mapping of voice command and PLC data register data, the principle of command operation performed by PLC is to modify data in corresponding register address.
7. The PLC speech recognition method based on machine learning of claim 1, wherein the step of matching the features of the speech command with the model in step g is as follows,
g1, importing the trained model group;
g2, creating a prediction score list;
g3, matching the input speech with each model of the model group;
g4, calculating a matching score and storing the matching score in a prediction score list;
g5, selecting the model with the highest score;
g6, outputting the voice signal mark corresponding to the model.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110319744.XA CN113643692B (en) | 2021-03-25 | 2021-03-25 | PLC voice recognition method based on machine learning |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110319744.XA CN113643692B (en) | 2021-03-25 | 2021-03-25 | PLC voice recognition method based on machine learning |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113643692A true CN113643692A (en) | 2021-11-12 |
CN113643692B CN113643692B (en) | 2024-03-26 |
Family
ID=78415711
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110319744.XA Active CN113643692B (en) | 2021-03-25 | 2021-03-25 | PLC voice recognition method based on machine learning |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113643692B (en) |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102324232A (en) * | 2011-09-12 | 2012-01-18 | 辽宁工业大学 | Method for recognizing sound-groove and system based on gauss hybrid models |
CN104078039A (en) * | 2013-03-27 | 2014-10-01 | 广东工业大学 | Voice recognition system of domestic service robot on basis of hidden Markov model |
CN106395516A (en) * | 2016-10-13 | 2017-02-15 | 东华大学 | Passenger elevator intelligent control system based on speech recognition |
CN106601230A (en) * | 2016-12-19 | 2017-04-26 | 苏州金峰物联网技术有限公司 | Logistics sorting place name speech recognition method, system and logistics sorting system based on continuous Gaussian mixture HMM |
CN107331384A (en) * | 2017-06-12 | 2017-11-07 | 平安科技(深圳)有限公司 | Audio recognition method, device, computer equipment and storage medium |
CN109243428A (en) * | 2018-10-15 | 2019-01-18 | 百度在线网络技术(北京)有限公司 | A kind of method that establishing speech recognition modeling, audio recognition method and system |
CN109448726A (en) * | 2019-01-14 | 2019-03-08 | 李庆湧 | A kind of method of adjustment and system of voice control accuracy rate |
US20190266998A1 (en) * | 2017-06-12 | 2019-08-29 | Ping An Technology(Shenzhen) Co., Ltd. | Speech recognition method and device, computer device and storage medium |
CN209433234U (en) * | 2019-03-15 | 2019-09-24 | 陕西中烟工业有限责任公司 | There is the technology for making tobacco threds parameter monitor device of voice alarm function based on Raspberry Pi |
-
2021
- 2021-03-25 CN CN202110319744.XA patent/CN113643692B/en active Active
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102324232A (en) * | 2011-09-12 | 2012-01-18 | 辽宁工业大学 | Method for recognizing sound-groove and system based on gauss hybrid models |
CN104078039A (en) * | 2013-03-27 | 2014-10-01 | 广东工业大学 | Voice recognition system of domestic service robot on basis of hidden Markov model |
CN106395516A (en) * | 2016-10-13 | 2017-02-15 | 东华大学 | Passenger elevator intelligent control system based on speech recognition |
CN106601230A (en) * | 2016-12-19 | 2017-04-26 | 苏州金峰物联网技术有限公司 | Logistics sorting place name speech recognition method, system and logistics sorting system based on continuous Gaussian mixture HMM |
CN107331384A (en) * | 2017-06-12 | 2017-11-07 | 平安科技(深圳)有限公司 | Audio recognition method, device, computer equipment and storage medium |
US20190266998A1 (en) * | 2017-06-12 | 2019-08-29 | Ping An Technology(Shenzhen) Co., Ltd. | Speech recognition method and device, computer device and storage medium |
CN109243428A (en) * | 2018-10-15 | 2019-01-18 | 百度在线网络技术(北京)有限公司 | A kind of method that establishing speech recognition modeling, audio recognition method and system |
CN109448726A (en) * | 2019-01-14 | 2019-03-08 | 李庆湧 | A kind of method of adjustment and system of voice control accuracy rate |
CN209433234U (en) * | 2019-03-15 | 2019-09-24 | 陕西中烟工业有限责任公司 | There is the technology for making tobacco threds parameter monitor device of voice alarm function based on Raspberry Pi |
Also Published As
Publication number | Publication date |
---|---|
CN113643692B (en) | 2024-03-26 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110364171B (en) | Voice recognition method, voice recognition system and storage medium | |
WO2021051544A1 (en) | Voice recognition method and device | |
US20210125603A1 (en) | Acoustic model training method, speech recognition method, apparatus, device and medium | |
CN109147774B (en) | Improved time-delay neural network acoustic model | |
WO2015171646A1 (en) | Method and system for speech input | |
CN104157285A (en) | Voice recognition method and device, and electronic equipment | |
JPH0422276B2 (en) | ||
CN109377981B (en) | Phoneme alignment method and device | |
CN111105785B (en) | Text prosody boundary recognition method and device | |
CN107093422B (en) | Voice recognition method and voice recognition system | |
JPS58100195A (en) | Continuous voice recognition equipment | |
CN102945673A (en) | Continuous speech recognition method with speech command range changed dynamically | |
CN111696522B (en) | Tibetan language voice recognition method based on HMM and DNN | |
CN114360557B (en) | Voice tone conversion method, model training method, device, equipment and medium | |
WO2023245389A1 (en) | Song generation method, apparatus, electronic device, and storage medium | |
CN111951796A (en) | Voice recognition method and device, electronic equipment and storage medium | |
CN111554279A (en) | Multi-mode man-machine interaction system based on Kinect | |
CN108364655A (en) | Method of speech processing, medium, device and computing device | |
CN113327575B (en) | Speech synthesis method, device, computer equipment and storage medium | |
JPH09319392A (en) | Voice recognition device | |
Yang et al. | Linguistically-Informed Training of Acoustic Word Embeddings for Low-Resource Languages. | |
CN113643692B (en) | PLC voice recognition method based on machine learning | |
Islam et al. | Improvement of text dependent speaker identification system using neuro-genetic hybrid algorithm in office environmental conditions | |
Dua et al. | Noise robust automatic speech recognition: review and analysis | |
CN112951270B (en) | Voice fluency detection method and device and electronic equipment |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |