CN110660396A - Language processing system and method based on MEMS - Google Patents

Language processing system and method based on MEMS Download PDF

Info

Publication number
CN110660396A
CN110660396A CN201810607268.XA CN201810607268A CN110660396A CN 110660396 A CN110660396 A CN 110660396A CN 201810607268 A CN201810607268 A CN 201810607268A CN 110660396 A CN110660396 A CN 110660396A
Authority
CN
China
Prior art keywords
mems
target vector
pulse
unit
processing system
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201810607268.XA
Other languages
Chinese (zh)
Inventor
周自伟
孔祥文
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Jiangsu Dexinke Intelligent Sensor Research Institute Co Ltd
Original Assignee
Jiangsu Dexinke Intelligent Sensor Research Institute Co Ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Jiangsu Dexinke Intelligent Sensor Research Institute Co Ltd filed Critical Jiangsu Dexinke Intelligent Sensor Research Institute Co Ltd
Priority to CN201810607268.XA priority Critical patent/CN110660396A/en
Publication of CN110660396A publication Critical patent/CN110660396A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L15/00Speech recognition
    • G10L15/28Constructional details of speech recognition systems
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/03Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters
    • G10L25/12Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the type of extracted parameters the extracted parameters being prediction coefficients
    • GPHYSICS
    • G10MUSICAL INSTRUMENTS; ACOUSTICS
    • G10LSPEECH ANALYSIS TECHNIQUES OR SPEECH SYNTHESIS; SPEECH RECOGNITION; SPEECH OR VOICE PROCESSING TECHNIQUES; SPEECH OR AUDIO CODING OR DECODING
    • G10L25/00Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00
    • G10L25/27Speech or voice analysis techniques not restricted to a single one of groups G10L15/00 - G10L21/00 characterised by the analysis technique

Landscapes

  • Engineering & Computer Science (AREA)
  • Computational Linguistics (AREA)
  • Health & Medical Sciences (AREA)
  • Audiology, Speech & Language Pathology (AREA)
  • Human Computer Interaction (AREA)
  • Physics & Mathematics (AREA)
  • Acoustics & Sound (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression, Expansion, Code Conversion, And Decoders (AREA)

Abstract

The invention relates to a language processing system based on MEMS and a method thereof, which mainly comprises the following steps: the system comprises an MEMS (micro electro mechanical system) identification sensing unit, a target vector unit and a multi-pulse analysis unit, wherein the MEMS identification sensing unit is used for quantifying the multi-pulse analysis function; the target vector unit acts on a target vector, determines an initial gain level for the pulse sequence, completes single-gain multi-pulse analysis for multiple times, and provides a sequence which can represent the target vector most as an output signal for different gain levels each time; the output end of the MEMS identification sensing unit is connected with at least one pulse series multi-pulse analysis unit, the target vector unit can be used as a pulse series sequence, and the pulse series is composed of a group of single-gain pulses. The invention belongs to an ultra-short time analysis system, and greatly improves the reaction time and the reaction accuracy of a language processing system.

Description

Language processing system and method based on MEMS
Technical Field
The invention relates to a language processing system, in particular to a language processing system based on MEMS and a method thereof.
Background
At present, human beings have more and more closely related, and the language difference and other factors make the language processing system and method more and more important, so that the problems of the response time and the response accuracy of the language processing system are urgently needed to be solved.
Disclosure of Invention
In view of the above problems, an object of the present invention is to provide a MEMS-based language processing system and method thereof, so as to shorten the response time of the language processing system and improve the response accuracy.
In order to achieve the purpose, the invention adopts the technical scheme that:
a language processing system based on MEMS and its method includes:
MEMS discerns induction element: the output end of the MEMS identification sensing unit is connected with an analyzer, and the input end of the MEMS is used for inputting a speech signal to generate the ultrashort characteristic of the input speech signal;
target vector unit: the target vector unit generates a target vector by at least the input speech signal and the optional ultra-short time characteristic, and the multi-pulse analysis unit of the output line of the target vector unit is connected with the multi-pulse analysis unit of the output line of the target vector unit;
a multi-pulse analysis unit: the multi-pulse analysis unit generates a group of pulse sequences with equal amplitude, variable symbols and variable intervals, the multi-pulse analysis unit outputs a signal corresponding to one pulse sequence with equal amplitude, variable symbols and variable intervals corresponding to the multi-pulse analysis unit, and the sequence is a sequence which can judge that the sequence can represent the target vector most through judgment according to the maximum likelihood.
Further, a MEMS speech processing system and method thereof includes an ultra-short time analyzer for generating ultra-short time characteristics by performing linear prediction coefficients on an input speech signal.
Further, the initial pulse of each of the equal amplitude, variable sign, variable spacing burst sequences is located at the same position.
Further, the target vector unit comprises a global decision determiner comprising a perceptual weighting filter for filtering the constant amplitude, variable sign, variable spacing pulse sequences.
Further, the ultra-short time analyzer has functions and characteristics that combine MEMS recognition sensing and switching.
Furthermore, the input end of the MEMS identification sensing unit inputs a voice signal and can also generate long-term characteristics, the long-term characteristics at least comprise a tone value of the input voice signal and short-term characteristics of the MEMS identification sensing output end, and the target vector unit is used for generating a target vector at least by the input voice signal and optional short-term and long-term characteristics.
Further, the initial pulses of the pulse sequence are located at the same sample position.
The invention has the beneficial effects that:
the technical scheme of the invention is designed based on MEMS, belongs to an ultra-short time analysis system, greatly shortens the reaction time of a language processing system and improves the reaction accuracy.
Drawings
FIG. 1 is a schematic diagram of the MEMS-based language processing system and method of the present invention.
The specific implementation mode is as follows:
the present invention will be described in further detail below with reference to the accompanying drawings and technical solutions, and embodiments of the present invention will be described in detail by way of preferred examples, but the embodiments of the present invention are not limited thereto.
A language processing system based on MEMS and its method includes:
MEMS discerns induction element: the output end of the MEMS identification sensing unit is connected with an analyzer, and the input end of the MEMS is used for inputting a speech signal to generate the ultrashort characteristic of the input speech signal;
target vector unit: the target vector unit generates a target vector by at least the input speech signal and the optional ultra-short time characteristic, and the multi-pulse analysis unit of the output line of the target vector unit is connected with the multi-pulse analysis unit of the output line of the target vector unit;
a multi-pulse analysis unit: the multi-pulse analysis unit generates a group of pulse sequences with equal amplitude, variable symbols and variable intervals, the multi-pulse analysis unit outputs a signal corresponding to one pulse sequence with equal amplitude, variable symbols and variable intervals corresponding to the multi-pulse analysis unit, and the sequence is a sequence which can judge that the sequence can represent the target vector most through judgment according to the maximum likelihood.
Further, a MEMS speech processing system and method thereof includes an ultra-short time analyzer for generating ultra-short time characteristics by performing linear prediction coefficients on an input speech signal.
Further, the initial pulse of each of the equal amplitude, variable sign, variable spacing burst sequences is located at the same position.
Further, the target vector unit comprises a global decision determiner comprising a perceptual weighting filter for filtering the constant amplitude, variable sign, variable spacing pulse sequences.
Further, the ultra-short time analyzer has functions and characteristics that combine MEMS recognition sensing and switching.
Furthermore, the input end of the MEMS identification sensing unit inputs a voice signal and can also generate long-term characteristics, the long-term characteristics at least comprise a tone value of the input voice signal and short-term characteristics of the MEMS identification sensing output end, and the target vector unit is used for generating a target vector at least by the input voice signal and optional short-term and long-term characteristics.
Further, the initial pulses of the pulse sequence are located at the same sample position.
The above description is only a preferred embodiment of the present invention and is not limited to the present invention, it should be noted that, for those skilled in the art, many modifications and variations can be made without departing from the principle of the present invention, and these changes and variations should also be construed as the protection scope of the present invention.

Claims (7)

1. A language processing system based on MEMS and its method are characterized by comprising:
MEMS discerns induction element: the output end of the MEMS identification sensing unit is connected with an analyzer, and the input end of the MEMS is used for inputting a speech signal to generate the ultrashort characteristic of the input speech signal;
target vector unit: said target vector unit generating a target vector from at least said input speech signal and optionally said ultra short time characteristic, a multi-pulse analysis unit connected to an output line of said target vector unit and a multi-pulse analysis of an output line of said target vector unit;
a multi-pulse analysis unit: the multi-pulse analysis unit generates a group of pulse sequences with equal amplitude, variable symbols and variable intervals, and outputs a signal corresponding to the pulse sequences with equal amplitude, variable symbols and variable intervals with the multi-pulse analysis unit, wherein the sequences can be judged to represent the target vector most according to the maximum likelihood.
2. The language processing system of MEMS and the method thereof according to claim 1, wherein: an ultra-short time analyzer is included for generating ultra-short time characteristics by linear prediction coefficients of an input speech signal.
3. The language processing system of MEMS and the method thereof according to claim 1, wherein: the initial pulse of each of the series of equal amplitude, variable sign, variable spacing bursts is located at the same position.
4. The language processing system of MEMS and the method thereof according to claim 1, wherein: the target vector unit comprises a global decision determiner comprising an perceptual weighting filter for filtering the constant amplitude, variable sign, variable spacing pulse sequences.
5. The language processing system of MEMS and the method thereof according to claim 2, wherein: the ultra-short time analyzer has the functions and characteristics of combining MEMS recognition sensing and conversion.
6. The language processing system of MEMS and method thereof according to claims 1-5, wherein: the MEMS recognition sensing unit comprises an input end, a target vector unit and a sensing output end, wherein the input end of the MEMS recognition sensing unit inputs a voice signal and can also generate long-term characteristics, the long-term characteristics at least comprise a tone value of the input voice signal and short-term characteristics of the MEMS recognition sensing output end, and the target vector unit is used for generating a target vector at least by the input voice signal and optional short-term and long-term characteristics.
7. The language processing system of MEMS and the method thereof according to claim 6, wherein: the initial pulses of the pulse sequence are located at the same sample position.
CN201810607268.XA 2018-06-13 2018-06-13 Language processing system and method based on MEMS Pending CN110660396A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201810607268.XA CN110660396A (en) 2018-06-13 2018-06-13 Language processing system and method based on MEMS

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201810607268.XA CN110660396A (en) 2018-06-13 2018-06-13 Language processing system and method based on MEMS

Publications (1)

Publication Number Publication Date
CN110660396A true CN110660396A (en) 2020-01-07

Family

ID=69026169

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201810607268.XA Pending CN110660396A (en) 2018-06-13 2018-06-13 Language processing system and method based on MEMS

Country Status (1)

Country Link
CN (1) CN110660396A (en)

Citations (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1995030222A1 (en) * 1994-04-29 1995-11-09 Sherman, Jonathan, Edward A multi-pulse analysis speech processing system and method

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO1995030222A1 (en) * 1994-04-29 1995-11-09 Sherman, Jonathan, Edward A multi-pulse analysis speech processing system and method
CN1153566A (en) * 1994-04-29 1997-07-02 乔纳森·爱德华·谢尔曼 Multi-pulse analysis speech processing system and method

Similar Documents

Publication Publication Date Title
US10930270B2 (en) Processing audio waveforms
CN110827837B (en) Whale activity audio classification method based on deep learning
CN106782501B (en) Speech feature extraction method and device based on artificial intelligence
Gevaert et al. Neural networks used for speech recognition
US8924209B2 (en) Identifying spoken commands by templates of ordered voiced and unvoiced sound intervals
CN106653056B (en) Fundamental frequency extraction model and training method based on LSTM recurrent neural network
US9454976B2 (en) Efficient discrimination of voiced and unvoiced sounds
CN106297826A (en) Speech emotional identification system and method
RU2016127609A (en) METHOD AND DEVICE FOR DETERMINING SIGNALS IN DATA
JPS62231997A (en) Voice recognition system and method
CN106098079B (en) Method and device for extracting audio signal
Rudoy et al. Time-varying autoregressions in speech: Detection theory and applications
Tsenov et al. Speech recognition using neural networks
CN105336344A (en) Noise detection method and apparatus thereof
CN104732984B (en) A kind of method and system of quick detection single-frequency prompt tone
CN110660396A (en) Language processing system and method based on MEMS
JP2008170789A (en) Voice section detection apparatus and voice section detection method
Raczynski Speech processing algorithm for isolated words recognition
Teacher et al. Experimental, limited vocabulary, speech recognizer
CN107039046B (en) Voice sound effect mode detection method based on feature fusion
Zhao et al. Online noise estimation using stochastic-gain HMM for speech enhancement
Ho et al. Naaloss: Rethinking the objective of speech enhancement
US20150364146A1 (en) Method for Providing Visual Feedback for Vowel Quality
Thakur et al. Speech enhancement using Open-Unmix music source separation architecture
KR101647059B1 (en) Independent vector analysis followed by HMM-based feature enhancement for robust speech recognition

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20200107

RJ01 Rejection of invention patent application after publication