WO2023216172A1 - Poultry voiceprint recognition method and system - Google Patents

Poultry voiceprint recognition method and system Download PDF

Info

Publication number
WO2023216172A1
WO2023216172A1 PCT/CN2022/092354 CN2022092354W WO2023216172A1 WO 2023216172 A1 WO2023216172 A1 WO 2023216172A1 CN 2022092354 W CN2022092354 W CN 2022092354W WO 2023216172 A1 WO2023216172 A1 WO 2023216172A1
Authority
WO
WIPO (PCT)
Prior art keywords
sound
poultry
state
recording information
voiceprint identification
Prior art date
Application number
PCT/CN2022/092354
Other languages
French (fr)
Chinese (zh)
Inventor
林玠佑
张光甫
黄醴万
Original Assignee
智逐科技股份有限公司
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by 智逐科技股份有限公司 filed Critical 智逐科技股份有限公司
Priority to PCT/CN2022/092354 priority Critical patent/WO2023216172A1/en
Publication of WO2023216172A1 publication Critical patent/WO2023216172A1/en

Links

Images

Classifications

    • GPHYSICS
    • G05CONTROLLING; REGULATING
    • G05DSYSTEMS FOR CONTROLLING OR REGULATING NON-ELECTRIC VARIABLES
    • G05D1/00Control of position, course, altitude or attitude of land, water, air or space vehicles, e.g. using automatic pilots
    • G05D1/02Control of position or course in two dimensions

Definitions

  • This case relates to a voiceprint identification method and system, and in particular, to a poultry voiceprint identification method and system.
  • this case proposes a poultry voiceprint identification method and system that can improve the aforementioned common problems.
  • a poultry voiceprint identification method including the following steps: receiving a recording information of a poultry house in a time period; and analyzing the recording information to determine a sound of the recording information
  • the sound state includes a normal poultry sound state or an abnormal poultry sound state.
  • the poultry voiceprint identification method further includes the following steps: converting the recorded information into several sound features, wherein the step of converting the recorded information into several sound features includes the following steps : filtering out the recorded information to generate filtered recorded information having a specific frequency range; dividing the filtered recorded information into a plurality of sound information; and extracting the plurality of sound features from the plurality of sound information , wherein the step of analyzing the recording information to determine the sound state of the recording information is analyzing each of the plurality of sound features to determine the sound state of each of the plurality of sound characteristics.
  • the step of filtering the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
  • the step of extracting the plurality of sound features from the plurality of sound information is implemented using openSMILE, Wavelet or short-time Fourier transform.
  • the poultry voiceprint identification method further includes the following steps: storing the voice status of the recording information in a database through a network; and allowing a user to view the recorded information through the network. the sound status in the database.
  • the step of analyzing each of the plurality of sound features to determine the sound state of each of the plurality of sound features is determined through an artificial intelligence sound model
  • the artificial intelligence sound model generates a training group based on the plurality of sound characteristics, and the training group includes an identification condition recording the normal poultry sound state and the abnormal poultry sound state.
  • the artificial intelligence sound model determines each of the several sound features as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state , the artificial intelligence sound model determines each of the plurality of sound features belonging to the poultry sound state as the positive poultry sound state or the abnormal poultry sound state via the support vector machine.
  • a poultry voiceprint recognition system which is installed on a server connected to a network and includes: a receiver, which is provided in a poultry house and is used to receive all the data in a time period. A recording information of the poultry house; and a feature analysis module for analyzing the recording information to determine a sound state of the recording information, where the sound state includes a normal poultry sound state or an abnormal poultry sound state.
  • the poultry voiceprint recognition system further includes: a feature processing module for converting the recorded information into several sound features; wherein the feature processing module further includes: a filtering unit , used to filter out the recorded information to generate filtered recorded information with a specific frequency range; a dividing unit used to divide the filtered recorded information into several pieces of sound information; and an extraction unit used to extract the filtered recorded information from the The plurality of sound features are extracted from the plurality of sound information, wherein the feature analysis module is used to analyze the plurality of sound features in the step of analyzing the recording information to determine the sound state of the recording information. Each to determine the sound state of each of the plurality of sound features.
  • the filtering unit is further configured to: the step of filtering the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
  • the extraction unit is further configured to: extract the plurality of sound features from the plurality of sound information by using openSMILE, Wavelet or short-time Fourier transform.
  • the poultry voiceprint recognition system further includes: a database for storing the voice status of the recording information through the network; and a user interface for using the The network provides a user with access to the sound status in the database.
  • the feature analysis module is further configured to: in the step of analyzing each of the several sound features to determine the sound state of each of the several sound features, Determined by an artificial intelligence sound model, the artificial intelligence sound model generates a training group based on the several sound characteristics and uses the training group to train the feature analysis module.
  • the training group includes records of the normal poultry sounds. status and an identification condition for the abnormal poultry sound status.
  • the artificial intelligence sound model determines each of the several sound features as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state , the artificial intelligence sound model determines each of the plurality of sound features belonging to the poultry sound state as the positive poultry sound state or the abnormal poultry sound state via the support vector machine.
  • abnormal poultry sounds were identified among the many sounds in intensively farmed poultry houses.
  • production of abnormal poultry sounds can be accurately and quickly identified using artificial intelligence, reducing the manpower required and enabling early detection of abnormal poultry sounds so that response measures can be taken as early as possible.
  • Figure 1 is a schematic diagram of a poultry voiceprint recognition system according to an embodiment of the present invention.
  • Figure 2 is a schematic diagram of a poultry voiceprint identification method according to an embodiment of the present case.
  • Figure 3 is a schematic diagram of a poultry voiceprint identification method according to another embodiment of the present invention.
  • Figure 4A shows a waveform diagram of original recording information of poultry according to an embodiment of the present invention.
  • Figure 4B shows a time-frequency diagram of original recording information of poultry according to an embodiment of the present invention.
  • Figure 5A shows a waveform diagram of the original recording information of poultry after band-pass filtering according to an embodiment of the present invention.
  • Figure 5B shows a time-frequency diagram of the original recording information of poultry after band-pass filtering according to an embodiment of the present invention.
  • Figure 6A shows a waveform diagram of the original recording information of poultry after spectral subtraction according to an embodiment of the present invention.
  • Figure 6B shows a time-frequency diagram of the original poultry recording information after spectral subtraction according to an embodiment of the present case.
  • Figure 7A shows a time-frequency diagram of poultry sounds in the original recording information according to an embodiment of the present invention.
  • FIG. 7B shows a time-frequency diagram of poultry sounds in the filtered recording information after band-pass filtering according to an embodiment of the present invention.
  • Figure 7C shows a time-frequency diagram of poultry sounds in the filtered recording information after spectral subtraction according to an embodiment of the present invention.
  • FIG. 8A shows a waveform diagram of poultry sounds in the filtered recording information after band-pass filtering and spectral subtraction processing according to an embodiment of the present invention.
  • Figure 8B shows a time-frequency diagram of poultry sounds in the filtered recording information after band-pass filtering and spectral subtraction processing according to an embodiment of the present invention.
  • Figure 9 is a schematic diagram of a portion of voice activity intercepted by VAD according to an embodiment of the present invention.
  • Figure 10A shows a sample of normal poultry sounds of native chickens according to an embodiment of the present invention.
  • Figure 10B shows a sample of abnormal poultry sounds of native chickens according to an embodiment of the present invention.
  • Figure 11A shows a sample of normal poultry sounds of laying hens according to an embodiment of the present invention.
  • Figure 11B shows a sample of abnormal poultry sounds of laying hens according to an embodiment of the present invention.
  • Figure 12 is a schematic diagram of a classification diagram of a support vector machine according to an embodiment of this case.
  • Figure 13 shows the temperature and humidity information in the poultry house of each age according to an embodiment of the present invention.
  • Figure 14 illustrates the prediction results of the number of voice data on the same day for each age according to an embodiment of the present case.
  • Figure 15 shows the prediction results of the number of abnormal poultry sounds of each day of age according to an embodiment of this case.
  • Figure 16 shows the proportion of each category in the prediction results of the day according to an embodiment of this case.
  • Figure 17A shows the prediction results of the number of sound data from 6 a.m. to 2 p.m. at each age according to an embodiment of the present case.
  • Figure 17B shows the prediction results of the number of sound data from 2 pm to 10 pm for each day according to an embodiment of the present case.
  • Figure 17C shows the prediction results of the number of sound data from 10 pm to 6 am the next day according to an embodiment of the present case.
  • Figure 18A shows the prediction results of the number of abnormal poultry sound data of each age from 6 am to 2 pm according to an embodiment of the present case.
  • Figure 18B shows the prediction results of the number of abnormal poultry sound data of each age from 2 pm to 10 pm according to an embodiment of the present case.
  • Figure 18C shows the prediction results of the number of abnormal poultry sound data from 10 pm to 6 am of each day of age according to an embodiment of the present case.
  • FIG. 1 illustrates a schematic diagram of a poultry voiceprint recognition system 100 according to an embodiment of the present invention.
  • the poultry voiceprint recognition system 100 is installed on a server (not shown) connected to a network (not shown).
  • the poultry voiceprint identification system 100 includes a receiver 110 and a processor 115.
  • the processor 115 includes a feature processing module 120 and a feature analysis module 130 .
  • the receiver 110 is disposed in a poultry house and is used for receiving a recording information IR of the poultry house in a time period.
  • the feature processing module 120 is used to transform the recording information IR into several sound features CS .
  • the feature analysis module 130 is used to analyze the recording information I R to determine the sound state S S of the recording information I R , or to determine the sound state S S of each of the plurality of sound features CS .
  • the sound state S S includes a normal poultry sound state and/or an abnormal poultry sound state .
  • the receiver 110 and the processor 115 may be integrated into a device (not shown; such as a single-chip microcomputer Raspberry Pi) provided in the poultry house.
  • the feature processing module 120 further includes a filtering unit 121, a segmentation unit 122 and an extraction unit 123.
  • the filtering unit 121 is used to filter out the recording information I R to generate filtered recording information I FR having a specific frequency range.
  • the dividing unit 122 is used to divide the filtered recording information I FR into several pieces of sound information IS .
  • the extraction unit 123 is used to extract several sound features CS from several pieces of sound information IS .
  • the poultry voiceprint recognition system 100 further includes a database 140 and a user interface 150.
  • the database 140 is used to store the sound state S S of the recording information I R (or the sound state S S of each of the plurality of sound features C S ) through the network.
  • the user interface 150 is used for a user to view the sound status S S in the database 140 through the network.
  • the feature analysis module 130 is further configured to make a determination through an artificial intelligence sound model 160 in the step of analyzing each of the plurality of sound characteristics CS to determine the sound state S S of each of the plurality of sound characteristics CS ,
  • the artificial intelligence sound model 160 generates a training group T based on the plurality of sound features CS and uses the training group T to train the feature analysis module 120.
  • the training group T includes an identification condition that records normal poultry sound states and abnormal poultry sound states.
  • the artificial intelligence sound model 160 determines each of the plurality of sound features C S as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state, the artificial intelligence sound model uses a support vector machine to determine whether it belongs to the poultry sound state. Each of the several sound characteristics CS of the sound state is determined to be a positive poultry sound state or an abnormal poultry sound state.
  • FIG. 2 illustrates a schematic diagram of a poultry voiceprint identification method 200 according to an embodiment of the present case, which at least includes steps S210 and S220, as detailed below.
  • step S210 please also refer to FIG. 1.
  • the receiver 110 receives a recording information IR of a poultry house in a time period.
  • step S220 the feature analysis module 130 analyzes and determines a sound state S S of the recording information IR .
  • FIG. 3 illustrates a schematic diagram of a poultry voiceprint identification method 300 according to another embodiment of the present invention, which at least includes steps S310, S320, S330, S340, S350 and S360, as detailed below.
  • the receiver 110 receives a recording information IR of a poultry house in a time period.
  • the receiver 110 is at least one of an omnidirectional microphone and a directional microphone.
  • receiver 110 is preferably a directional microphone.
  • the recording information IR received by the receiver 110 may include a waveform diagram or a time-frequency diagram.
  • the waveform of the recording information IR presented in this waveform diagram is not obvious due to the interference of noise.
  • the recording information IR presented in this frequency diagram includes background noise with a frequency similar to that of poultry sounds.
  • the background noise includes but is not limited to the sound of fans running, the sound of cars and rain outside the poultry house. However, these background noises will cause errors in subsequent feature value calculations and affect the classification results. Therefore, the recording information IR needs to be filtered before analysis.
  • step S320 the feature processing module 120 transforms the recording information IR into several sound features CS .
  • Step S320 also includes step S321, step S322 and step S323.
  • the feature processing module 120 includes a filtering unit 121, a segmentation unit 122 and an extraction unit 123.
  • step S321 the filtering unit 121 filters out the recording information I R to generate filtered recording information I FR having a specific frequency range.
  • the filtered recording information I FR has a specific frequency range between about 500 Hz and about 5 kHz.
  • methods for implementing step S321 include, but are not limited to, using filters, spectral subtraction, spectral gating, noise gating, and multi-microphone noise reduction.
  • filters include but are not limited to adapter filters, finite impulse filters (FIR) or infinite impulse filters (IIR) ), where the finite impulse filter (FIR) or the infinite impulse filter (IIR) is, for example, a high-pass filter (high-pass filter), a band-pass filter (band-pass filter), a low-pass filter (low-pass filter) ) or band-stop filter.
  • FIR finite impulse filters
  • IIR infinite impulse filters
  • the filtering unit 121 includes, for example, a Butterworth filter, and performs step S321 with bandpass filtering.
  • the amplitude gain of the n-order Butterworth low-pass filter can be expressed as the following formula (1) , where Ha is the transfer function, N is the order of the filter, ⁇ is the angular frequency of the signal, and ⁇ c is the cut-off frequency when the amplitude drops by 3dB.
  • Ha is the transfer function
  • N is the order of the filter
  • the angular frequency of the signal
  • ⁇ c the cut-off frequency when the amplitude drops by 3dB.
  • the order of the filter is higher, the amplitude of the filter attenuates faster in the stop band, and the filtering effect is better. Frequencies lower than ⁇ c will pass with gain, and frequencies higher than ⁇ c will be suppressed.
  • the step S321 is performed by spectral subtraction, as shown in Figures 6A to 6B.
  • Spectral subtraction is based on a simple assumption, and this assumption is that the noise in the speech signal only exists additive noise. As long as the spectrum of the noisy signal is subtracted from the spectrum of the noise signal, a relatively clean pure speech spectrum can be obtained.
  • the signal model in the time domain can be expressed as the following formula (2), where y(m) is the noisy signal, x(m) is the additive noise, d(m) is the pure speech signal, and m is time.
  • the background noise after spectral subtraction (shown in Figure 6B) is suppressed to a much greater extent than after band-pass filtering (shown in Figure 5B). Therefore, compared with band-pass filtering, spectral subtraction can effectively eliminate background noise that is similar to the frequency of poultry sounds (that is, it can effectively eliminate background noise within the band-pass band). However, please refer to Figure 6B. Compared with band-pass filtering, spectral subtraction cannot effectively suppress low-frequency noise. In addition, please refer to Figures 7A to 7C. When the signal-to-noise ratio is too low, spectral subtraction may cause sound distortion after filtering (as shown in Figure 7C).
  • the step S321 is performed using bandpass filtering and spectral subtraction. Please refer to Figure 8A to Figure 8B. After most of the high- and low-frequency noise is suppressed through band-pass filtering, spectral subtraction is then used to suppress the background noise in the pass-band. This also reduces the distortion of poultry sounds that may be caused by spectral subtraction. (As shown in Figure 8B).
  • step S322 the dividing unit 122 divides the filtered recording information I FR into several pieces of sound information IS .
  • the recording information I R and the filtered recording information I FR are files with continuous recording duration of 5 minutes, including many normal poultry sound clips, abnormal poultry sound clips, non-poultry sound clips, no sound clips, etc. This makes it difficult to define the classification of this file, causing the subsequent extraction of acoustic features in this file to lose its unity. Therefore, it is necessary to segment the filtered recording information I FR to reduce the number of poultry sounds contained in a single information file.
  • methods for implementing step S322 include, but are not limited to, voice activity detection (VAD), autocorrelation function (autocorrelation function, ACF), or other voice feature recognition methods.
  • step S322 is performed using Voice Activity Detection (VAD).
  • VAD Voice Activity Detection
  • Voice activity detection is based on a frame length of 10 milliseconds, and is characterized by six sub-band energies, namely 80 ⁇ 250Hz, 250 ⁇ 500Hz, 500 ⁇ 1000Hz, 1 ⁇ 2kHz, 2 ⁇ 3kHz, 3 ⁇ 4kHz, for the entire segment Filter the recording information I FR for detection, and calculate the probability that each frame is speech and noise respectively.
  • the final criterion for judging that there is speech activity is that the total likelihood ratio of speech in any subband or six subbands is greater than 0.9, then it is judged as having speech activity. Voice activities.
  • the program when voice activity is detected, the program will start recording until the voice likelihood ratio is detected to be less than 0.9, that is, when there is no voice activity, the program will end the recording and intercept the documentary segment into a new one with a shorter duration.
  • Audio files for example, files usually within 2 seconds. Please refer to Figure 9.
  • the number of poultry sounds contained in these intercepted sound information IS is relatively small. Compared with the original 5-minute recording information I R and the filtered recording information I FR , it is more single and can target poultry. Sounds are classified more accurately.
  • step S323 the extraction unit 123 extracts several sound features CS from several pieces of sound information IS .
  • short-term analysis is usually the main method. Because the audio changes greatly and the recording environment of this study is in a commercial poultry house, the audio content is more complex, so short-term data analysis is relatively stable. , and short-term analysis usually calculates the feature values of a piece of audio content in units of sound frames (Frame) with a length of about 30 milliseconds.
  • Step S323 may also include a feature acquisition method, a feature transformation method or a feature extraction method.
  • the feature acquisition method in step S323 includes but is not limited to signal transformation (such as wavelet analysis (Wavelet Analysis), power spectral density (PSD)), openSMILE human emotion feature set, process Zero-crossing rate (ZCR), short-time Fourier transform, Fast Fourier Transform (FFT) (such as energy intensity, fundamental frequency), Mel-frequency cepstral coefficient (MFCC) ), Cepstral peak prominence (CPP) and Welchz method.
  • the feature transformation method in step S323 includes but is not limited to standardization, normalization, binarization or other numerical transformation, scaling, and function transformation methods.
  • the feature extraction methods in step S323 include but are not limited to principal component analysis (PCA), linear discriminant analysis (LDA), local linear embedding (Local Linear Embedding) , LLE), Laplacian Eigenmap (LE), Stochastic Neighbor Embedding (SNE), T-Stochastic Neighbor Embedding (T-SNE), kernel Principal component analysis (KPCA), transfer component analysis (TCA) or other feature dimensionality reduction and feature extraction methods.
  • PCA principal component analysis
  • LDA linear discriminant analysis
  • LLE Laplacian Eigenmap
  • SNE Stochastic Neighbor Embedding
  • T-Stochastic Neighbor Embedding T-Stochastic Neighbor Embedding
  • KPCA kernel Principal component analysis
  • TCA transfer component analysis
  • openSMILE is used to perform step S323.
  • openSMILE is an open source toolkit suitable for signal processing and audio acoustic feature extraction. This embodiment uses the "The INTERSPEECH 2009 Emotion Challenge feature set" feature set, and a total of 384 acoustic features are extracted for each audio file, including the root-mean-square signal frame energy (root-mean-square signal frame). energy), Mel-Frequency cepstral coefficients 1-12, zero-crossing rate of time signal, the voicing probability calculated from the ACF ACF), the fundamental frequency computed from the Cepstrum, etc., a total of 16 low-level descriptors (LLDs).
  • LLDs low-level descriptors
  • Mel-Frequency cepstral coefficients are a set of key coefficients used to establish Mel-Frequency cepstral coefficients.
  • Mel-Frequency cepstral coefficients are a spectrum used to represent short-term audio. The principle is based on the use of nonlinear Mel scale representation of the logarithmic spectrum and its linear cosine transform. Its biggest feature is that the frequency bands on the Mel cepstrum are evenly distributed on the Mel scale. This representation method is closer to the human nonlinear auditory system. This acoustic feature also takes into account the human ear's perception of different frequencies. , so it is particularly suitable for speech recognition, where the transformation between the Mel scale (m) and the actual frequency (f) can be expressed as the following formula (3), where f is the actual frequency value. Its reference point defines 1000Hz as 1000mel.
  • step S330 the artificial intelligence sound model 160 generates a training group T according to several sound features CS .
  • the artificial intelligence sound model 160 uses the training group T to train the feature analysis module 130.
  • the training group T includes an identification condition that records normal poultry sound states, abnormal poultry sound states, and non-poultry sound states.
  • methods for implementing step S330 include, but are not limited to, supervised learning, unsupervised learning, semi-supervised learning, and reinforcement learning. (reinforcement learning).
  • Supervised learning includes but is not limited to classifiers (classification) and regression (regression).
  • Classifiers include but are not limited to random forest (random forest), K-Nearest Neighbor (k-NN), support vector machine ( support vector machine (SVM), artificial neural network (ANN), support vector domain description (SVDD), sparse representation classifier (SRC).
  • Unsupervised learning includes but is not limited to clustering and dimensionality reduction.
  • FIGS. 10A and 10B show normal poultry sounds and abnormal poultry sounds collected in an experimental example.
  • the normal sound data of a 20-day-old native chicken is displayed. It has been confirmed by practitioners that the normal sound frequency of a native chicken is between about 2.5kHz and about 3.5kHz, and the duration of a single sound is about 0.1 seconds.
  • abnormal sound data of a 20-day-old native chicken is displayed. The frequency of the abnormal sound is between about 1 kHz and about 1.5 kHz, and the duration of this single sound is between about 0.67 seconds and 0.7 seconds.
  • this abnormal sound was a symptom of rales.
  • One of the characteristics of rales is a prolonged sound caused by excessive mucus in the trachea and obstruction.
  • Figures 11A to 11B show normal poultry sounds and abnormal poultry sounds collected in another experimental example.
  • the normal sound data of a 19-day-old native chicken is displayed. It has been confirmed by practitioners that the normal sound frequency of a native chicken is between about 2.5kHz and about 4kHz, and the duration of a single sound is about 0.15 seconds.
  • the abnormal sound data of a 20-day-old native chicken is displayed. The frequency of the abnormal sound is between about 1kHz and about 1.5kHz, and the duration of this single sound is about 0.7 seconds, which was confirmed by practitioners. , this abnormal sound is a symptom of rales. One of the characteristics of rales is a prolonged sound caused by excessive mucus in the trachea and obstruction.
  • the frequency difference between normal poultry sounds and abnormal poultry sounds is about 2 kHz, and there is a difference between sound durations of about 0.5 seconds to 0.6 seconds. Therefore, for example, the identification condition in step S330 may be set according to the frequency difference or duration difference of the sounds.
  • FIG. 12 illustrates a classification diagram 1200 of a support vector machine according to an embodiment of the present application.
  • the artificial intelligence sound model 160 determines each of the several sound features C S as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state, the artificial intelligence sound model 160 uses a support vector machine to determine whether it belongs to the poultry sound state. Each of the several sound characteristics CS of the poultry sound state is determined to be a positive poultry sound state or an abnormal poultry sound state.
  • the artificial intelligence sound model 160 uses a total of 150 pieces of normal poultry sound data, 150 pieces of abnormal poultry sound data, and 150 pieces of non-poultry sound data to train the model.
  • the three types of training set data are shown in Table 2. .
  • the trained artificial intelligence sound model 160 has a verification accuracy of 84.2% in identifying these three types of data.
  • the verification results of the artificial intelligence sound model 160 are shown in Table 3.
  • step S340 the feature analysis module 130 analyzes and determines a sound state S S of each of the plurality of sound features C S .
  • the step of analyzing each of the plurality of sound characteristics CS to determine the sound state S S of each of the plurality of sound characteristics CS is through an artificial intelligence sound model 160 (eg, step S330 shown).
  • step S350 the sound state S S of the recording information I R (or the sound state S S of each of the plurality of sound features C S ) is stored in a database 140 through a network.
  • the database 140 is a cloud database for storing historical information of the sound state S S.
  • step S360 a user is allowed to view the sound status S S in the database 140 through the network.
  • Figure 13 illustrates the temperature and humidity information in the poultry house of each age according to an embodiment of the present case.
  • Figure 14 illustrates the prediction results of the number of sound data on the same day for each age according to an embodiment of the present case.
  • Figure 15 shows the prediction results of the number of abnormal poultry sounds of each age on the day according to an embodiment of this case.
  • Figure 16 shows the proportion of each category in the prediction results of the day according to an embodiment of this case.
  • the human observations shown in Figures 14 to 16 are The abnormality day D is the date when the practitioner observed abnormal poultry sounds (such as rales in the poultry), and the artificial observation of the abnormality day D is also applicable to the description of the subsequent figures, so it will be explained first.
  • Figure 17A shows the prediction results of the number of sound data from 6 a.m. to 2 p.m. for each day age according to an embodiment of the present case.
  • Figure 17B shows the prediction results for the sound data quantity from 2 p.m. for each day age according to an embodiment of the present case.
  • FIG. 17C shows the prediction results of the number of sound data from 10 p.m. to 6 a.m. the next day according to an embodiment of the present case.
  • Figure 18A shows the prediction results of the number of abnormal poultry sound data from 6 am to 2 pm for each age according to an embodiment of this case.
  • Figure 18B shows the prediction results of the number of abnormal poultry sound data from 2 pm to 10 pm of each age according to an embodiment of this case
  • Figure 18C shows the prediction results of the number of abnormal poultry sound data from 10 pm to 6 am of each day of age according to an embodiment of this case. Prediction results of the number of abnormal poultry sound data points.
  • the prediction results of abnormal poultry sounds by the system and method proposed in this case can be provided to a user (such as a practitioner) as a basis for assessing the health status of poultry.
  • the system and method proposed in this case can observe abnormal poultry sounds at 18 days of age, which can help users (such as practitioners) take measures faster.
  • this case discloses a poultry voiceprint recognition system and method that can receive a recording information of a poultry house in a time period and convert the recording information into an image (such as a waveform diagram or a time-frequency diagram and other planar images). ), analyze the recording information based on several image indicators of the image (such as frequency gap or duration gap and other identification conditions) to determine the sound status of the recording information.
  • the sound status includes normal poultry sound status and/or abnormal poultry sound status.
  • the above-mentioned step of analyzing the recording information based on several image indicators of the image to determine the sound state of the recording information can also be implemented in conjunction with an artificial intelligence sound model.
  • the poultry voiceprint identification system and method based on the embodiment of this case can accurately and quickly identify through the improvement of computer software/hardware and the combination of artificial intelligence sound models, even when the poultry house receives recording information with many mixed sounds. The production of abnormal poultry sounds.

Landscapes

  • Engineering & Computer Science (AREA)
  • Aviation & Aerospace Engineering (AREA)
  • Radar, Positioning & Navigation (AREA)
  • Remote Sensing (AREA)
  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Automation & Control Theory (AREA)
  • Housing For Livestock And Birds (AREA)

Abstract

A poultry voiceprint recognition system (100), installed on a server connected to a network. The poultry voiceprint recognition system (100) comprises a receiver (100), a feature processing module (120), a feature analysis module (130), and an artificial intelligence voice model (160). The receiver (100) is arranged in a poultry house and is used for receiving recording information of the poultry house in a period of time. The feature processing module (120) is used for converting the recording information into a plurality of voice features. The feature analysis module (130) is used for analyzing each of the plurality of voice features so as to determine a voice state of each of the plurality of voice features by means of the artificial intelligence voice model (160). The voice state comprises a normal poultry voice state or an abnormal poultry voice state. The artificial intelligence voice model (160) generates a training set according to the plurality of voice features.

Description

家禽声纹辨识方法及系统Poultry voiceprint identification method and system 技术领域Technical field
本案是有关于一种声纹辨识方法及系统,且特别是有关于一种家禽声纹辨识方法及系统。This case relates to a voiceprint identification method and system, and in particular, to a poultry voiceprint identification method and system.
背景技术Background technique
现代家禽产业大多规模化并密集性养殖,促进了疾病的传播,当禽舍内有家禽感染时,疾病于禽舍内扩散至整批家禽的速度非常快,因此每年都对家禽产业带来巨大的经济损失。这些疾病由病毒方面所引起的主要有家禽流行性感冒(Avian Influenza,AI)、新城病(Newcastle Disease,ND)、传染性支气管炎(Infectious Bronchitis,IB)、传染性喉头气管炎(Infectious Laryngotracheitis)等,而其中传染性支气管炎(IB)又为亚洲地区最重要的呼吸道疾病之一。一般而言,受感染的家禽会有气管啰音、咳嗽、喷嚏、流鼻涕、产蛋率下降及饲料换肉率下降等状况产生。此外,通常在衍生成较为严重的症状前,受感染的家禽在声音上都会先有所变化。若能提早发现家禽声音的改变,尤其是发现受感染的家禽声音,将提供一个很好的预警效果。Most modern poultry industries are large-scale and intensively farmed, which promotes the spread of diseases. When poultry is infected in a poultry house, the disease spreads very quickly in the poultry house to the entire batch of poultry. Therefore, it has a huge impact on the poultry industry every year. economic losses. These diseases are mainly caused by viruses: Avian Influenza (AI), Newcastle Disease (ND), Infectious Bronchitis (IB), Infectious Laryngotracheitis (Infectious Laryngotracheitis) Among them, infectious bronchitis (IB) is one of the most important respiratory diseases in Asia. Generally speaking, infected poultry will have tracheal rales, coughing, sneezing, runny nose, reduced egg production rate and reduced feed meat conversion rate. In addition, infected birds often show changes in their vocalizations before developing more severe symptoms. If changes in the sounds of poultry can be detected early, especially the sounds of infected poultry, it will provide a good early warning effect.
技术问题technical problem
然而,对于家禽疾病的观察,常仰赖人工经验判断。此外,由于现代家禽产业大多规模化并密集性养殖,常交杂许多家禽声音及非家禽声音,故又更难以从众多声音当中辨识出异常家禽声音。However, the observation of poultry diseases often relies on manual empirical judgment. In addition, since most modern poultry industries are large-scale and intensively farmed, many poultry sounds and non-poultry sounds are often mixed together, so it is more difficult to identify abnormal poultry sounds among the many sounds.
技术解决方案Technical solutions
因此,本案提出一种家禽声纹辨识方法及系统,可改善前述习知问题。Therefore, this case proposes a poultry voiceprint identification method and system that can improve the aforementioned common problems.
根据本案的一实施例,提出一种家禽声纹辨识方法,包括以下步骤:接收在一时间段中的一禽舍的一录音信息;以及分析所述录音信息以判断所述录音信息的一声音状态,所述声音状态包括一正常家禽声音状态或一异常家禽声音状态。According to an embodiment of this case, a poultry voiceprint identification method is proposed, including the following steps: receiving a recording information of a poultry house in a time period; and analyzing the recording information to determine a sound of the recording information The sound state includes a normal poultry sound state or an abnormal poultry sound state.
在本案的一些实施例中,所述家禽声纹辨识方法更包括以下步骤:将所述录音信息变换成数个声音特征,其中在将所述录音信息变换成数个声音特征的步骤包括以下步骤:滤除所述录音信息以产生具有一特定频率范围的一过滤录音信息;将所述过滤录音信息分割成数个声音信息;以及从所述数个声音信息中提取出所 述数个声音特征,其中在分析所述录音信息以判断所述录音信息的所述声音状态的步骤为分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态。In some embodiments of this case, the poultry voiceprint identification method further includes the following steps: converting the recorded information into several sound features, wherein the step of converting the recorded information into several sound features includes the following steps : filtering out the recorded information to generate filtered recorded information having a specific frequency range; dividing the filtered recorded information into a plurality of sound information; and extracting the plurality of sound features from the plurality of sound information , wherein the step of analyzing the recording information to determine the sound state of the recording information is analyzing each of the plurality of sound features to determine the sound state of each of the plurality of sound characteristics. .
在本案的一些实施例中,在滤除所述录音信息以产生具有所述特定频率范围的所述过滤录音信息的步骤是以带通滤波及谱减法实现。In some embodiments of this case, the step of filtering the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
在本案的一些实施例中,在从所述数个声音信息中提取出所述数个声音特征的步骤是以openSMILE、Wavelet或短时距傅立叶变换实现。In some embodiments of this case, the step of extracting the plurality of sound features from the plurality of sound information is implemented using openSMILE, Wavelet or short-time Fourier transform.
在本案的一些实施例中,所述家禽声纹辨识方法更包括以下步骤:透过一网络将所述录音信息的所述声音状态储存于一数据库;以及透过所述网络供一用户查看所述数据库中的所述声音状态。In some embodiments of this case, the poultry voiceprint identification method further includes the following steps: storing the voice status of the recording information in a database through a network; and allowing a user to view the recorded information through the network. the sound status in the database.
在本案的一些实施例中,在分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态的步骤中是经由一人工智能声音模型所判断,所述人工智能声音模型根据所述数个声音特征产生一训练组,所述训练组包括记载所述正常家禽声音状态及所述异常家禽声音状态的一辨识条件。In some embodiments of the present case, the step of analyzing each of the plurality of sound features to determine the sound state of each of the plurality of sound features is determined through an artificial intelligence sound model, The artificial intelligence sound model generates a training group based on the plurality of sound characteristics, and the training group includes an identification condition recording the normal poultry sound state and the abnormal poultry sound state.
在本案的一些实施例中,所述人工智能声音模型经由一支持向量机将所述数个声音特征的每一者判断为一家禽声音状态或一非家禽声音状态,若属于所述家禽声音状态,所述人工智能声音模型经由所述支持向量机再将属于所述家禽声音状态的所述数个声音特征的每一者判断为所述正家禽声音状态或所述异常家禽声音状态。In some embodiments of this case, the artificial intelligence sound model determines each of the several sound features as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state , the artificial intelligence sound model determines each of the plurality of sound features belonging to the poultry sound state as the positive poultry sound state or the abnormal poultry sound state via the support vector machine.
根据本案的另一实施例,提出一种家禽声纹辨识系统,安装于连接至一网络的一伺服器,包括:一接收器,设置于一禽舍中,用以接收一时间段中的所述禽舍的一录音信息;以及一特征分析模块,用以分析所述录音信息以判断所述录音信息的一声音状态,所述声音状态包括一正常家禽声音状态或一异常家禽声音状态。According to another embodiment of the present case, a poultry voiceprint recognition system is proposed, which is installed on a server connected to a network and includes: a receiver, which is provided in a poultry house and is used to receive all the data in a time period. A recording information of the poultry house; and a feature analysis module for analyzing the recording information to determine a sound state of the recording information, where the sound state includes a normal poultry sound state or an abnormal poultry sound state.
在本案的一些实施例中,所述家禽声纹辨识系统更包括:一特征处理模块,用以将所述录音信息变换成数个声音特征;其中所述特征处理模块更包括:一滤除单元,用以滤除所述录音信息以产生具有一特定频率范围的一过滤录音信息;一分割单元,用以将所述过滤录音信息分割成数个声音信息;以及一提取单元,用以从所述数个声音信息中提取出所述数个声音特征,其中所述特征分析模块在 分析所述录音信息以判断所述录音信息的所述声音状态的步骤用以分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态。In some embodiments of this case, the poultry voiceprint recognition system further includes: a feature processing module for converting the recorded information into several sound features; wherein the feature processing module further includes: a filtering unit , used to filter out the recorded information to generate filtered recorded information with a specific frequency range; a dividing unit used to divide the filtered recorded information into several pieces of sound information; and an extraction unit used to extract the filtered recorded information from the The plurality of sound features are extracted from the plurality of sound information, wherein the feature analysis module is used to analyze the plurality of sound features in the step of analyzing the recording information to determine the sound state of the recording information. Each to determine the sound state of each of the plurality of sound features.
在本案的一些实施例中,所述滤除单元更用以:在滤除所述录音信息以产生具有所述特定频率范围的所述过滤录音信息的步骤是以带通滤波及谱减法实现。In some embodiments of this case, the filtering unit is further configured to: the step of filtering the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
在本案的一些实施例中,所述提取单元更用以:在从所述数个声音信息中提取出所述数个声音特征的步骤是以openSMILE、Wavelet或短时距傅立叶变换实现。In some embodiments of this case, the extraction unit is further configured to: extract the plurality of sound features from the plurality of sound information by using openSMILE, Wavelet or short-time Fourier transform.
在本案的一些实施例中,所述家禽声纹辨识系统更包括:一数据库,用以透过所述网络储存所述录音信息的所述声音状态;以及一用户介面,用以透过所述网络供一用户查看所述数据库中的所述声音状态。In some embodiments of this case, the poultry voiceprint recognition system further includes: a database for storing the voice status of the recording information through the network; and a user interface for using the The network provides a user with access to the sound status in the database.
在本案的一些实施例中,所述特征分析模块更用以:在分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态的步骤中是经由一人工智能声音模型所判断,所述人工智能声音模型根据所述数个声音特征产生一训练组并使用所述训练组训练所述特征分析模块,所述训练组包括记载所述正常家禽声音状态及所述异常家禽声音状态的一辨识条件。In some embodiments of the present case, the feature analysis module is further configured to: in the step of analyzing each of the several sound features to determine the sound state of each of the several sound features, Determined by an artificial intelligence sound model, the artificial intelligence sound model generates a training group based on the several sound characteristics and uses the training group to train the feature analysis module. The training group includes records of the normal poultry sounds. status and an identification condition for the abnormal poultry sound status.
在本案的一些实施例中,所述人工智能声音模型经由一支持向量机将所述数个声音特征的每一者判断为一家禽声音状态或一非家禽声音状态,若属于所述家禽声音状态,所述人工智能声音模型经由所述支持向量机再将属于所述家禽声音状态的所述数个声音特征的每一者判断为所述正家禽声音状态或所述异常家禽声音状态。In some embodiments of this case, the artificial intelligence sound model determines each of the several sound features as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state , the artificial intelligence sound model determines each of the plurality of sound features belonging to the poultry sound state as the positive poultry sound state or the abnormal poultry sound state via the support vector machine.
有益效果beneficial effects
在本案中,透过在一时间段中的一禽舍的一录音信息,并将录音信息过滤、分割、提取或转变成图像,以判断录音信息的声音状态,能够在现代家禽产业大多规模化并密集性养殖的禽舍中的众多声音当中辨识出异常家禽声音。此外,能够以人工智能方式准确地及快速地辨识出异常家禽声音的产生,减少所需人力并能提早发现异常家禽声音而尽早采取应对措施。In this case, by filtering, segmenting, extracting or converting the recorded information from a poultry house in a time period to determine the sound state of the recorded information, it can be used on a large scale in most modern poultry industries. And abnormal poultry sounds were identified among the many sounds in intensively farmed poultry houses. In addition, the production of abnormal poultry sounds can be accurately and quickly identified using artificial intelligence, reducing the manpower required and enabling early detection of abnormal poultry sounds so that response measures can be taken as early as possible.
附图说明Description of the drawings
为了对本案的上述及其他方面有更佳的了解,下文特举实施例,并配合所附图式详细说明如下:In order to have a better understanding of the above and other aspects of this case, examples are given below, and detailed descriptions are as follows with the accompanying drawings:
图1绘示依照本案一实施例的家禽声纹辨识系统的示意图。Figure 1 is a schematic diagram of a poultry voiceprint recognition system according to an embodiment of the present invention.
图2绘示依照本案一实施例的家禽声纹辨识方法的示意图。Figure 2 is a schematic diagram of a poultry voiceprint identification method according to an embodiment of the present case.
图3绘示依照本案另一实施例的家禽声纹辨识方法的示意图。Figure 3 is a schematic diagram of a poultry voiceprint identification method according to another embodiment of the present invention.
图4A绘示依照本案一实施例的家禽原始录音信息的波形图。Figure 4A shows a waveform diagram of original recording information of poultry according to an embodiment of the present invention.
图4B绘示依照本案一实施例的家禽原始录音信息的时频图。Figure 4B shows a time-frequency diagram of original recording information of poultry according to an embodiment of the present invention.
图5A绘示依照本案一实施例的经带通滤波后的家禽原始录音信息的波形图。Figure 5A shows a waveform diagram of the original recording information of poultry after band-pass filtering according to an embodiment of the present invention.
图5B绘示依照本案一实施例的经带通滤波后的家禽原始录音信息的时频图。Figure 5B shows a time-frequency diagram of the original recording information of poultry after band-pass filtering according to an embodiment of the present invention.
图6A绘示依照本案一实施例的经谱减法后的家禽原始录音信息的波形图。Figure 6A shows a waveform diagram of the original recording information of poultry after spectral subtraction according to an embodiment of the present invention.
图6B绘示依照本案一实施例的经谱减法后的家禽原始录音信息的时频图。Figure 6B shows a time-frequency diagram of the original poultry recording information after spectral subtraction according to an embodiment of the present case.
图7A绘示依照本案一实施例的原始录音信息中的家禽声音的时频图。Figure 7A shows a time-frequency diagram of poultry sounds in the original recording information according to an embodiment of the present invention.
图7B绘示依照本案一实施例的经带通滤波后的过滤录音信息中的家禽声音的时频图。FIG. 7B shows a time-frequency diagram of poultry sounds in the filtered recording information after band-pass filtering according to an embodiment of the present invention.
图7C绘示依照本案一实施例的经谱减法后的过滤录音信息中的家禽声音的时频图。Figure 7C shows a time-frequency diagram of poultry sounds in the filtered recording information after spectral subtraction according to an embodiment of the present invention.
图8A绘示依照本案一实施例的经带通滤波及谱减法处理后的过滤录音信息中的家禽声音的波形图。8A shows a waveform diagram of poultry sounds in the filtered recording information after band-pass filtering and spectral subtraction processing according to an embodiment of the present invention.
图8B绘示依照本案一实施例的经带通滤波及谱减法处理后的过滤录音信息中的家禽声音的时频图。Figure 8B shows a time-frequency diagram of poultry sounds in the filtered recording information after band-pass filtering and spectral subtraction processing according to an embodiment of the present invention.
图9绘示依照本案一实施例的VAD截取有语音活动部分示意图。Figure 9 is a schematic diagram of a portion of voice activity intercepted by VAD according to an embodiment of the present invention.
图10A绘示依照本案一实施例的土鸡的正常家禽声音的样本。Figure 10A shows a sample of normal poultry sounds of native chickens according to an embodiment of the present invention.
图10B绘示依照本案一实施例的土鸡的异常家禽声音的样本。Figure 10B shows a sample of abnormal poultry sounds of native chickens according to an embodiment of the present invention.
图11A绘示依照本案一实施例的蛋鸡的正常家禽声音的样本。Figure 11A shows a sample of normal poultry sounds of laying hens according to an embodiment of the present invention.
图11B绘示依照本案一实施例的蛋鸡的异常家禽声音的样本。Figure 11B shows a sample of abnormal poultry sounds of laying hens according to an embodiment of the present invention.
图12绘示依照本案一实施例的支持向量机的分类示意图。Figure 12 is a schematic diagram of a classification diagram of a support vector machine according to an embodiment of this case.
图13绘示出依照本发明一实施例的各日龄禽舍内温度及湿度信息。Figure 13 shows the temperature and humidity information in the poultry house of each age according to an embodiment of the present invention.
图14绘示依照本案一实施例的各日龄当天声音数据数量预测结果。Figure 14 illustrates the prediction results of the number of voice data on the same day for each age according to an embodiment of the present case.
图15绘示依照本案一实施例的各日龄当天异常家禽声音数量预测结果。Figure 15 shows the prediction results of the number of abnormal poultry sounds of each day of age according to an embodiment of this case.
图16绘示依照本案一实施例的各类别所占当天预测结果比例。Figure 16 shows the proportion of each category in the prediction results of the day according to an embodiment of this case.
图17A绘示依照本案一实施例的各日龄早上6点至下午2点的声音数据数量预测结果。Figure 17A shows the prediction results of the number of sound data from 6 a.m. to 2 p.m. at each age according to an embodiment of the present case.
图17B绘示依照本案一实施例的各日龄下午2点至晚上10点的声音数据数量预测结果。Figure 17B shows the prediction results of the number of sound data from 2 pm to 10 pm for each day according to an embodiment of the present case.
图17C绘示依照本案一实施例的晚上10点至隔日早上6点的声音数据数量预测结果。Figure 17C shows the prediction results of the number of sound data from 10 pm to 6 am the next day according to an embodiment of the present case.
图18A绘示依照本案一实施例的各日龄早上6点至下午2点的异常家禽声音数据数量预测结果。Figure 18A shows the prediction results of the number of abnormal poultry sound data of each age from 6 am to 2 pm according to an embodiment of the present case.
图18B绘示依照本案一实施例的各日龄下午2点至晚上10点的异常家禽声音数据数量预测结果。Figure 18B shows the prediction results of the number of abnormal poultry sound data of each age from 2 pm to 10 pm according to an embodiment of the present case.
图18C绘示依照本案一实施例的各日龄晚上10点至隔日早上6点的异常家禽声音数据数量预测结果。Figure 18C shows the prediction results of the number of abnormal poultry sound data from 10 pm to 6 am of each day of age according to an embodiment of the present case.
本发明的最佳实施方式Best Mode of Carrying Out the Invention
请参照图1,其绘示依照本案一实施例的家禽声纹辨识系统100的示意图。家禽声纹辨识系统100安装于连接至一网络(未绘示)的一伺服器(未绘示)。家禽声纹辨识系统100包括接收器110及处理器115。处理器115包括特征处理模块120及特征分析模块130。接收器110设置于一禽舍中,用以接收一时间段中的禽舍的一录音信息I R。特征处理模块120用以将录音信息I R变换成数个声音特征C S。特征分析模块130用以分析录音信息I R以判断录音信息I R的声音状态S S,或是用以数个声音特征C S的每一者以判断数个声音特征C S的每一者的声音状态S S,声音状态S S包括一正常家禽声音状态及/或一异常家禽声音状态。在一些实施例中,接收器110及处理器115可整合设置于禽舍中的一装置(未绘示;例如单晶片微电脑树莓派)中。 Please refer to FIG. 1 , which illustrates a schematic diagram of a poultry voiceprint recognition system 100 according to an embodiment of the present invention. The poultry voiceprint recognition system 100 is installed on a server (not shown) connected to a network (not shown). The poultry voiceprint identification system 100 includes a receiver 110 and a processor 115. The processor 115 includes a feature processing module 120 and a feature analysis module 130 . The receiver 110 is disposed in a poultry house and is used for receiving a recording information IR of the poultry house in a time period. The feature processing module 120 is used to transform the recording information IR into several sound features CS . The feature analysis module 130 is used to analyze the recording information I R to determine the sound state S S of the recording information I R , or to determine the sound state S S of each of the plurality of sound features CS . The sound state S S includes a normal poultry sound state and/or an abnormal poultry sound state . In some embodiments, the receiver 110 and the processor 115 may be integrated into a device (not shown; such as a single-chip microcomputer Raspberry Pi) provided in the poultry house.
特征处理模块120更包括滤除单元121、分割单元122及提取单元123。滤除单元121用以滤除录音信息I R以产生具有一特定频率范围的一过滤录音信息 I FR。分割单元122用以将过滤录音信息I FR分割成数个声音信息I S。提取单元123用以从数个声音信息I S中提取出数个声音特征C SThe feature processing module 120 further includes a filtering unit 121, a segmentation unit 122 and an extraction unit 123. The filtering unit 121 is used to filter out the recording information I R to generate filtered recording information I FR having a specific frequency range. The dividing unit 122 is used to divide the filtered recording information I FR into several pieces of sound information IS . The extraction unit 123 is used to extract several sound features CS from several pieces of sound information IS .
家禽声纹辨识系统100更包括数据库140及用户介面150。数据库140用以透过网络储存录音信息I R的声音状态S S(或数个声音特征C S的每一者的声音状态S S)。用户介面150用以透过网络供一用户查看数据库140中的声音状态S SThe poultry voiceprint recognition system 100 further includes a database 140 and a user interface 150. The database 140 is used to store the sound state S S of the recording information I R (or the sound state S S of each of the plurality of sound features C S ) through the network. The user interface 150 is used for a user to view the sound status S S in the database 140 through the network.
特征分析模块130更用以在分析数个声音特征C S的每一者以判断数个声音特征C S的每一者的声音状态S S的步骤中经由一人工智能声音模型160来进行判断,人工智能声音模型160根据所述数个声音特征C S产生一训练组T并使用训练组T训练特征分析模块120,训练组T包括记载正常家禽声音状态及异常家禽声音状态的一辨识条件。 The feature analysis module 130 is further configured to make a determination through an artificial intelligence sound model 160 in the step of analyzing each of the plurality of sound characteristics CS to determine the sound state S S of each of the plurality of sound characteristics CS , The artificial intelligence sound model 160 generates a training group T based on the plurality of sound features CS and uses the training group T to train the feature analysis module 120. The training group T includes an identification condition that records normal poultry sound states and abnormal poultry sound states.
人工智能声音模型160经由一支持向量机将数个声音特征C S的每一者判断为家禽声音状态或非家禽声音状态,若属于家禽声音状态,人工智能声音模型经由支持向量机再将属于家禽声音状态的数个声音特征C S的每一者判断为正家禽声音状态或异常家禽声音状态。 The artificial intelligence sound model 160 determines each of the plurality of sound features C S as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state, the artificial intelligence sound model uses a support vector machine to determine whether it belongs to the poultry sound state. Each of the several sound characteristics CS of the sound state is determined to be a positive poultry sound state or an abnormal poultry sound state.
请参照图2,其绘示依照本案一实施例的家禽声纹辨识方法200的示意图,其中至少包含步骤S210及S220,详述如下。Please refer to Figure 2, which illustrates a schematic diagram of a poultry voiceprint identification method 200 according to an embodiment of the present case, which at least includes steps S210 and S220, as detailed below.
在步骤S210中,请同时参照图1,接收器110接收一时间段中的一禽舍的一录音信息I RIn step S210, please also refer to FIG. 1. The receiver 110 receives a recording information IR of a poultry house in a time period.
在步骤S220中,特征分析模块130分析及判断录音信息I R的一声音状态S SIn step S220, the feature analysis module 130 analyzes and determines a sound state S S of the recording information IR .
请参照图3,其绘示依照本案另一实施例的家禽声纹辨识方法300的示意图,其中至少包含步骤S310、S320、S330、S340、S350及S360,详述如下。Please refer to Figure 3, which illustrates a schematic diagram of a poultry voiceprint identification method 300 according to another embodiment of the present invention, which at least includes steps S310, S320, S330, S340, S350 and S360, as detailed below.
在步骤S310中,接收器110接收一时间段中的一禽舍的一录音信息I R。在一些实施例中,接收器110为一全向型麦克风及一指向型麦克风中的至少一者。在一些实施例中,接收器110较佳为指向型麦克风。如图4A至图4B所示,接收器110所接收的录音信息I R可包括波形图或时频图。如图4A所示,此波形图呈现的录音信息I R由于杂讯的干扰,导致家禽声音内容部分的波形不明显。如图4B所示,此时频图呈现的录音信息I R包括频率与家禽声音相近的背景噪音,背景噪音包括但不限于风扇运转声以及禽舍外的车声及雨声。然而,这些背景噪 音会造成后续进行特征值计算时的误差,影响分类结果,因此在进行分析前需要对录音信息I R进行滤波的处理。 In step S310, the receiver 110 receives a recording information IR of a poultry house in a time period. In some embodiments, the receiver 110 is at least one of an omnidirectional microphone and a directional microphone. In some embodiments, receiver 110 is preferably a directional microphone. As shown in FIGS. 4A and 4B , the recording information IR received by the receiver 110 may include a waveform diagram or a time-frequency diagram. As shown in Figure 4A, the waveform of the recording information IR presented in this waveform diagram is not obvious due to the interference of noise. As shown in Figure 4B, the recording information IR presented in this frequency diagram includes background noise with a frequency similar to that of poultry sounds. The background noise includes but is not limited to the sound of fans running, the sound of cars and rain outside the poultry house. However, these background noises will cause errors in subsequent feature value calculations and affect the classification results. Therefore, the recording information IR needs to be filtered before analysis.
在步骤S320中,特征处理模块120将录音信息I R变换成数个声音特征C S。步骤S320还包括步骤S321、步骤S322及步骤S323。特征处理模块120包括滤除单元121、分割单元122及提取单元123。 In step S320, the feature processing module 120 transforms the recording information IR into several sound features CS . Step S320 also includes step S321, step S322 and step S323. The feature processing module 120 includes a filtering unit 121, a segmentation unit 122 and an extraction unit 123.
在步骤S321中,滤除单元121滤除录音信息I R以产生具有一特定频率范围的一过滤录音信息I FR。在一些实施例中,过滤录音信息I FR具有在约500Hz至约5kHz之间的特定频率范围。在一些实施例中,实现所述步骤S321的方法包括但不限于使用滤波器(filter)、谱减法(spectral subtraction)、频谱噪声门(spectral gating)、噪声门(noise gating)、多麦克风降噪(multi-microphone noise reduction)或其他可提升信噪比的方法,其中滤波器(filter)包括但不限于适配器录除器(adapter filter)、有限脉冲滤波器(FIR)或无限脉冲滤波器(IIR),其中有限脉冲滤波器(FIR)或无限脉冲滤波器(IIR)例如为高通滤波器(high-pass filter)、带通滤波器(band-pass filter)、低通滤波器(low-pass filter)或带阻滤波器(band-stop filter)。 In step S321, the filtering unit 121 filters out the recording information I R to generate filtered recording information I FR having a specific frequency range. In some embodiments, the filtered recording information I FR has a specific frequency range between about 500 Hz and about 5 kHz. In some embodiments, methods for implementing step S321 include, but are not limited to, using filters, spectral subtraction, spectral gating, noise gating, and multi-microphone noise reduction. (multi-microphone noise reduction) or other methods that can improve the signal-to-noise ratio, where filters include but are not limited to adapter filters, finite impulse filters (FIR) or infinite impulse filters (IIR) ), where the finite impulse filter (FIR) or the infinite impulse filter (IIR) is, for example, a high-pass filter (high-pass filter), a band-pass filter (band-pass filter), a low-pass filter (low-pass filter) ) or band-stop filter.
在一些实施例中,滤除单元121包括例如一巴特沃斯滤波器,并以带通滤波进行所述步骤S321,n阶巴特沃斯低通滤波器的幅度增益可表示为下式(1),其中H a为传递函数,N为滤波器的阶数,ω为信号的角频率,ω c为振幅下降3dB时的截止频率。随着滤波器的阶数越高,滤波器在阻频带振幅衰减速度越快,滤波效果越好,而低于ω c的频率会以增益通过,高于ω c的频率则会被抑制。 In some embodiments, the filtering unit 121 includes, for example, a Butterworth filter, and performs step S321 with bandpass filtering. The amplitude gain of the n-order Butterworth low-pass filter can be expressed as the following formula (1) , where Ha is the transfer function, N is the order of the filter, ω is the angular frequency of the signal, and ω c is the cut-off frequency when the amplitude drops by 3dB. As the order of the filter is higher, the amplitude of the filter attenuates faster in the stop band, and the filtering effect is better. Frequencies lower than ω c will pass with gain, and frequencies higher than ω c will be suppressed.
Figure PCTCN2022092354-appb-000001
Figure PCTCN2022092354-appb-000001
在所接收的录音信息I R中,除了高、低频的背景噪音干扰之外,还有与家禽声音频率相近的背景噪音干扰,如图5B所示。然而,此情况如果直接使用带通滤波抑制与家禽声音频率相近的背景噪音,将同时导致家禽声音的消除。 In the received recording information IR , in addition to high- and low-frequency background noise interference, there is also background noise interference with a frequency similar to that of poultry sounds, as shown in Figure 5B. However, in this case, if band-pass filtering is directly used to suppress background noise with a frequency similar to that of poultry sounds, it will also lead to the elimination of poultry sounds.
此外,在另一些实施例中,以谱减法进行所述步骤S321,如图6A至图6B所示。谱减法基于一简单的假设,而此假设为语音讯号中的噪声只有加性噪声存在,只要将带噪讯号的频谱减去噪声讯号的频谱,即可得到相对干净的纯语音频谱。时域中的讯号模型可表示为下式(2),其中y(m)为带噪讯号,x(m)为加性噪声,d(m)为纯语音讯号,m为时间。In addition, in other embodiments, the step S321 is performed by spectral subtraction, as shown in Figures 6A to 6B. Spectral subtraction is based on a simple assumption, and this assumption is that the noise in the speech signal only exists additive noise. As long as the spectrum of the noisy signal is subtracted from the spectrum of the noise signal, a relatively clean pure speech spectrum can be obtained. The signal model in the time domain can be expressed as the following formula (2), where y(m) is the noisy signal, x(m) is the additive noise, d(m) is the pure speech signal, and m is time.
y(m)=x(m)+d(m),变换成d(m)=y(m)-x(m),(2)y(m)=x(m)+d(m), transformed into d(m)=y(m)-x(m), (2)
在整段过滤录音信息I FR中,谱减法后的背景噪音(如图6B所示)被抑制的程度远大于带通滤波后(如图5B所示)。因此,相较于带通滤波,谱减法能够有效地消除与家禽声音频率相近的背景噪音(即,能够有效地消除带通带内的背景噪音)。然而,请参照图6B,相较于带通滤波,谱减法无法有效地抑制低频噪音。此外,请参照图7A至图7C,当讯噪比过低时,谱减法在滤波处理后可能导致声音失真(如图7C所示)。 In the entire filtered recording information I FR , the background noise after spectral subtraction (shown in Figure 6B) is suppressed to a much greater extent than after band-pass filtering (shown in Figure 5B). Therefore, compared with band-pass filtering, spectral subtraction can effectively eliminate background noise that is similar to the frequency of poultry sounds (that is, it can effectively eliminate background noise within the band-pass band). However, please refer to Figure 6B. Compared with band-pass filtering, spectral subtraction cannot effectively suppress low-frequency noise. In addition, please refer to Figures 7A to 7C. When the signal-to-noise ratio is too low, spectral subtraction may cause sound distortion after filtering (as shown in Figure 7C).
此外,在又一些实施例中,以带通滤波及谱减法进行所述步骤S321。请参照图8A至图8B,透过带通滤波将大部分的高、低频噪音抑制后,再利用谱减法抑制通频带内的背景噪音,也减少了谱减法可能导致的家禽声音失真的状况发生(如图8B所示)。In addition, in some embodiments, the step S321 is performed using bandpass filtering and spectral subtraction. Please refer to Figure 8A to Figure 8B. After most of the high- and low-frequency noise is suppressed through band-pass filtering, spectral subtraction is then used to suppress the background noise in the pass-band. This also reduces the distortion of poultry sounds that may be caused by spectral subtraction. (As shown in Figure 8B).
在步骤S322中,分割单元122将过滤录音信息I FR分割成数个声音信息I S。举例而言,录音信息I R及过滤录音信息I FR为具有连续录音5分钟的时长的档案,包括许多正常家禽声音片段、异常家禽声音片段、非家禽声音片段、无声音片段等。此导致难以界定此档案的归类,造成后续提取此档案中的声学特征失去了单一性,因此对过滤录音信息I FR进行分割,减少单一信息档案中所包含的家禽声音数量具有其必要性。在一些实施例中,实现所述步骤S322的方法包括但不限于语音活动检测(Voice Activity Detection,VAD)、自相关函数(autocorrelation function,ACF)或其他声音特征辨识方法。 In step S322, the dividing unit 122 divides the filtered recording information I FR into several pieces of sound information IS . For example, the recording information I R and the filtered recording information I FR are files with continuous recording duration of 5 minutes, including many normal poultry sound clips, abnormal poultry sound clips, non-poultry sound clips, no sound clips, etc. This makes it difficult to define the classification of this file, causing the subsequent extraction of acoustic features in this file to lose its unity. Therefore, it is necessary to segment the filtered recording information I FR to reduce the number of poultry sounds contained in a single information file. In some embodiments, methods for implementing step S322 include, but are not limited to, voice activity detection (VAD), autocorrelation function (autocorrelation function, ACF), or other voice feature recognition methods.
在一些实施例中,使用语音活动检测(Voice Activity Detection,VAD)进行所述步骤S322。语音活动检测以长度10毫秒一帧为单位,并以六个子带能量作为特征,分别为80~250Hz、250~500Hz、500~1000Hz、1~2kHz、2~3kHz、3~4kHz,对整段过滤录音信息I FR进行检测,计算出每帧分别为语音及噪音的概率,而其最终判断具有语音活动的标准为任一子带或六个子带的语音总似然比大于0.9则判定为具有语音活动。此外,当检测到具有语音活动时程式将开始记录,直到检测到语音似然比小于0.9,即不具有语音活动时程式将结束纪录,并将所述纪录片段截取出来成为一个时长较短的新音讯档案(例如时常多为2秒内的档案)。请参照图9,这些截取出来的声音信息I S中所包含的家禽声音数量相对较少,相较 于原来5分钟的录音信息I R、过滤录音信息I FR更具有单一性,并且能够针对家禽声音进行更准确的分类。 In some embodiments, step S322 is performed using Voice Activity Detection (VAD). Voice activity detection is based on a frame length of 10 milliseconds, and is characterized by six sub-band energies, namely 80~250Hz, 250~500Hz, 500~1000Hz, 1~2kHz, 2~3kHz, 3~4kHz, for the entire segment Filter the recording information I FR for detection, and calculate the probability that each frame is speech and noise respectively. The final criterion for judging that there is speech activity is that the total likelihood ratio of speech in any subband or six subbands is greater than 0.9, then it is judged as having speech activity. Voice activities. In addition, when voice activity is detected, the program will start recording until the voice likelihood ratio is detected to be less than 0.9, that is, when there is no voice activity, the program will end the recording and intercept the documentary segment into a new one with a shorter duration. Audio files (for example, files usually within 2 seconds). Please refer to Figure 9. The number of poultry sounds contained in these intercepted sound information IS is relatively small. Compared with the original 5-minute recording information I R and the filtered recording information I FR , it is more single and can target poultry. Sounds are classified more accurately.
此外,请参照下表1,显示出相较于仅使用带通滤波、或仅使用谱减法,带通滤波后进行谱减法具有较佳的声音截取结果。In addition, please refer to Table 1 below, which shows that compared to using only band-pass filtering or only using spectral subtraction, spectral subtraction after band-pass filtering has better sound interception results.
表1:Table 1:
Figure PCTCN2022092354-appb-000002
Figure PCTCN2022092354-appb-000002
在步骤S323中,提取单元123从数个声音信息I S中提取出数个声音特征C S。在分析一段音讯时,通常以短时距分析为主,因为音讯的变化量大且本研究的录音环境在商业禽舍中,音讯内容更为复杂,因此短时间的数据分析是相对较为稳定的,而短时距分析通常会将一段音讯内容以长度30毫秒左右的音框(Frame)为单位分别进行特征值的计算。步骤S323还可包括特征取得方法、特征变换方法或特征提取方法。在一些实施例中,实现所述步骤S323中的特征取得方法包括但不限于信号变换(例如小波分析(Wavelet Analysis)、功率谱密度(power spectral density,PSD))、openSMILE人类情感特征集、过零率(zero-crossing rate,ZCR)、短时距傅立叶变换、快速傅立叶变换(Fast Fourier Transform,FFT)(例如能量强度、基频)、梅尔频率倒谱系数(Mel-frequency cepstral coefficient,MFCC)、倒谱峰突出(Cepstral peak prominence,CPP)及Welchz方法。在一些实施例中,实现所述步骤S323中的特征变换方法包括但不限于标准化(Standardization)、正规化(Normalization)、二值化(Binarization)或其他数值变换、缩放、函数变换的方式。在一些实施例中,实现所述步骤S323中的特征提取方法包括但不限于主成分分析(Principal Component Analysis,PCA)、线性判别分析(Linear Discrimination Analysis,LDA)、局部线性嵌入法(Local Linear Embedding,LLE)、拉普拉斯特征映射(Laplacian Eigenmap,LE)、随机邻域嵌入法(Stochastic Neighbor Embedding, SNE)、T-随机邻域嵌入法(T-Stochastic Neighbor Embedding,T-SNE)、核主成分分析(KPCA)、迁移成分分析(TCA)或其他特征降维、特征提取方式。 In step S323, the extraction unit 123 extracts several sound features CS from several pieces of sound information IS . When analyzing a piece of audio, short-term analysis is usually the main method. Because the audio changes greatly and the recording environment of this study is in a commercial poultry house, the audio content is more complex, so short-term data analysis is relatively stable. , and short-term analysis usually calculates the feature values of a piece of audio content in units of sound frames (Frame) with a length of about 30 milliseconds. Step S323 may also include a feature acquisition method, a feature transformation method or a feature extraction method. In some embodiments, the feature acquisition method in step S323 includes but is not limited to signal transformation (such as wavelet analysis (Wavelet Analysis), power spectral density (PSD)), openSMILE human emotion feature set, process Zero-crossing rate (ZCR), short-time Fourier transform, Fast Fourier Transform (FFT) (such as energy intensity, fundamental frequency), Mel-frequency cepstral coefficient (MFCC) ), Cepstral peak prominence (CPP) and Welchz method. In some embodiments, the feature transformation method in step S323 includes but is not limited to standardization, normalization, binarization or other numerical transformation, scaling, and function transformation methods. In some embodiments, the feature extraction methods in step S323 include but are not limited to principal component analysis (PCA), linear discriminant analysis (LDA), local linear embedding (Local Linear Embedding) , LLE), Laplacian Eigenmap (LE), Stochastic Neighbor Embedding (SNE), T-Stochastic Neighbor Embedding (T-SNE), kernel Principal component analysis (KPCA), transfer component analysis (TCA) or other feature dimensionality reduction and feature extraction methods.
在一些实施例中,使用openSMILE来进行所述步骤S323。openSMILE是一种适用于信号处理及音频声学特征提取的开源工具包。本实施例使用其中的“The INTERSPEECH 2009 Emotion Challenge feature set”特征集,而每笔音讯档案皆会进行共384个声学特征的提取,其中包含均方根讯号帧能量(root-mean-square signal frame energy)、梅尔倒频谱系数1至12(Mel-Frequency cepstral coefficients 1-12)、时间信号过零率(zero-crossing rate of time signal)、根据ACF计算的发声概率(the voicing probability computed from the ACF)、由倒谱计算出的基本频率(the fundamental frequency computed from the Cepstrum)等,共16个低阶描述符(low-level Descriptors,LLDs)。In some embodiments, openSMILE is used to perform step S323. openSMILE is an open source toolkit suitable for signal processing and audio acoustic feature extraction. This embodiment uses the "The INTERSPEECH 2009 Emotion Challenge feature set" feature set, and a total of 384 acoustic features are extracted for each audio file, including the root-mean-square signal frame energy (root-mean-square signal frame). energy), Mel-Frequency cepstral coefficients 1-12, zero-crossing rate of time signal, the voicing probability calculated from the ACF ACF), the fundamental frequency computed from the Cepstrum, etc., a total of 16 low-level descriptors (LLDs).
梅尔倒频谱系数(Mel-Frequency cepstral coefficients,MFCCs)是一组用来建立梅尔倒频谱的关键系数,梅尔倒频谱则是一个用来代表短期音讯的频谱,原理是基于用非线性的梅尔刻度表示的对数频谱及其线性余弦变换上。其最大特色在于,梅尔倒频谱上的频带是均匀分布于梅尔刻度上,这种表示方法及人类的非线性听觉系统较为接近,也由于此声学特征考虑到人耳对不同频率的感受程度,因此特别适合用于语音辨识,其中梅尔刻度(m)与实际频率(f)之间的变换可以表示为下式(3),其中f为实际频率值。其参考点定义1000Hz为1000mel。Mel-Frequency cepstral coefficients (MFCCs) are a set of key coefficients used to establish Mel-Frequency cepstral coefficients. Mel-Frequency cepstral coefficients are a spectrum used to represent short-term audio. The principle is based on the use of nonlinear Mel scale representation of the logarithmic spectrum and its linear cosine transform. Its biggest feature is that the frequency bands on the Mel cepstrum are evenly distributed on the Mel scale. This representation method is closer to the human nonlinear auditory system. This acoustic feature also takes into account the human ear's perception of different frequencies. , so it is particularly suitable for speech recognition, where the transformation between the Mel scale (m) and the actual frequency (f) can be expressed as the following formula (3), where f is the actual frequency value. Its reference point defines 1000Hz as 1000mel.
Figure PCTCN2022092354-appb-000003
Figure PCTCN2022092354-appb-000003
在步骤S330中,人工智能声音模型160根据数个声音特征C S产生一训练组T。人工智能声音模型160使用训练组T训练特征分析模块130。训练组T包括记载正常家禽声音状态、异常家禽声音状态及非家禽声音状态的一辨识条件。在一些实施例中,实现所述步骤S330的方法包括但不限于监督式学习(supervised learning)、非监督式学习(un-supervised learning)、半监督式学习(semi-supervised learning)及强化式学习(reinforcement learning)。监督式学习包括但不限于分类器(classification)及回归(regression),其中分类器包括但不限于随机森林(random forest)、K-近邻法(K Nearest Neighbor,k-NN)、支持向量机(support vector machine,SVM)、人工神经网络(artificial neural network,ANN)、支持向量域描述法(support  vector domain description,SVDD)、稀疏表示分类器(sparse representation classifier,SRC)。非监督式学习包括但不限于聚类分析(clustering)及降维(dimensionality reduction)。 In step S330, the artificial intelligence sound model 160 generates a training group T according to several sound features CS . The artificial intelligence sound model 160 uses the training group T to train the feature analysis module 130. The training group T includes an identification condition that records normal poultry sound states, abnormal poultry sound states, and non-poultry sound states. In some embodiments, methods for implementing step S330 include, but are not limited to, supervised learning, unsupervised learning, semi-supervised learning, and reinforcement learning. (reinforcement learning). Supervised learning includes but is not limited to classifiers (classification) and regression (regression). Classifiers include but are not limited to random forest (random forest), K-Nearest Neighbor (k-NN), support vector machine ( support vector machine (SVM), artificial neural network (ANN), support vector domain description (SVDD), sparse representation classifier (SRC). Unsupervised learning includes but is not limited to clustering and dimensionality reduction.
举例而言,请参照图10A至图10B,显示出一实验例所收集的正常家禽声音及异常家禽声音。如图10A所示,显示出20日龄的土鸡正常声音数据,经从业人员确认,土鸡正常声音频率在约2.5kHz至约3.5kHz之间,单一个发声的持续时间约为0.1秒。如图10B所示,显示出20日龄的土鸡异常声音数据,所述异常声音的频率在约1kHz至约1.5kHz之间,而此单一个发声的持续时间约在0.67秒至0.7秒之间,经从业人员确认,此异常声音为啰音症状的表现,其中啰音的特征之一为由于气管中黏液过多而堵塞所导致的拉长音状况。For example, please refer to FIGS. 10A and 10B , which show normal poultry sounds and abnormal poultry sounds collected in an experimental example. As shown in Figure 10A, the normal sound data of a 20-day-old native chicken is displayed. It has been confirmed by practitioners that the normal sound frequency of a native chicken is between about 2.5kHz and about 3.5kHz, and the duration of a single sound is about 0.1 seconds. As shown in Figure 10B, abnormal sound data of a 20-day-old native chicken is displayed. The frequency of the abnormal sound is between about 1 kHz and about 1.5 kHz, and the duration of this single sound is between about 0.67 seconds and 0.7 seconds. During the period, practitioners confirmed that this abnormal sound was a symptom of rales. One of the characteristics of rales is a prolonged sound caused by excessive mucus in the trachea and obstruction.
此外,请参照某图11A至图11B,显示出另一实验例所收集的正常家禽声音及异常家禽声音。如图11A所示,显示出19日龄的土鸡正常声音数据,经从业人员确认,土鸡正常声音频率在约2.5kHz至约4kHz之间,单一个发声的持续时间约为0.15秒。如图11B所示,显示出20日龄的土鸡异常声音数据,所述异常声音的频率在约1kHz至约1.5kHz之间,而此单一个发声的持续时间约0.7秒,经从业人员确认,此异常声音为啰音症状的表现,其中啰音的特征之一为由于气管中黏液过多而堵塞所导致的拉长音状况。In addition, please refer to Figures 11A to 11B, which show normal poultry sounds and abnormal poultry sounds collected in another experimental example. As shown in Figure 11A, the normal sound data of a 19-day-old native chicken is displayed. It has been confirmed by practitioners that the normal sound frequency of a native chicken is between about 2.5kHz and about 4kHz, and the duration of a single sound is about 0.15 seconds. As shown in Figure 11B, the abnormal sound data of a 20-day-old native chicken is displayed. The frequency of the abnormal sound is between about 1kHz and about 1.5kHz, and the duration of this single sound is about 0.7 seconds, which was confirmed by practitioners. , this abnormal sound is a symptom of rales. One of the characteristics of rales is a prolonged sound caused by excessive mucus in the trachea and obstruction.
如图10A至图11B所示,正常家禽声音及异常家禽声音在频率上差距约2kHz,且在声音的持续时间上具有约0.5秒至0.6秒的差距。因此,举例而言,可根据声音的频率差距或持续时间差距来设定在步骤S330中的所述辨识条件。As shown in Figures 10A to 11B, the frequency difference between normal poultry sounds and abnormal poultry sounds is about 2 kHz, and there is a difference between sound durations of about 0.5 seconds to 0.6 seconds. Therefore, for example, the identification condition in step S330 may be set according to the frequency difference or duration difference of the sounds.
请参照图12,图12绘示依照本案一实施例的支持向量机的分类示意图1200。人工智能声音模型160经由一支持向量机将数个声音特征C S的每一者判断为家禽声音状态或非家禽声音状态,若属于家禽声音状态,人工智能声音模型160经由支持向量机再将属于家禽声音状态的数个声音特征C S的每一者判断为正家禽声音状态或异常家禽声音状态。 Please refer to FIG. 12 , which illustrates a classification diagram 1200 of a support vector machine according to an embodiment of the present application. The artificial intelligence sound model 160 determines each of the several sound features C S as a poultry sound state or a non-poultry sound state through a support vector machine. If it belongs to the poultry sound state, the artificial intelligence sound model 160 uses a support vector machine to determine whether it belongs to the poultry sound state. Each of the several sound characteristics CS of the poultry sound state is determined to be a positive poultry sound state or an abnormal poultry sound state.
在一些实施例中,人工智能声音模型160共使用了150笔正常家禽声音数据、150笔异常家禽声音数据及150笔非家禽声音数据进行模型的训练,此三类训练集数据如表2所示。训练完成的人工智能声音模型160透过混淆矩阵及10-fold 交叉验证,得到在辨识这三类的数据有84.2%的验证准确率,人工智能声音模型160验证结果如表3所示。In some embodiments, the artificial intelligence sound model 160 uses a total of 150 pieces of normal poultry sound data, 150 pieces of abnormal poultry sound data, and 150 pieces of non-poultry sound data to train the model. The three types of training set data are shown in Table 2. . Through confusion matrix and 10-fold cross-validation, the trained artificial intelligence sound model 160 has a verification accuracy of 84.2% in identifying these three types of data. The verification results of the artificial intelligence sound model 160 are shown in Table 3.
表2:Table 2:
Figure PCTCN2022092354-appb-000004
Figure PCTCN2022092354-appb-000004
表3:table 3:
Figure PCTCN2022092354-appb-000005
Figure PCTCN2022092354-appb-000005
在步骤S340中,特征分析模块130分析及判断数个声音特征C S的每一者的一声音状态S S。在一些实施例中,在分析数个声音特征C S的每一者以判断数个声音特征C S的每一者的声音状态S S的步骤中是经由一人工智能声音模型160(如步骤S330所示)所判断。 In step S340, the feature analysis module 130 analyzes and determines a sound state S S of each of the plurality of sound features C S . In some embodiments, the step of analyzing each of the plurality of sound characteristics CS to determine the sound state S S of each of the plurality of sound characteristics CS is through an artificial intelligence sound model 160 (eg, step S330 shown).
在步骤S350中,透过一网络将录音信息I R的声音状态S S(或数个声音特征C S的每一者的声音状态S S)储存于一数据库140。在一些实施例中,数据库140为一云端数据库,用以储存声音状态S S的历史信息。 In step S350, the sound state S S of the recording information I R (or the sound state S S of each of the plurality of sound features C S ) is stored in a database 140 through a network. In some embodiments, the database 140 is a cloud database for storing historical information of the sound state S S.
在步骤S360中,透过网络供一用户查看数据库140中的声音状态S SIn step S360, a user is allowed to view the sound status S S in the database 140 through the network.
请参照图13至图16,图13绘示依照本案一实施例的各日龄禽舍内温度及湿度信息,图14绘示依照本案一实施例的各日龄当天声音数据数量预测结果,图15绘示依照本案一实施例的各日龄当天异常家禽声音数量预测结果,图16绘示依照本案一实施例的各类别所占当天预测结果比例,其中图14至图16所示的人为观察异状日D表示为从业人员观察到异常家禽声音(例如家禽有啰音的状况)的日期,且人为观察异状日D也适用于后续图式的叙述,合先叙明。如图14及图16所示,虽然此从业人员提到于20日龄时发现有些家禽有啰音的状况,由于正常家禽声音的数量远大于异常家禽声音的数量,难以观察到异常家禽声音的数量变化。如图15所示,独立整理出异常家禽声音的数量变化,可知异常家禽声音的比例在18日龄开始明显增加,并在23日龄时比例开始下降。Please refer to Figures 13 to 16. Figure 13 illustrates the temperature and humidity information in the poultry house of each age according to an embodiment of the present case. Figure 14 illustrates the prediction results of the number of sound data on the same day for each age according to an embodiment of the present case. Figure 15 shows the prediction results of the number of abnormal poultry sounds of each age on the day according to an embodiment of this case. Figure 16 shows the proportion of each category in the prediction results of the day according to an embodiment of this case. The human observations shown in Figures 14 to 16 are The abnormality day D is the date when the practitioner observed abnormal poultry sounds (such as rales in the poultry), and the artificial observation of the abnormality day D is also applicable to the description of the subsequent figures, so it will be explained first. As shown in Figures 14 and 16, although this practitioner mentioned that some poultry were found to have rales at 20 days of age, since the number of normal poultry sounds is much greater than the number of abnormal poultry sounds, it is difficult to observe the abnormal poultry sounds. Quantity changes. As shown in Figure 15, the changes in the number of abnormal poultry sounds were independently sorted out. It can be seen that the proportion of abnormal poultry sounds began to increase significantly at the age of 18 days, and the proportion began to decrease at the age of 23 days.
请参照图17A至图17C,图17A绘示依照本案一实施例的各日龄早上6点至下午2点的声音数据数量预测结果,图17B绘示依照本案一实施例的各日龄下午2点至晚上10点的声音数据数量预测结果,图17C绘示依照本案一实施例的晚上10点至隔日早上6点的声音数据数量预测结果。Please refer to Figures 17A to 17C. Figure 17A shows the prediction results of the number of sound data from 6 a.m. to 2 p.m. for each day age according to an embodiment of the present case. Figure 17B shows the prediction results for the sound data quantity from 2 p.m. for each day age according to an embodiment of the present case. The prediction results of the number of sound data from 10 p.m. to 10 p.m., FIG. 17C shows the prediction results of the number of sound data from 10 p.m. to 6 a.m. the next day according to an embodiment of the present case.
进一步而言,独立整理出异常家禽声音的数量变化,请参照图18A至图18C,图18A绘示依照本案一实施例的各日龄早上6点至下午2点的异常家禽声音数据数量预测结果,图18B绘示依照本案一实施例的各日龄下午2点至晚上10点的异常家禽声音数据数量预测结果,图18C绘示依照本案一实施例的各日龄晚上10点至隔日早上6点的异常家禽声音数据数量预测结果。于此实施例中,经从业人员确认后,在这2周期间的数据下,平均每日搜集约12,000笔声音数据中,实际为异常家禽声音且被辨识为异常家禽声音的准确率为99.3%。因此,本案所提出的系统及方法对于异常家禽声音的预测结果可提供给一用户(例如从业人员)作为评估家禽健康状况的依据。此外,相对于从业人员于20日龄观察到异常家禽声音,本案所提出的系统及方法能够于18日龄观察到异常家禽声音,如此可帮助用户(例如从业人员)更快采行措施。Furthermore, to independently sort out the changes in the number of abnormal poultry sounds, please refer to Figures 18A to 18C. Figure 18A shows the prediction results of the number of abnormal poultry sound data from 6 am to 2 pm for each age according to an embodiment of this case. , Figure 18B shows the prediction results of the number of abnormal poultry sound data from 2 pm to 10 pm of each age according to an embodiment of this case, and Figure 18C shows the prediction results of the number of abnormal poultry sound data from 10 pm to 6 am of each day of age according to an embodiment of this case. Prediction results of the number of abnormal poultry sound data points. In this example, after confirmation by practitioners, under the data during this 2-week period, an average of about 12,000 pieces of sound data were collected every day, and the accuracy rate of being actually abnormal poultry sounds and being identified as abnormal poultry sounds was 99.3%. . Therefore, the prediction results of abnormal poultry sounds by the system and method proposed in this case can be provided to a user (such as a practitioner) as a basis for assessing the health status of poultry. In addition, compared with practitioners observing abnormal poultry sounds at 20 days of age, the system and method proposed in this case can observe abnormal poultry sounds at 18 days of age, which can help users (such as practitioners) take measures faster.
根据一些实施例,本案揭示一种家禽声纹辨识系统及方法,可接收在一时间段中的一禽舍的一录音信息,将录音信息转变成图像(例如波形图或时频图等平面图像),根据图像的数个图像指标(例如频率差距或持续时间差距等辨识条件)分析录音信息以判断录音信息的声音状态,声音状态包括正常家禽声音状态及/ 或异常家禽声音状态。根据一些实施例,上述根据图像的数个图像指标分析录音信息以判断录音信息的声音状态的步骤亦可结合人工智能声音模型实现。According to some embodiments, this case discloses a poultry voiceprint recognition system and method that can receive a recording information of a poultry house in a time period and convert the recording information into an image (such as a waveform diagram or a time-frequency diagram and other planar images). ), analyze the recording information based on several image indicators of the image (such as frequency gap or duration gap and other identification conditions) to determine the sound status of the recording information. The sound status includes normal poultry sound status and/or abnormal poultry sound status. According to some embodiments, the above-mentioned step of analyzing the recording information based on several image indicators of the image to determine the sound state of the recording information can also be implemented in conjunction with an artificial intelligence sound model.
基于本案实施例的家禽声纹辨识系统及方法,即使在禽舍接收许多声音交杂的录音信息下,透过对计算机软/硬件的改良并结合人工智能声音模型,能准确地及快速地辨识出异常家禽声音的产生。The poultry voiceprint identification system and method based on the embodiment of this case can accurately and quickly identify through the improvement of computer software/hardware and the combination of artificial intelligence sound models, even when the poultry house receives recording information with many mixed sounds. The production of abnormal poultry sounds.
综上所述,虽然本案已以实施例揭露如上,然其并非用以限定本案。本领域的技术人员,在不脱离本案的精神及范围内,当可作各种的更动与润饰。因此,本案的保护范围当视后附的权利要求所界定者为准。In summary, although the present case has been disclosed as above using embodiments, they are not used to limit the present case. Those skilled in the art can make various changes and modifications without departing from the spirit and scope of the present invention. Therefore, the scope of protection in this case shall be determined by the appended claims.

Claims (14)

  1. 一种家禽声纹辨识方法,其特征在于,所述家禽声纹辨识方法包括以下步骤:A poultry voiceprint identification method, characterized in that the poultry voiceprint identification method includes the following steps:
    接收在一时间段中的一禽舍的一录音信息;以及receiving a recorded message for a poultry house during a time period; and
    分析所述录音信息以判断所述录音信息的一声音状态,所述声音状态包括一正常家禽声音状态或一异常家禽声音状态。The recorded information is analyzed to determine a sound state of the recorded information, where the sound state includes a normal poultry sound state or an abnormal poultry sound state.
  2. 如权利要求1所述的家禽声纹辨识方法,其特征在于,所述家禽声纹辨识方法更包括以下步骤:The poultry voiceprint identification method according to claim 1, wherein the poultry voiceprint identification method further includes the following steps:
    将所述录音信息变换成数个声音特征,其中在将所述录音信息变换成数个声音特征的步骤包括以下步骤:Converting the recorded information into several sound features, wherein the step of converting the recorded information into several sound features includes the following steps:
    滤除所述录音信息以产生具有一特定频率范围的一过滤录音信息;filtering the recorded information to generate filtered recorded information having a specific frequency range;
    将所述过滤录音信息分割成数个声音信息;以及Divide the filtered recording information into several pieces of sound information; and
    从所述数个声音信息中提取出所述数个声音特征,extract the plurality of sound features from the plurality of sound information,
    其中在分析所述录音信息以判断所述录音信息的所述声音状态的步骤为分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态。The step of analyzing the recording information to determine the sound state of the recording information is analyzing each of the plurality of sound features to determine the sound state of each of the plurality of sound characteristics.
  3. 如权利要求2所述的家禽声纹辨识方法,其特征在于,在滤除所述录音信息以产生具有所述特定频率范围的所述过滤录音信息的步骤是以带通滤波及谱减法实现。The poultry voiceprint identification method according to claim 2, wherein the step of filtering out the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
  4. 如权利要求2所述的家禽声纹辨识系统,其特征在于,在从所述数个声音信息中提取出所述数个声音特征的步骤是以openSMILE、Wavelet或短时距傅立叶变换实现。The poultry voiceprint identification system of claim 2, wherein the step of extracting the plurality of sound features from the plurality of sound information is implemented using openSMILE, Wavelet or short-time Fourier transform.
  5. 如权利要求1所述的家禽声纹辨识方法,其特征在于,所述家禽声纹辨识方法更包括以下步骤:The poultry voiceprint identification method according to claim 1, wherein the poultry voiceprint identification method further includes the following steps:
    透过一网络将所述录音信息的所述声音状态储存于一数据库;以及Storing the sound status of the recording information in a database through a network; and
    透过所述网络供一用户查看所述数据库中的所述声音状态。A user is provided to view the sound status in the database through the network.
  6. 如权利要求2所述的家禽声纹辨识方法,其特征在于,在分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态的步骤中是经由一人工智能声音模型所判断,所述人工智能声音模型根据所述数个声音特征产生一训练组,所述训练组包括记载所述正常家禽声音状态及所述异常家禽声音状态的一辨识条件。The poultry voiceprint identification method according to claim 2, wherein in the step of analyzing each of the plurality of sound characteristics to determine the sound state of each of the plurality of sound characteristics: Judging by an artificial intelligence sound model, the artificial intelligence sound model generates a training group based on the plurality of sound characteristics. The training group includes an identification condition that records the normal poultry sound state and the abnormal poultry sound state. .
  7. 如权利要求6所述的家禽声纹辨识方法,其特征在于,所述人工智能声音模型经由一支持向量机将所述数个声音特征的每一者判断为一家禽声音状态或一非家禽声音状态,若属于所述家禽声音状态,所述人工智能声音模型经由所述支持向量机再将属于所述家禽声音状态的所述数个声音特征的每一者判断为所述正家禽声音状态或所述异常家禽声音状态。The poultry voiceprint identification method of claim 6, wherein the artificial intelligence sound model determines each of the plurality of sound features as a poultry sound state or a non-poultry sound through a support vector machine. state, if it belongs to the poultry sound state, the artificial intelligence sound model determines each of the several sound features belonging to the poultry sound state as the positive poultry sound state or The abnormal poultry sound condition.
  8. 一种家禽声纹辨识系统,其特征在于,所述家禽声纹辨识系统安装于连接至一网络的一伺服器,所述家禽声纹辨识系统包括:A poultry voiceprint identification system, characterized in that the poultry voiceprint identification system is installed on a server connected to a network, and the poultry voiceprint identification system includes:
    一接收器,设置于一禽舍中,用以接收一时间段中的所述禽舍的一录音信息;以及A receiver, disposed in a poultry house, for receiving a recorded message of the poultry house in a time period; and
    一特征分析模块,用以分析所述录音信息以判断所述录音信息的一声音状态,所述声音状态包括一正常家禽声音状态或一异常家禽声音状态。A feature analysis module is used to analyze the recording information to determine a sound state of the recording information. The sound state includes a normal poultry sound state or an abnormal poultry sound state.
  9. 如权利要求8所述的家禽声纹辨识系统,其特征在于,所述家禽声纹辨识系统更包括:The poultry voiceprint identification system according to claim 8, wherein the poultry voiceprint identification system further includes:
    一特征处理模块,用以将所述录音信息变换成数个声音特征,其中所述特征处理模块更包括:A feature processing module for converting the recorded information into several sound features, wherein the feature processing module further includes:
    一滤除单元,用以滤除所述录音信息以产生具有一特定频率范围的一过滤录音信息;a filtering unit for filtering the recording information to generate filtered recording information having a specific frequency range;
    一分割单元,用以将所述过滤录音信息分割成数个声音信息;以及A dividing unit used to divide the filtered recording information into several pieces of sound information; and
    一提取单元,用以从所述数个声音信息中提取出所述数个声音特征,an extraction unit, used to extract the plurality of sound features from the plurality of sound information,
    其中所述特征分析模块在分析所述录音信息以判断所述录音信息的所述声音状态的步骤用以分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态。In the step of analyzing the recording information to determine the sound state of the recording information, the feature analysis module is used to analyze each of the plurality of sound characteristics to determine each of the plurality of sound characteristics. of said sound state.
  10. 如权利要求9所述的家禽声纹辨识系统,其特征在于,所述滤除单元更用以:The poultry voiceprint identification system according to claim 9, wherein the filtering unit is further used for:
    在滤除所述录音信息以产生具有所述特定频率范围的所述过滤录音信息的步骤是以带通滤波及谱减法实现。The step of filtering out the recording information to generate the filtered recording information having the specific frequency range is implemented by bandpass filtering and spectral subtraction.
  11. 如权利要求9所述的家禽声纹辨识系统,其特征在于,所述提取单元更用以:The poultry voiceprint identification system according to claim 9, wherein the extraction unit is further used to:
    在从所述数个声音信息中提取出所述数个声音特征的步骤是以openSMILE、Wavelet或短时距傅立叶变换实现。The step of extracting the plurality of sound features from the plurality of sound information is implemented by openSMILE, Wavelet or short-time Fourier transform.
  12. 如权利要求6所述的家禽声纹辨识系统,其特征在于,所述家禽声纹辨识系 统更包括:The poultry voiceprint identification system according to claim 6, wherein the poultry voiceprint identification system further includes:
    一数据库,用以透过所述网络储存所述录音信息的所述声音状态;以及a database for storing the sound status of the recording information through the network; and
    一用户介面,用以透过所述网络供一用户查看所述数据库中的所述声音状态。A user interface for a user to view the sound status in the database through the network.
  13. 如权利要求9所述的家禽声纹辨识系统,其特征在于,所述特征分析模块更用以:The poultry voiceprint identification system according to claim 9, wherein the feature analysis module is further used to:
    在分析所述数个声音特征的每一者以判断所述数个声音特征的每一者的所述声音状态的步骤中是经由一人工智能声音模型所判断,所述人工智能声音模型根据所述数个声音特征产生一训练组并使用所述训练组训练所述特征分析模块,所述训练组包括记载所述正常家禽声音状态及所述异常家禽声音状态的一辨识条件。In the step of analyzing each of the plurality of sound characteristics to determine the sound state of each of the plurality of sound characteristics, the determination is made through an artificial intelligence sound model, and the artificial intelligence sound model is based on the The several sound features generate a training group and use the training group to train the feature analysis module. The training group includes an identification condition that records the normal poultry sound state and the abnormal poultry sound state.
  14. 如权利要求13所述的家禽声纹辨识系统,其特征在于,所述人工智能声音模型经由一支持向量机将所述数个声音特征的每一者判断为一家禽声音状态或一非家禽声音状态,若属于所述家禽声音状态,所述人工智能声音模型经由所述支持向量机再将属于所述家禽声音状态的所述数个声音特征的每一者判断为所述正家禽声音状态或所述异常家禽声音状态。The poultry voiceprint identification system of claim 13, wherein the artificial intelligence sound model determines each of the plurality of sound features as a poultry sound state or a non-poultry sound through a support vector machine. state, if it belongs to the poultry sound state, the artificial intelligence sound model determines each of the several sound features belonging to the poultry sound state as the positive poultry sound state or The abnormal poultry sound condition.
PCT/CN2022/092354 2022-05-12 2022-05-12 Poultry voiceprint recognition method and system WO2023216172A1 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/092354 WO2023216172A1 (en) 2022-05-12 2022-05-12 Poultry voiceprint recognition method and system

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
PCT/CN2022/092354 WO2023216172A1 (en) 2022-05-12 2022-05-12 Poultry voiceprint recognition method and system

Publications (1)

Publication Number Publication Date
WO2023216172A1 true WO2023216172A1 (en) 2023-11-16

Family

ID=88729369

Family Applications (1)

Application Number Title Priority Date Filing Date
PCT/CN2022/092354 WO2023216172A1 (en) 2022-05-12 2022-05-12 Poultry voiceprint recognition method and system

Country Status (1)

Country Link
WO (1) WO2023216172A1 (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8915215B1 (en) * 2012-06-21 2014-12-23 Scott A. Helgeson Method and apparatus for monitoring poultry in barns
CN106691451A (en) * 2016-07-11 2017-05-24 山东省农业科学院家禽研究所(山东省无特定病原鸡研究中心) Sick poultry and location system and location method for floor-rearing birdhouse
CN109599120A (en) * 2018-12-25 2019-04-09 哈尔滨工程大学 One kind being based on large-scale farming field factory mammal abnormal sound monitoring method
CN113219964A (en) * 2021-03-30 2021-08-06 广州朗国电子科技有限公司 Poultry house environment inspection and regulation method, equipment and medium
CN113539294A (en) * 2021-05-31 2021-10-22 河北工业大学 Method for collecting and identifying sound of abnormal state of live pig

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8915215B1 (en) * 2012-06-21 2014-12-23 Scott A. Helgeson Method and apparatus for monitoring poultry in barns
CN106691451A (en) * 2016-07-11 2017-05-24 山东省农业科学院家禽研究所(山东省无特定病原鸡研究中心) Sick poultry and location system and location method for floor-rearing birdhouse
CN109599120A (en) * 2018-12-25 2019-04-09 哈尔滨工程大学 One kind being based on large-scale farming field factory mammal abnormal sound monitoring method
CN113219964A (en) * 2021-03-30 2021-08-06 广州朗国电子科技有限公司 Poultry house environment inspection and regulation method, equipment and medium
CN113539294A (en) * 2021-05-31 2021-10-22 河北工业大学 Method for collecting and identifying sound of abnormal state of live pig

Similar Documents

Publication Publication Date Title
CN110570880B (en) Snore signal identification method
CN108701469B (en) Cough sound recognition method, device, and storage medium
CN106653032B (en) Based on the animal sounds detection method of multiband Energy distribution under low signal-to-noise ratio environment
CN110970042B (en) Pulmonary ralated artificial intelligence real-time classification method, system and device of electronic stethoscope and readable storage medium
Alonso et al. Automatic anuran identification using noise removal and audio activity detection
Chee et al. MFCC based recognition of repetitions and prolongations in stuttered speech using k-NN and LDA
Mahdavian et al. Acoustic features of vocalization signal in poultry health monitoring
WO2016155047A1 (en) Method of recognizing sound event in auditory scene having low signal-to-noise ratio
Ulukaya et al. Overcomplete discrete wavelet transform based respiratory sound discrimination with feature and decision level fusion
US10548534B2 (en) System and method for anhedonia measurement using acoustic and contextual cues
WO2019232848A1 (en) Voice distinguishing method and device, computer device and storage medium
Noda et al. Using bioacoustic signals and support vector machine for automatic classification of insects
CN110444230A (en) A kind of chickens respiratory disease automatic identification equipment based on Audiotechnica
CN115410711B (en) White feather broiler health monitoring method based on sound signal characteristics and random forest
JP2020513914A (en) How to characterize sleep-disordered breathing
Abdollahpur et al. Cycle selection and neuro-voting system for classifying heart sound recordings
WO2020000523A1 (en) Signal processing method and apparatus
Fischer et al. Classification of breath and snore sounds using audio data recorded with smartphones in the home environment
Patti et al. Methods for classification of nocturnal migratory bird vocalizations using Pseudo Wigner-Ville Transform
WO2023216172A1 (en) Poultry voiceprint recognition method and system
Chen et al. An intelligent nocturnal animal vocalization recognition system
Jha et al. Discriminant feature vectors for characterizing ailment cough vs. simulated cough
US20230320329A1 (en) Poultry voiceprint identification method and system
Godino-Llorente et al. Discriminative methods for the detection of voice disorders
TW202341127A (en) Poultry voiceprint identification method and system

Legal Events

Date Code Title Description
121 Ep: the epo has been informed by wipo that ep was designated in this application

Ref document number: 22941130

Country of ref document: EP

Kind code of ref document: A1