CN118094336A - Underwater sound signal identification method based on integrated learning and sample synthesis - Google Patents

Underwater sound signal identification method based on integrated learning and sample synthesis Download PDF

Info

Publication number
CN118094336A
CN118094336A CN202410302626.1A CN202410302626A CN118094336A CN 118094336 A CN118094336 A CN 118094336A CN 202410302626 A CN202410302626 A CN 202410302626A CN 118094336 A CN118094336 A CN 118094336A
Authority
CN
China
Prior art keywords
sample
data set
classifier
weak
neural network
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202410302626.1A
Other languages
Chinese (zh)
Inventor
夏飞
窦钰涛
赵祥
赵飞
张洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hunan University
Original Assignee
Hunan University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hunan University filed Critical Hunan University
Priority to CN202410302626.1A priority Critical patent/CN118094336A/en
Publication of CN118094336A publication Critical patent/CN118094336A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • G06F18/2148Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the process organisation or structure, e.g. boosting cascade
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N20/00Machine learning
    • G06N20/20Ensemble learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/0464Convolutional networks [CNN, ConvNet]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2218/00Aspects of pattern recognition specially adapted for signal processing
    • G06F2218/12Classification; Matching
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/30Assessment of water resources

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Artificial Intelligence (AREA)
  • Evolutionary Computation (AREA)
  • General Physics & Mathematics (AREA)
  • Software Systems (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Computing Systems (AREA)
  • Mathematical Physics (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Evolutionary Biology (AREA)
  • Medical Informatics (AREA)
  • Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)

Abstract

The application belongs to the technical field of underwater acoustic signal processing, and particularly relates to an underwater acoustic signal identification method based on integrated learning and sample synthesis. The method comprises the following steps: collecting underwater sound signal sample data and manufacturing an initial data set; collecting underwater sound background sample data and manufacturing a background noise data set; iteratively training a weak classifier; constructing a random initialized neural network to perform repeated training to obtain a weak classifier; calculating the data set by using a weak classifier, and calculating total errors and weights; adding a sample to be identified by a synthesis method to generate a new data set; combining the weak classifiers into a strong classifier Q in a weighting way; the application trains a plurality of weak classifiers for classifying underwater acoustic signals by adopting an integrated learning method, and integrates the weak classifiers in a weighting mode, thereby obtaining a classifier with more powerful functions. Especially, the recognition errors of all weak classifiers are considered in the process of designing weights, so that the recognition capability of the integrated classifier is further improved.

Description

Underwater sound signal identification method based on integrated learning and sample synthesis
Technical Field
The invention belongs to the technical field of underwater acoustic signal processing, and particularly relates to an underwater acoustic signal identification method based on integrated learning and sample synthesis.
Background
With the increasing activity of human exploration of the ocean, marine environmental awareness has become a popular research direction. Because seawater has strong attenuation on electromagnetic signals, the electromagnetic signals cannot be adopted to carry out long-distance information transmission underwater, and the activities of human beings in the ocean are seriously hindered. Fortunately, sound waves have better transmission capability in sea water as a mechanical wave, so that the sound waves become a main mode of current underwater information transmission and information perception.
In a seawater environment, different targets produce acoustic signals of different characteristics. The classification of underwater sound targets and clutter plays an important role in fishery production, marine information industry, scientific investigation and research and national defense military. However, the ocean has a lot of background noise at any time due to the wave, tide, ocean current, etc., and the recognition of the water supply sound signal is very difficult. Accurate identification of underwater acoustic signals is a very challenging task that has become an important issue to be addressed in marine environmental awareness. The underwater sound signal identification comprises the stages of data acquisition, feature extraction, classification identification and the like. Because of the complex environment and the very rare objects, the acquisition of the underwater acoustic signals is very difficult, and the obtained sample library also usually has the problem of unbalanced categories, namely, a plurality of samples in certain categories and a very small number of samples in certain rare categories. On the other hand, the background noise of the underwater sound signal is large, so that the signal characteristics are difficult to extract, the prior cognitive deviation exists in the traditional manual characteristic extraction method, and certain distinguishing characteristics are possibly omitted, so that the recognition accuracy is low.
Disclosure of Invention
Aiming at the existing problems, the invention provides an integrated learning based on boost, which is used as a basic recognition framework, and a plurality of individual learners are trained successively to improve the recognition accuracy of signals; the method is characterized in that the category with low recognition rate is effectively processed, the number of samples is increased by adopting a signal synthesis method with adjustable parameters, and the method is used for recognizing the underwater sound signals based on the characteristic of multiple sound signals as the input of an individual learner so as to improve the richness and the diversity of information.
In order to achieve the above purpose, the present invention adopts the following technical scheme.
1. An underwater sound signal identification method based on integrated learning and sample synthesis comprises the following steps:
Step 1: collecting underwater sound signal sample data and manufacturing an initial data set D; wherein the data set D contains C categories of underwater acoustic signals;
step 2: collecting underwater sound background sample data and manufacturing a background noise data set Z; wherein, the underwater sound background sample does not contain a target sound signal;
Step 3: iteratively training T weak classifiers; the method specifically comprises the following steps: initializing operation, namely constructing a data set D 1 = D required by the 1 st weak classifier; training the 2 nd to the T th weak classifiers and corresponding training sets D 2 to D T in turn on the basis; wherein, the training process of the t weak classifier is as follows from 3.1 to 3.4:
3.1: constructing a random initialized neural network N t, wherein the neural network adopts a convolutional neural network, and parameters of the neural network N t are randomly set;
3.2: randomly sampling samples from the dataset D t to train the neural network N t;
3.2.1: randomly extracting a sample x from the data set D t, and calculating a Mel spectrum m of the x;
3.2.2: inputting m to a neural network N t, and adjusting the parameter value of N t by using a back propagation algorithm;
3.2.3: repeating training based on the steps 2.2.1-2.2.2 to obtain a neural network Nt, namely a t weak classifier F t;
3.3: performing one-time operation on all samples in the data set D t by using a weak classifier F t, calculating the total error e t of F t on the data set D t, and calculating the total weight a t of F t on the final strong classifier by using e t;
e t is calculated as follows:
Wherein, P () represents a desire, I () is an indication function, x i represents an I-th sample in the data set D t, and y i represents a class label corresponding to the I-th sample;
the calculation formula of a t is as follows:
Step 3.4: adding a sample to be identified by a synthesis method to generate a new data set D t+1; specifically:
3.4.1: for one sample x in the data set D t, carrying out reasoning and prediction operation on the sample x by using a classifier Ft, wherein the last layer of the neural network of F t outputs a probability distribution g which represents the probability that the sample is judged to be of each category;
If sample x is classified correctly, sample x need not synthesize a new sample;
If sample x is not classified correctly, then a new sample needs to be synthesized using x; the calculation formula of the synthesized sample number n is as follows:
Wherein the degree of difficulty in recognition of the sample is evaluated with 1-max (g);
Detecting an effective fragment s in a sample x by adopting a double-threshold endpoint detection algorithm [2], cutting the effective fragment s out of the sample x, randomly extracting n pieces of background noise from a background noise library Z, and then overlapping the effective fragment s on the background noise to form n new samples;
3.4.2: the operations of step 2.4.1 are performed for each sample in the dataset D t;
3.4.3: adding all new samples synthesized in step 2.4.2 to dataset D t, generating new dataset D t+1;Dt+1 to be used for training of the next weak classifier;
Step 4: combining the weak classifiers into a strong classifier Q in a weighting way; the algorithm for the weighted combination is as follows: Wherein a t is the weight of the t weak classifier calculated in step 3.3, and F t is the trained t weak classifier.
The beneficial effects are that:
A plurality of weak classifiers for classifying underwater acoustic signals are trained by adopting an ensemble learning method, and are integrated in a weighting mode, so that a classifier with more powerful functions is obtained. Especially, the recognition errors of all weak classifiers are considered when the weights are designed (the weak classifiers with smaller errors occupy larger weights), so that the recognition capability of the integrated classifier is further improved.
In the training process of the weak classifier, a sample repetition synthesis method is adopted, so that the problem of insufficient data of the underwater sound samples is solved, and especially, different numbers of new samples can be iteratively synthesized according to the recognition difficulty degree of different samples in the sample synthesis process (namely, samples which are successfully recognized do not synthesize new samples any more, and the more difficult samples are recognized, the more new samples are synthesized), so that the pertinence and the effectiveness of sample synthesis are improved.
Drawings
Fig. 1 is a schematic flow diagram of a method for identifying underwater acoustic signals based on ensemble learning and sample synthesis.
Detailed Description
The present invention will be described in detail with reference to specific examples.
The underwater sound signal identification method based on integrated learning and sample synthesis is designed aiming at the problem of insufficient samples in underwater sound data, and achieves the purposes of reducing classifier errors, expanding data sample capacity and improving underwater sound signal identification efficiency and accuracy by means of integrated learning classification, sample repeated synthesis and the like.
The method mainly comprises the following steps:
Step 1: collecting underwater sound signal sample data and manufacturing an initial data set D; wherein the data set D contains C categories of underwater acoustic signals;
step 2: collecting underwater sound background sample data and manufacturing a background noise data set Z; wherein, the underwater sound background sample does not contain a target sound signal;
Step 3: iteratively training T weak classifiers; the method specifically comprises the following steps: initializing operation, namely constructing a data set D 1 = D required by the 1 st weak classifier; training the 2 nd to the T th weak classifiers and corresponding training sets D 2 to D T in turn on the basis; wherein, the training process of the t weak classifier is as follows from 3.1 to 3.4:
3.1: constructing a random initialized neural network N t, wherein the neural network adopts a convolutional neural network, and parameters of the neural network N t are randomly set;
3.2: randomly sampling samples from the dataset D t to train the neural network N t;
3.2.1: randomly extracting a sample x from the data set D t, and calculating a Mel spectrum m of the x;
3.2.2: inputting m to a neural network N t, and adjusting the parameter value of N t by using a back propagation algorithm;
3.2.3: repeating training based on the steps 2.2.1-2.2.2 to obtain a neural network Nt, namely a t weak classifier F t;
3.3: performing one-time operation on all samples in the data set D t by using a weak classifier F t, calculating the total error e t of F t on the data set D t, and calculating the total weight a t of F t on the final strong classifier by using e t;
e t is calculated as follows:
Wherein, P () represents a desire, I () is an indication function, x i represents an I-th sample in the data set D t, and y i represents a class label corresponding to the I-th sample;
the calculation formula of a t is as follows:
Step 3.4: adding a sample to be identified by a synthesis method to generate a new data set D t+1; specifically:
3.4.1: for one sample x in the dataset Dt, carrying out reasoning and prediction operation on the sample x by using a classifier F t, wherein the last layer of the neural network of F t outputs a probability distribution g which represents the probability that the sample is judged to be of each category;
If sample x is classified correctly, sample x need not synthesize a new sample;
If sample x is not classified correctly, then a new sample needs to be synthesized using x; the calculation formula of the synthesized sample number n is as follows:
Wherein the degree of difficulty in recognition of the sample is evaluated with 1-max (g);
Detecting an effective fragment s in a sample x by adopting a double-threshold endpoint detection algorithm [2], cutting the effective fragment s out of the sample x, randomly extracting n pieces of background noise from a background noise library Z, and then overlapping the effective fragment s on the background noise to form n new samples;
3.4.2: the operations of step 2.4.1 are performed for each sample in the dataset D t;
3.4.3: adding all new samples synthesized in step 2.4.2 to dataset D t, generating new dataset D t+1;Dt+1 to be used for training of the next weak classifier;
Step 4: combining the weak classifiers into a strong classifier Q in a weighting way; the algorithm for the weighted combination is as follows: Wherein a t is the weight of the t weak classifier calculated in step 3.3, and F t is the trained t weak classifier.
The following is a description of specific embodiments.
Firstly, a water sound classification data set is collected and constructed, and mainly comprises water sound data of six categories of ships, namely, a tanker, a tug, a passenger ship, a cargo ship, a yacht and a submarine. The first five ships collect 100 underwater sound samples respectively, the occurrence rate of the submarine is low, signals are difficult to capture, and therefore 20 underwater sound samples are collected.
Based on the principle and the steps, and the classification processing scheme design comparison experiment commonly used at present, the recognition conditions of the classifier under different conditions are compared.
The convolutional neural network suitable for classification adopts Resnet, 3 weak classifiers (i.e. t=3) are adopted in boosting ensemble learning, and the training round of each weak classifier is 100 (i.e. e=100).
The identification accuracy of six kinds of ships under different conditions is shown in table 1:
TABLE 1 identification accuracy of six ships under different conditions
Oil tanker Tugboat Passenger ship Cargo ship Yacht Submarine craft
Single weak classifier 63% 68% 59% 66% 62% 35%
The 3 weak classifiers are integrated without adopting a data synthesis algorithm 81% 84% 78% 85% 84% 65%
3 Weak classifiers are integrated and a data synthesis algorithm is adopted 93% 95% 91% 96% 93% 90%
As can be seen from table 1:
(1) The identification accuracy of a single weak classifier on five common ships (oil tankers, tugboats, passenger ships, cargo ships and yachts) is about 60%, and the identification accuracy of a submarine is only 35% because of few samples.
(2) After the integrated learning is adopted, the recognition accuracy of all ships is obviously improved, the recognition accuracy of five common ships reaches about 80%, and the recognition accuracy of the submarine is improved to 65%.
(3) When the integrated learning is adopted and the data synthesis algorithm is adopted, the recognition accuracy of five common ships is improved to more than 90%, and the recognition accuracy of the submarine is also improved to 90%. The improvement of the accuracy of the underwater vehicle is most obvious mainly because the sample recognition difficulty of the underwater vehicle is the greatest, and the data synthesis algorithm synthesizes the most new samples aiming at the samples of the underwater vehicle, so that the recognition accuracy of the underwater vehicle is remarkably improved.
Finally, it should be noted that the above embodiments are only for illustrating the technical solution of the present invention, and not for limiting the scope of the present invention, and although the present invention has been described in detail with reference to the preferred embodiments, it should be understood by those skilled in the art that modifications or equivalent substitutions can be made to the technical solution of the present invention without departing from the spirit and scope of the technical solution of the present invention.

Claims (1)

1. The underwater sound signal identification method based on integrated learning and sample synthesis is characterized by comprising the following steps of:
Step 1: collecting underwater sound signal sample data and manufacturing an initial data set D; wherein the data set D contains C categories of underwater acoustic signals;
step 2: collecting underwater sound background sample data and manufacturing a background noise data set Z; wherein, the underwater sound background sample does not contain a target sound signal;
Step 3: iteratively training T weak classifiers; the method specifically comprises the following steps: initializing operation, namely constructing a data set D 1 = D required by the 1 st weak classifier; training the 2 nd to the T th weak classifiers in turn on the basis of the training, and constructing corresponding training sets D 2 to D T; wherein, the training process of the t weak classifier is as follows from 3.1 to 3.4:
3.1: constructing a random initialized neural network N t, wherein the neural network adopts a convolutional neural network, and parameters of the neural network N t are randomly set;
3.2: randomly sampling samples from the dataset D t to train the neural network N t;
3.2.1: randomly extracting a sample x from the data set D t, and calculating a Mel spectrum m of the x;
3.2.2: inputting m to a neural network N t, and adjusting the parameter value of N t by using a back propagation algorithm;
3.2.3: repeating training based on the steps 3.2.1-3.2.2 to obtain a neural network N t, namely a t weak classifier F t;
3.3: performing one-time operation on all samples in the data set D t by using a weak classifier F t, calculating the total error e t of F t on the data set D t, and calculating the total weight a t of F t on the final strong classifier by using e t;
e t is calculated as follows:
Wherein, P () represents a desire, I () is an indication function, x i represents an I-th sample in the data set D t, and y i represents a class label corresponding to the I-th sample;
the calculation formula of a t is as follows:
Step 3.4: adding a sample to be identified by a synthesis method to generate a new data set D t+1; specifically:
3.4.1: for one sample x in the data set D t, carrying out reasoning and prediction operation on the sample x by using a classifier F t, wherein the last layer of the neural network of F t outputs a probability distribution g which represents the probability that the sample is judged to be of each category;
If sample x is classified correctly, sample x need not synthesize a new sample;
If sample x is not classified correctly, then a new sample needs to be synthesized using x; the calculation formula of the synthesized sample number n is as follows:
Wherein the degree of difficulty in recognition of the sample is evaluated with 1-max (g);
Detecting an effective fragment s in a sample x by adopting a double-threshold endpoint detection algorithm [2], cutting the effective fragment s out of the sample x, randomly extracting n pieces of background noise from a background noise library Z, and then overlapping the effective fragment s on the background noise to form n new samples;
3.4.2: the operations of step 2.4.1 are performed for each sample in the dataset D t;
3.4.3: adding all new samples synthesized in step 2.4.2 to dataset D t, generating new dataset D t+1;Dt+1 to be used for training of the next weak classifier;
Step 4: combining the weak classifiers into a strong classifier Q in a weighting way; the algorithm for the weighted combination is as follows: Wherein a t is the weight of the t weak classifier calculated in step 3.3, and F t is the trained t weak classifier.
CN202410302626.1A 2024-03-18 2024-03-18 Underwater sound signal identification method based on integrated learning and sample synthesis Pending CN118094336A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202410302626.1A CN118094336A (en) 2024-03-18 2024-03-18 Underwater sound signal identification method based on integrated learning and sample synthesis

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202410302626.1A CN118094336A (en) 2024-03-18 2024-03-18 Underwater sound signal identification method based on integrated learning and sample synthesis

Publications (1)

Publication Number Publication Date
CN118094336A true CN118094336A (en) 2024-05-28

Family

ID=91161410

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202410302626.1A Pending CN118094336A (en) 2024-03-18 2024-03-18 Underwater sound signal identification method based on integrated learning and sample synthesis

Country Status (1)

Country Link
CN (1) CN118094336A (en)

Similar Documents

Publication Publication Date Title
CN101470194B (en) Torpedo target recognition method
CN104732970A (en) Ship radiation noise recognition method based on comprehensive features
CN109711314B (en) Radar radiation source signal classification method based on feature fusion and SAE
CN112666533B (en) Repetition frequency change steady target identification method based on spatial pyramid pooling network
CN115114949A (en) Intelligent ship target identification method and system based on underwater acoustic signals
CN113640768B (en) Low-resolution radar target identification method based on wavelet transformation
CN111010356A (en) Underwater acoustic communication signal modulation mode identification method based on support vector machine
Jin et al. Offshore ship recognition based on center frequency projection of improved EMD and KNN algorithm
Mouy et al. Automatic recognition of fin and blue whale calls for real-time monitoring in the St. Lawrence
CN1609631A (en) Method for testing and identifying underwater sound noise based on small wave area
CN103116740A (en) Method and device for identifying underwater targets
White et al. More than a whistle: Automated detection of marine sound sources with a convolutional neural network
Alouani et al. A spatio-temporal deep learning approach for underwater acoustic signals classification
Zhou et al. An attention-based multi-scale convolution network for intelligent underwater acoustic signal recognition
Ju et al. A new low SNR underwater acoustic signal classification method based on intrinsic modal features maintaining dimensionality reduction
CN118094336A (en) Underwater sound signal identification method based on integrated learning and sample synthesis
Vahidpour et al. An automated approach to passive sonar classification using binary image features
Hummel et al. A survey on machine learning in ship radiated noise
CN103308918A (en) Fish identification method and system based on segmented time-domain centroid features
Bach Improving the classification of propeller ships using LOFAR and triple loss variational auto encoder
Farrokhrooz et al. Ship noise classification using probabilistic neural network and AR model coefficients
Xie et al. DBM-based underwater acoustic source recognition
CN114545422A (en) Active sonar target identification method based on multiple physical characteristics
Lopez-Risueno et al. Target detection in sea clutter using convolutional neural networks
CN111624585A (en) Underwater target passive detection method based on convolutional neural network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination