AU2020287677B2 - Sound source position estimation method, readable storage medium and computer device - Google Patents
Sound source position estimation method, readable storage medium and computer device Download PDFInfo
- Publication number
- AU2020287677B2 AU2020287677B2 AU2020287677A AU2020287677A AU2020287677B2 AU 2020287677 B2 AU2020287677 B2 AU 2020287677B2 AU 2020287677 A AU2020287677 A AU 2020287677A AU 2020287677 A AU2020287677 A AU 2020287677A AU 2020287677 B2 AU2020287677 B2 AU 2020287677B2
- Authority
- AU
- Australia
- Prior art keywords
- signal
- window
- signals
- sound source
- length
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Ceased
Links
- 238000000034 method Methods 0.000 title claims abstract description 45
- 238000004422 calculation algorithm Methods 0.000 claims abstract description 17
- 238000013528 artificial neural network Methods 0.000 claims abstract description 7
- 230000000306 recurrent effect Effects 0.000 claims abstract description 7
- 239000002245 particle Substances 0.000 claims description 22
- 230000006870 function Effects 0.000 claims description 20
- 239000013589 supplement Substances 0.000 claims description 20
- 238000001228 spectrum Methods 0.000 claims description 13
- 230000004927 fusion Effects 0.000 claims description 9
- 238000004364 calculation method Methods 0.000 claims description 7
- 238000007476 Maximum Likelihood Methods 0.000 claims description 6
- 238000004590 computer program Methods 0.000 claims description 6
- 238000009795 derivation Methods 0.000 claims description 4
- 208000001992 Autosomal Dominant Optic Atrophy Diseases 0.000 claims description 3
- 206010011906 Death Diseases 0.000 claims description 3
- 238000002790 cross-validation Methods 0.000 claims description 3
- 125000004122 cyclic group Chemical group 0.000 claims description 3
- 230000001360 synchronised effect Effects 0.000 claims description 3
- 230000008569 process Effects 0.000 description 5
- 238000005314 correlation function Methods 0.000 description 3
- 238000011160 research Methods 0.000 description 3
- 239000003643 water by type Substances 0.000 description 3
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000009286 beneficial effect Effects 0.000 description 1
- 238000007796 conventional method Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 230000018109 developmental process Effects 0.000 description 1
- 230000000694 effects Effects 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000011218 segmentation Effects 0.000 description 1
- 230000005236 sound signal Effects 0.000 description 1
- 230000004083 survival effect Effects 0.000 description 1
- 230000002123 temporal effect Effects 0.000 description 1
- 238000012360 testing method Methods 0.000 description 1
- 238000010200 validation analysis Methods 0.000 description 1
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S3/00—Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received
- G01S3/80—Direction-finders for determining the direction from which infrasonic, sonic, ultrasonic, or electromagnetic waves, or particle emission, not having a directional significance, are being received using ultrasonic, sonic or infrasonic waves
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S5/00—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations
- G01S5/18—Position-fixing by co-ordinating two or more direction or position line determinations; Position-fixing by co-ordinating two or more distance determinations using ultrasonic, sonic, or infrasonic waves
- G01S5/20—Position of source determined by a plurality of spaced direction-finders
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Remote Sensing (AREA)
- Radar, Positioning & Navigation (AREA)
- Theoretical Computer Science (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Biophysics (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Computational Linguistics (AREA)
- Biomedical Technology (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Artificial Intelligence (AREA)
- Life Sciences & Earth Sciences (AREA)
- Health & Medical Sciences (AREA)
- Measurement Of Velocity Or Position Using Acoustic Or Ultrasonic Waves (AREA)
Abstract
Disclosed is a sound source position estimation method. The method comprises: a single-vector hydrophone receiving multichannel signals sent by a sound source in the sea; fusing the received multichannel signals into an instant single-channel sound intensity signal by means of combined sliding of a fixed window and a dynamic window, dividing the instant single-channel sound intensity signal into signal segments containing a sufficient amount of information, and with the premise that the information amount is guaranteed, reducing data quantity, and improving the operating speed; performing self-supplementation on the signal by means of an expectation maximization algorithm, and expanding distances between the various signal segments, such that the resolution between the signal segments is improved, and information lost during interception in the previous step is supplemented to a certain extent; and estimating the position of the sound source by means of a recurrent neural network and using self-supplemented equal-length signals. According to the method, only the single-vector hydrophone is needed to collect signals, such that not only are arrangement difficulty and usage costs simplified, but the application range is also expanded. Further disclosed are a readable storage medium and a computer device.
Description
Technical Field
The present disclosure relates to the field of sound source position estimation technologies, and in particular, to a sound source position estimation method, a readable storage medium, and a computer device.
Related Art
The description in this section merely provides background information related to the present disclosure, and does not necessarily constitute the prior art.
With rapid and continuous growth in economy in China, population also continuously increases, and people's demands for resources and requirements for consumption levels also increase greatly accordingly. People are developing and utilizing the limited resources on the land to the greatest extent, and are also facing resource shortage at the same time. Therefore, new fields and new resources need to be further researched and developed on the basis of maximizing the application of existing resources. The ocean area accounts for more than 70% of the surface area of the earth, and therefore, it is crucial to develop and utilize marine resources. In recent years, countries in the world are quite fiercely competing for marine resources. How to develop and utilize marine resources to make full use of the resources has become a major issue that attracts much attention in recent years. China is in a unique position, with a total national territorial area of approximately 9.6 million square kilometers, of which the ocean area accounts for approximately 3% of the total area, and most of the ocean waters are shallow sea waters. Therefore, technical research on shallow sea waters will be of greater significance for human survival and development.
Compared with the deep sea environment, the temporal and spatial variability and uncertainty of the shallow sea environment have a severer impact on signal propagation. In addition, both reflected signals from the shallow sea floor and human activities in the shallow sea cause aliasing of a target signal, further affecting sound source position estimation. Therefore, sound source target estimation in the shallow sea environment has been a research difficulty in this field currently. Compared with a conventional sound pressure hydrophone, a vector hydrophone can acquire a sound pressure signal and triaxial acoustic particle velocity signals in orthogonal directions at a common point, and is of great practical value.
The inventor of the present disclosure finds during research that: currently, a single vector hydrophone is mostly used to estimate an azimuth direction of arrival (DOA) and a pitch DOA of a target, and to determine a position distance of the target, cross-estimation needs to be performed by using a vector hydrophone array due to an inherent defect of the single vector hydrophone, namely, insufficient distance resolution. Moreover, in actual engineering application of the single vector hydrophone, on the one hand, it is difficult for an actual parameter to meet a requirement of an ideal electroacoustic parameter characteristic due to limitations of process conditions, restricting the accuracy of position estimation of the single vector hydrophone; on the other hand, the single vector hydrophone is easily influenced by the environment to cause uncertain changes in attitude, so that acquisition of a real position of the target is further influenced. Such reasons result in insufficient application of the single vector hydrophone in target position estimation.
To overcome the shortcomings in the related art, the present disclosure provides a sound source position estimation method, a readable storage medium, and a computer device. Compared with a conventional sound source estimation model in which a complex vector hydrophone array needs to be deployed to receive a signal, in the sound source estimation method, only a single vector hydrophone is required to acquire a signal, thereby reducing deployment difficulty and use costs, and expanding the application scope.
To achieve the foregoing objective, the present disclosure uses the following technical solutions:
According to a first aspect, the present disclosure provides a sound source position estimation method.
The sound source position estimation method includes the following steps:
receiving, by a single vector hydrophone, multichannel signals from a sound source in the ocean;
fusing the received multichannel signals into an instantaneous single-channel sound intensity signal through fixed-dynamic window combined sliding, and dividing the received multichannel signals into signal segments including a sufficient information amount, to reduce a data amount and improve an operation speed while ensuring the information amount;
performing signal self-supplement by using the expectation maximization algorithm, and increasing a distance between the signal segments, to improve a resolution between the signal segments and supplement information lost in the previous step of capture to some extent;
estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement.
In some possible implementations, the multichannel signals are four-channel signals, including three acoustic particle velocity signals in orthogonal directions: an acoustic
particle velocity v- in an X -axis direction, an acoustic particle velocity vY in a Y -axis
direction, and an acoustic particle velocity v- in a z -axis direction, and a scalar sound
pressure signal P.
As a further limitation, the multichannel signals are fused into the instantaneous single-channel sound intensity signal by using a fixed window, dynamic windows of all lengths are traversed, a steepest ascent section of a Shannon entropy is found, an optimum dynamic window is determined, the instantaneous single-channel sound intensity signal in the fixed window are dynamically captured as unequal-length signals based on the Shannon entropy by using the optimum dynamic window, and signal self-supplement is performed on the captured unequal-length signals by using the expectation maximization algorithm.
As a still further limitation, the dividing the received multichannel signals into signal segments including a sufficient information amount through fixed-dynamic window combined sliding specifically includes the following steps:
401: specifying a fixed window length if and an initial window start point tf for the
acquired four-channel signals P, v, v, and v
402: fusing, by using a fixed window Wh_ whose window length and start point are
respectively If and tf , four-channel information in the window, to obtain an
instantaneous single-channel sound intensity signal whose length is ;
W 403: capturing a dynamic window d, whose window length and start point are
respectively I and t in the instantaneous single-channel sound intensity signal '
, where the length of the signal in the dynamic window ' is as short as possible provided that the signal includes a sufficient information amount; and
404: returning to 402, updating a start point tf of a fixed window W',' based on a
signal overlap rate q, and performing a cyclic operation.
As a still further limitation, in step 402, synchronous sliding is performed in channel signals through a time window of a fixed size, signals are extracted, and the information is fused into the instantaneous single-channel sound intensity signal by using a cross-spectrum method; and the step specifically includes the following steps:
501: specifying a fixed window length ' and a window start point t according to a signal fusion degree;
502: capturing signal segments of a window size by using the same start point t
in the signal channels of the sound pressure P and the acoustic particle velocities v and in the axial directions, where corresponding window signals are Wp,
" and ,Z; and
503: calculating, based on the cross-spectrum method, the instantaneous single-channel W sound intensity signal h obtained after the window signals are fused, to implement multi-sensor information fusion, where a calculation formula of the instantaneous single-channel sound intensity signal obtained after the fusion is:
Re[S, ] Re[S, ] I~p(t).V,(t)R "+p(t).v Re[S'() (t) Re[S'() ' +p(t).v,(t) 1" Re[S'()
5 where SP, " , and P' are respectively cross-spectrum functions of three
components x, y, and z, S is a spectrum function of P , f is a frequency,
Re[ I indicates Laplace transform, 0 and 9 are respectively a pitch direction of arrival (DOA) and an azimuth DOA of a sound source relative to a vector hydrophone, where the pitch DOA and the azimuth DOA respectively take the XOY plane and the x
-axis as 0°, and p(t) v(t) v,(t), and v(t) are respectively a sound pressure signal and acoustic particle velocity signals in different directions that are received by the vector hydrophone at a moment t.
As a still further limitation, in step 403, the dynamic windows of all lengths are W traversed for the instantaneous single-channel sound intensity signal ', in the fixed window, and the steepest ascent section of the Shannon entropy is found as the optimum W dynamic window f; and the step specifically includes the following steps:
601: traversing the entire fixed window 6 of the captured instantaneous
single-channel sound intensity signal I starting from the start points of the fixed window, and calculating a Shannon entropy of signals of all lengths by using the following
formula, to construct a Shannon entropy signal ',
Shannon(X )=p(x,)-log =-p(x,)-1ogp(x,) p(x1 ) i i
where 't is a possible value of a random event X, Shannon(X) is a Shannon
entropy included in the random event X, m is a total quantity of random events, and
p(x,) is a probability of occurrence of* ;
602: finding a steepest ascent section of SI, according to a derivation result S' of
marking the length as , andjumping to step 604;
603: marking the length as if no steepest ascent section is found in W considering that the signal in the fixed window h is an invalid signal or a noise signal, where the following two conditions are satisfied:
when S is relatively small, it is considered that the signal is a null signal or Shannon
entropy content of the signal in the fixed window is insufficient, and I °, where o is a preset minimum captured length; and
when S is relatively large, it is considered that the signal is a noise signal or a valid
signal including a relatively high Shannon entropy, and =/, where 1 is a preset maximum captured length;
604: capturing a signal segment whose length is starting from the start point tf in W the fixed window h as the dynamic window Wf' ,and marking an end time of the
window ast
As a still further limitation, signal self-supplement is performed by using the expectation maximization algorithm, segmented unequal-length signals are equivalent to
observation data X, supplemented equal-length signals are equivalent to complete data Y,
a supplementary signal is equivalent to unobserved data Z, and a result is iterated by using the expectation maximization algorithm to obtain a maximum value 0 of a parameter 0, that is, when a maximum value of a maximum likelihood function L() based on Y is 2 obtained, optimal solutions of a mean u and a variance 2 of a complete dataset are obtained, and an unknown dataset Z is obtained based on an observed dataset X, to supplement the complete dataset Y, which specifically includes the following steps:
701: assuming that an iteration quantity t = 0, initializing a parameter vector 0(0), 0
being a parameter vector formed by the mean and the variance of the dataset Y, and
calculating an initial maximum likelihood function LO) (0)
n
702: obtaining Qi ( ) from 0t, and ensuring that when 0 is specified, the
equal sign in In(E(X))E[In(X)]is satisfied, to establish a lower bound of ;
703: fixing Qt)(zi) Q~~z0t and taking 0( as a variable, taking a derivative of (t) ))in
step 702, and obtaining O(t+1) according to the formula
L(60'*0) Q ) z In P ( Zi 0t).=L(6t)) ; and
704: if L (, ending the iterative calculation; otherwise, assuming
that t=t+1, and returning to step 702, where a threshold 6 is a specified very small value;
where represents a distribution of unknown data Z, p(x z;0(t)) is a
probability of occurrence of x, z under a condition (t), the superscript i indicates an it value of a corresponding parameter, - is the threshold and is an initially specified small value used as a standard for ending the iteration, and E[ ] is a mathematical expectation.
As a still further limitation, the estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement specifically includes: using the supplemented signal segments as an input by using the expectation maximization algorithm, outputting azimuth DOAs and distances of the sound source in different signal segments, and performing cross-validation on estimation results of the different signal segments, to implement accurate positioning of the position of the sound source.
According to a second aspect, the present disclosure provides a computer-readable storage medium, storing a computer program, where when the program is executed by a processor, steps in the sound source position estimation method of the present disclosure are implemented.
According to a third aspect, the present disclosure provides a computer device, including a memory, a processor, and a computer program stored in the memory and capable of being run on the processor, where when the processor executes the program, steps in the sound source position estimation method of the present disclosure are implemented.
Compared with the prior art, the present disclosure has the following beneficial effects:
1. Compared with the conventional sound source estimation model, the sound source position estimation method of the present disclosure avoids a problem of deploying a complex vector hydrophone array to receive a signal. In the sound source estimation method of this application, only a single vector hydrophone is required to acquire a signal, thereby reducing deployment difficulty and use costs, and expanding the application scope.
2. In the sound source position estimation method of the present disclosure, a short-duration signal sample is divided into a large quantity of signal segments through dynamic window-fixed window combined sliding. Through mutual validation of the signal segments, the accuracy and stability of position estimation are improved. The data amount is reduced and the operation speed is improved while ensuring the information amount.
3. In the sound source position estimation method of the present disclosure, large samples are required to train a network only in an early stage, and complex operations are not required during use, thereby implementing real-time tracking of high-speed and highly maneuverable target trajectories.
FIG. 1 shows a relationship between a single vector hydrophone and a sound source position according to Embodiment 1 of the present disclosure.
FIG. 2 is a flowchart of a sound source position estimation method according to Embodiment 1 of the present disclosure.
FIG. 3 is a flowchart of fixed-dynamic window combined sliding according to Embodiment 1 of the present disclosure.
FIG. 4 is a curve graph of an estimation result of a position of an ultra-low frequency sound source according to Embodiment 1 of the present disclosure.
It should be noted that the following detailed descriptions are all exemplary and are intended to provide a further description of the present disclosure. Unless otherwise specified, all technical and scientific terms used herein have the same meaning as commonly understood by a person of ordinary skill in the art to which the present disclosure belongs.
It should be noted that the terms used herein are merely used for describing specific implementations, and are not intended to limit exemplary implementations of the present disclosure. As used herein, the singular form is intended to include the plural form, unless the context clearly indicates otherwise. In addition, it should further be understood that terms "comprise" and/or "include" used in this specification indicate that there are features, steps, operations, devices, components, and/or combinations thereof.
Embodiment 1:
As shown in FIG. 1 and FIG. 2, Embodiment 1 of the present disclosure provides a sound source position estimation method, including the following steps:
receiving, by a single vector hydrophone, multichannel signals from a sound source in the ocean, where the multichannel signals are four-channel signals, including three acoustic
particle velocity signals in orthogonal directions: an acoustic particle velocity v- in an X
-axis direction, an acoustic particle velocity v in a Y -axis direction, and an acoustic
particle velocity vZ in a z -axis direction, and a scalar sound pressure signal P;
fusing the received multichannel signals into an instantaneous single-channel sound intensity signal through fixed-dynamic window combined sliding, and dividing the received multichannel signals into signal segments including a sufficient information amount, to reduce a data amount and improve an operation speed while ensuring the information amount;
performing signal self-supplement by using the expectation maximization algorithm, and increasing a distance between the signal segments, to improve a resolution between the signal segments and supplement information lost in the previous step of capture to some extent;
estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement.
The multichannel signals are fused into the instantaneous single-channel sound intensity signal by using the fixed window, dynamic windows of all lengths are traversed, a steepest ascent section of a Shannon entropy is found, an optimum dynamic window is determined, the instantaneous single-channel sound intensity signal in the fixed window are dynamically captured as unequal-length signals based on the Shannon entropy by using the optimum dynamic window, and signal self-supplement is performed on the captured unequal-length signals by using the expectation maximization algorithm.
As shown in FIG. 3, the dividing the received multichannel signals into signal segments including a sufficient information amount through fixed-dynamic window combined sliding specifically includes the following steps:
301: specifying a fixed window length/J and an initial window start point tf for
the acquired four-channel signals P, v-, v", and v;
302: fusing, by using a fixed window T whose window length and start point are I tI respectively f and t , four-channel information in the window, to obtain an
instantaneous single-channel sound intensity signal whose length is ;
303: capturing a dynamic window fr whose window length and start point are
respectively and t in the instantaneous single-channel sound intensity signal
where the length of the signal in the dynamic window fr is as short as possible provided that the signal includes a sufficient information amount; and
304: returning to 302, updating a start point t of a fixed window j> based on a
signal overlap rate q, and performing a cyclic operation.
In step 302, synchronous sliding is performed in channel signals through a time window of a fixed size, signals are extracted, and the information is fused into the instantaneous single-channel sound intensity signal by using a cross-spectrum method; and the step specifically includes the following steps:
401: specifying a fixed window length J and a window start point t accordingtoa signal fusion degree;
402: fusing, by using a fixed window Wfi_ whose window length and start point are respectively If and tfi, four-channel information in the window, to obtain an instantaneous single-channel sound intensity signal whose length is j; and
403: calculating, based on the cross-spectrum method, the instantaneous single-channel W sound intensity signal obtained after the window signals are fused, to implement multi-sensor information fusion, where a calculation formula of the instantaneous single-channel sound intensity signal obtained after the fusion is:
SRe[S ] Re[S, ] Re[S ] I =p(t).v,(t ) " ±+p(t).v,(t )R[ ' +p(t ).v,(t ) 1 "eS *~ Re[S'() Re[S,1 (f) Re[S'()
A specific derivation process of the instantaneous single-channel sound intensity signal is:
Assuming that a sound signal P propagates in an isotropic noise field, a vector
hydrophone Q receives the signal, and outputs of the vector hydrophone have the following relationships:
Soundpressure p(t)=p(t)+p(t) (2)
Acoustic particle velocity component x v,(t)= v,(t)+v,(t) (3)
Acoustic particle velocity component y v,,(t)= v,(t)+v,(t) (4)
Acoustic particle velocity component z (t)= v-,(t)+v -(t) (5)
In the foregoing formulas, subscripts "s" and "n " respectively represent signal and noise. If noise sources are independent of each other and a mean is 0, a sound intensity in the X direction is:
I= p~|v,t W =( [p( W+ p, (t1 W -( + +v,(t01
=S pW tv.( W + PS W - v,.( W+ p, W - v.( W+ p,W - v,.( W
=S p(t)- v.( W (6)
Similarly:
Iz =S p(t)- vzS(t)(7
It can be learned from the foregoing formula that the sound intensity obtained from the outputs P , V, , and v of the vector hydrophone does not include noise energy, that is, the vector hydrophone can resist an isotropic noise.
At a moment t, a sound pressure signal and acoustic particle velocity signals in
different directions that are received by the vector hydrophone are respectively p(t)
v,(t) v,(t) and v,(t)
An approximate spatial position of the target is estimated by using the cross-spectrum
method. First, a cross-correlation operation is performed on the sound pressure P and each of the acoustic particle velocity components, to obtain the following cross-correlation functions:
T R = p(t)v,(t-r)dt= p(t)p(t-r)cosOcospdt
R ,=(p~t)v(t-r)dt=J p(t)p(t-r)cosOsinqdt
R),= p(t)vt-r)dt=f p(t)p(t-r)sinOdt (8)
Then, Fourier transform is performed on the foregoing cross-correlation functions, to obtain cross-spectrum functions of the cross-correlation functions:
S = S (f)cos0 CosP S = S (f)cos0 sinp S =S (f)sinO
Herein, 2(fis a spectrum function of P ,f is a frequency, Re[ ]
indicates Laplace transform, and 0, and 9 are respectively a pitch direction of arrival (DOA) and an azimuth DOA of a sound source relative to a vector hydrophone, where the
pitch DOA and the azimuth DOA respectively take theOY plane and the x -axis as0°.
In this case, the azimuth DOA and pitch DOA of the target are:
¶o = arctan SPvvM
0=arctan S132 VS P() (10)
The sound intensity of the vector hydrophone may be obtained with reference to formulas (6), (7), and (10):
I=(I cos0cossp+ cos0sins+I sino)
= p (t) -v (t) cos 0cossop+ ps (t) -v, (t) cos 0sinsop+ ps (t) -v,(t) sin 0 Re[S~ ] Re[S, ] Re[S ] p(t)-ve(t) e[ +pt-,(t) ' +p(t)-v,(t) Re[S'() Re[S'() Re[S ()
$ In step 303, the dynamic windows of all lengths are traversed for the instantaneous
single-channel sound intensity signal w' in the fixed window, and the steepest ascent
section of the Shannon entropy is found as the optimum dynamic window ';and the step specifically includes the following steps:
601: traversing the entire fixed window 6 of the captured instantaneous
single-channel sound intensity signal I starting from the start point of the fixed window, and calculating a Shannon entropy of signals of all lengths by using the following
formula, to construct a Shannon entropy signal ',
Shannon(X) =1p(x,). log =-$p(x,). logp(x,) 1i p(x,) i_1 (12)
where xi is a possible value of a random event X, Shannon(X) is a Shannon
entropy included in the random event X, m is a total quantity of random events, and
p(x,) is a probability of occurrence of2I;
602: finding a steepest ascent section of SI, according to a derivation result " of
marking the length as , andjumping to step 604;
603: marking the length as e if no steepest ascent section is found in S', W considering that the signal in the fixed window ', is an invalid signal or a noise signal, where the following two conditions are satisfied:
when ' is relatively small, it is considered that the signal is a null signal or Shannon
entropy content of the signal in the fixed window is insufficient, and I °, where o is a preset minimum captured length; and
when S is relatively large, it is considered that the signal is a noise signal or a valid
signal including a relatively high Shannon entropy, and t , where 1 is a preset maximum captured length;
604: capturing a signal segment whose length is starting from the start pointtf in W the fixed window h as the dynamic window Wf' ,and marking an end time of the
window as
For the shortcoming of different lengths of the signals obtained after fixed-dynamic window segmentation, signal self-supplement is performed by using the expectation maximization algorithm (EM algorithm). The segmented unequal-length signals are
equivalent to observation data X, the supplemented equal-length signals are equivalent to
complete data Y, and the supplementary signal is equivalent to unobserved data Z. The equal lengths of the signals facilitate subsequent calculation and comparison.
A result is iterated by using the expectation maximization algorithm to obtain a
maximum value 0* of a parameter 0, that is, when a maximum value of a maximum
likelihood function L() based on Y is obtained, optimal solutions of a mean ui and a variance 'i2 of a complete dataset are obtained, and an unknown dataset Z is obtained based on an observed dataset X , to supplement the complete dataset Y, which specifically includes the following steps:
701: assuming that an iteration quantity t 0, initializing a parameter vector (), 0
being a parameter vector formed by the mean and the variance of the dataset Y, and
calculating an initial maximum likelihood function
L(O) = L(xl, --- ,x,;))= p( x ;O)....- p( xn;))=$ p(x,.;0)
Q(t)(i0' 702: obtaining Q z) from 0(t), and ensuring that when 0(t) is specified, the
equal sign in In(E(X))E[In(X)]is satisfied, to establish a lower bound of ;
703: fixing Qt)(z~ and taking (t) as a variable, taking a derivative of ) in
step 702, and obtaining 0(t+1) according to the formula
L (60'0 ?) InQ(zXNi ) ln ) =L(6t)) P) ;and
704: if , ending the iterative calculation; otherwise, assuming
that t=t+1, and returning to step 702, where a threshold - is a specified very small value;
A specific iteration process is as follows:
Assuming that Z represents missing data, that is, unobserved data, X is observed
data, referred to as incomplete data, a sum of the missing data Z and the incomplete data
X is defined as complete data Y, and X is a function of Y, there are the following relations:
(L(O) =inp(X|O) ->max]->([In pY|O) ->max]->([0->0*| (13)
L(O)=L(N,---,x(");O)=p x ;0 -.. p(x";O =$p(x; = (14)
0 (15)
P(X )is a probability density function of an observed dataset, p(Yl@) is a
probability density function of a complete dataset, and u! and 2 are respectively a mean and a variance of a probability density function.
To find a maximum value of the likelihood function L(O) is to find 0 in a
parameter space 0 to maximize the likelihood function when sample points
{I I)are fixed, specifically:
0* = arg max, L (0) (16)
Since L(O) and InL(0) have extreme values at the same 0 , a logarithmic operation is performed on the likelihood function:
In(L(0))= Inp(x(i);o) (17)
A maximum likelihood estimate 0* of 0 may be obtained by solving the following equation:
d InLO)=0 dO (18)
Therefore, formula (13) can be transformed into:
L(O)= (z('))lIn ( 'z;) Qi (z) (19)
P ( )(20)
Qi represents a distribution of unknown data Z , and satisfies the following conditions:
z~Q(z()=
(z)> 0 (21)
from related definitions of a mathematical expectation and the Jensen's inequality:
E[f(x)]= f(x,)p(x,) (22)
ln(E(X))> E[ln(X)] (23)
It is learned with reference to formula (19) that:
0 L(6)= ln E PX'
i 1 - Q Z) (24)
It is learned with reference to formulas (19) and (20) that, in a t th iteration:
>W Q1 (z
) p( x Z ;0) ln E
p(z(') ln xz;0 >W Qz(O)
IQ (z() Iln PXiZi)O1).=L(OW i Z I ) (25)
Formula (25) may be seen as a process of finding a lower bound of L(O) Through
continuous iteration, the lower bound is increased until the lower bound
converges to the vicinity the likelihood function L(O) when the parameter 0 reaches a
maximum value 0*. In this case, the iteration ends.
The estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement specifically includes: using the supplemented signal segments as an input by using the expectation maximization algorithm, outputting azimuth DOAs and distances of the sound source in different signal segments, and performing cross-validation on estimation results of the different signal segments, to implement accurate positioning of the position of the sound source.
To further describe an implementation process of the method, the method is tested by using a signal acquired by a single vector hydrophone deployed at a position when a ship is sailing. Upon testing, it is found that a sound source position can be located in a very short time by using the method, with accuracy of 1.5 m. Compared with a conventional method, only a single vector hydrophone is required, not only improving the positioning accuracy, but also improving the stability. An estimation result is shown in FIG. 4.
Embodiment 2:
Embodiment 2 of the present disclosure provides a computer-readable storage medium, storing a computer program, where when the program is executed by a processor, steps in the sound source position estimation method in Embodiment 1 of the present disclosure are implemented.
Embodiment 3:
Embodiment 3 of the present disclosure provides a computer device, including a memory, a processor, and a computer program stored in the memory and capable of being run on the processor, where when the processor executes the program, steps in the sound source position estimation method in Embodiment 1 of the present disclosure are implemented.
The foregoing descriptions are merely exemplary embodiments of the present disclosure, but are not intended to limit the present disclosure. The present disclosure may include various modifications and changes for a person skilled in the art. Any modification, equivalent replacement, or improvement made within the spirit and principle of the present disclosure shall fall within the protection scope of the present disclosure.
Claims (10)
1. A sound source position estimation method, comprising the following steps:
receiving, by a single vector hydrophone, multichannel signals from a sound source in the ocean;
fusing the received multichannel signals into an instantaneous single-channel sound intensity signal through fixed-dynamic window combined sliding, and dividing the received multichannel signals into signal segments comprising a sufficient information amount;
performing signal self-supplement by using the expectation maximization algorithm, and increasing a distance between the signal segments; and
estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement.
2. The sound source position estimation method according to claim 1, wherein the multichannel signals are four-channel signals, comprising three acoustic particle velocity
signals in orthogonal directions: an acoustic particle velocity VX in an X-axis direction,
an acoustic particle velocity v in a Y -axis direction, and an acoustic particle velocity
in a z -axis direction, and a scalar sound pressure signal P .
3. The sound source position estimation method according to claim 2, wherein the multichannel signals are fused into the instantaneous single-channel sound intensity signal by using a fixed window, dynamic windows of all lengths are traversed, a steepest ascent section of a Shannon entropy is found, an optimum dynamic window is determined, the instantaneous single-channel sound intensity signal in the fixed window are dynamically captured as unequal-length signals based on the Shannon entropy by using the optimum dynamic window, and signal self-supplement is performed on the captured unequal-length signals by using the expectation maximization algorithm.
4. The sound source position estimation method according to claim 3, wherein the dividing the received multichannel signals into signal segments comprising a sufficient information amount through fixed-dynamic window combined sliding specifically comprises the following steps:
401: specifying a fixed window length/J and an initial window start point tf for
the acquired four-channel signals P, v-, v", and v;
402: fusing, by using a fixed window Wh_ whose window length and start point are
respectively J and t, four-channel information in the window, to obtain an
instantaneous single-channel sound intensity signal whose length is ;
403: capturing a dynamic window W't whose window length and start point are
respectively and t in the instantaneous single-channel sound intensity signal
wherein the length of the signal in the dynamic window ', is as short as possible provided that the signal comprises a sufficient information amount; and
404: returning to 402, updating a start point t of a fixed window j> basedona
signal overlap rate q, and performing a cyclic operation.
5. The sound source position estimation method according to claim 4, wherein in step 402, synchronous sliding is performed in channel signals through a time window of a fixed size, signals are extracted, and the information is fused into the instantaneous single-channel sound intensity signal by using a cross-spectrum method; and the step specifically comprises the following steps:
1t 501: specifying a fixed window length J and a window start point t according to a signal fusion degree;
502: capturing signal segments of a window sizeIf by using the same start pointtl;
in the signal channels of the sound pressure P and the acoustic particle velocitiesvr W W and z in the axial directions, wherein corresponding window signals are A ,
W"Y, and W-,; and
503: calculating, based on the cross-spectrum method, the instantaneous w single-channel sound intensity signal 6 obtained after the window signals are fused, to implement multi-sensor information fusion, wherein a calculation formula of the instantaneous single-channel sound intensity signal obtained after the fusion is:
Re[S ] Re[S, ] Re[S ] I~p(t).v,(t) "2 +p(t).v,(t) *(1)+p(t).v,(t) Re[S (f)] +Re[S (f)
" Re[S (f)]
5 wherein SP , SP , and 'r are respectively cross-spectrum functions of three
components x, y, and z, S(f)is a spectrum function of P , f is a frequency, Re[ I indicates Laplace transform, 0 and 9 are respectively a pitch direction of arrival (DOA) and an azimuth DOA of a sound source relative to a vector hydrophone, wherein the pitch DOA and the azimuth DOA respectively take the XOY plane and the x
-axis as 0°, and p(t) v (t) v,(t), and v(t) are respectively a sound pressure signal and acoustic particle velocity signals in different directions that are received by the vector hydrophone at a moment t.
6. The sound source position estimation method according to claim 4, wherein in step 403, the dynamic windows of all lengths are traversed for the instantaneous single-channel
sound intensity signal ', in the fixed window, and the steepest ascent section of the
Shannon entropy is found as the optimum dynamic window U; and the step specifically comprises the following steps:
601: traversing the entire fixed window 'I of the captured instantaneous
single-channel sound intensity signal I starting from the start point of the fixed window, and calculating a Shannon entropy of signals of all lengths by using the following
formula, to construct a Shannon entropy signal 6,
Shannon(X) =p(x). log = -p(x). logp(x) p(x1 ) i I
wherein i is a possible value of a random event X, Shannon(X) is a Shannon entropy comprised in the random event X, m is a total quantity of random events, and
p(x,) is a probability of occurrence of* ;
602: finding a steepest ascent section of SI, according to a derivation result si of
marking the length as , andjumping to step 604;
603: marking the length as e if no steepest ascent section is found in S6, W considering that the signal in the fixed window ', is an invalid signal or a noise signal, wherein the following two conditions are satisfied:
when S' is relatively small, it is considered that the signal is a null signal or
Shannon entropy content of the signal in the fixed window is insufficient, and =/0
wherein 1 is a preset minimum captured length; and
when S is relatively large, it is considered that the signal is a noise signal or a valid
signal comprising a relatively high Shannon entropy, and s 1, wherein 1 is a preset maximum captured length;
604: capturing a signal segment whose length is is starting from the start point t
in the fixed window 6 as the dynamic window ,and marking an end time of the
window as
7. The sound source position estimation method according to claim 1, wherein
signal self-supplement is performed by using the expectation maximization algorithm, segmented unequal-length signals are equivalent to observation data X, supplemented equal-length signals are equivalent to complete data Y, a supplementary signal is equivalent to unobserved data Z, and a result is iterated by using the expectation maximization algorithm to obtain a maximum value 0* ofaparameter , thatis,whena maximum value of a maximum likelihood function L(0) based on Y is obtained, 2 optimal solutions of a mean u and a variance 2 of a complete dataset are obtained, and an unknown dataset Z is obtained based on an observed dataset X, to supplement the complete dataset Y, which specifically comprises the following steps:
701: assuming that an iteration quantity t = 0, initializing a parameter vector 0(0), 0
being a parameter vector formed by the mean and the variance of the dataset Y, and
calculating an initial maximum likelihood function LO) (0)
L())= L( xl, --- ,xn;))= p( x ;O)..... p( xn;))=$ p( x,;0
) 702: obtainingQ(z from 0(t), and ensuring that when 0 is specified, the
equal sign in In(E(X))E[In(X)]is satisfied, to establish a lower bound of ;
703: fixing Qt)(~)andtaking 0( as a variable, taking a derivative ofL ) in
step 702, and obtaining O(t+1) according to the formula
L (60'0 ?) InQ(zN),?) ln ) =L(6t)) P); and
704: if ) , ending the iterative calculation; otherwise, assuming
that t = t +1, and returning to step 702, wherein a threshold - is a specified very small value; wherein Qi represents a distribution of unknown data Z , p () z W; ) is a probability of occurrence of x '),z' under a condition 0(), the superscript i indicates an ith value of a corresponding parameter, - is the threshold and is an initially specified small value used as a standard for ending the iteration, and E[ ] is a mathematical expectation.
8. The sound source position estimation method according to claim 1, wherein the estimating, through a recurrent neural network, a position of the sound source by using equal-length signals obtained after the self-supplement specifically comprises: using the supplemented signal segments as an input by using the expectation maximization algorithm, outputting azimuth DOAs and distances of the sound source in different signal segments, and performing cross-validation on estimation results of the different signal segments, to implement accurate positioning of the position of the sound source.
9. A computer-readable storage medium, storing a computer program, wherein when the program is executed by a processor, steps in the sound source position estimation method according to any one of claims 1 to 8 are implemented.
10. A computer device, comprising a memory, a processor, and a computer program stored in the memory and capable of being run on the processor, wherein when the processor executes the program, steps in the sound source position estimation method according to any one of claims 1 to 8 are implemented.
Applications Claiming Priority (3)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201910493374.4A CN110146846B (en) | 2019-06-06 | 2019-06-06 | Sound source position estimation method, readable storage medium and computer equipment |
CN201910493374.4 | 2019-06-06 | ||
PCT/CN2020/089894 WO2020244359A1 (en) | 2019-06-06 | 2020-05-13 | Sound source position estimation method, readable storage medium and computer device |
Publications (2)
Publication Number | Publication Date |
---|---|
AU2020287677A1 AU2020287677A1 (en) | 2021-03-04 |
AU2020287677B2 true AU2020287677B2 (en) | 2022-07-07 |
Family
ID=67590874
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
AU2020287677A Ceased AU2020287677B2 (en) | 2019-06-06 | 2020-05-13 | Sound source position estimation method, readable storage medium and computer device |
Country Status (3)
Country | Link |
---|---|
CN (1) | CN110146846B (en) |
AU (1) | AU2020287677B2 (en) |
WO (1) | WO2020244359A1 (en) |
Families Citing this family (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110146846B (en) * | 2019-06-06 | 2021-04-13 | 青岛理工大学 | Sound source position estimation method, readable storage medium and computer equipment |
CN111427008B (en) * | 2020-04-19 | 2022-09-16 | 中国电子科技集团公司电子科学研究院 | Position determination method for underwater autonomous vehicle cluster and related device |
CN111580048B (en) * | 2020-05-09 | 2020-12-29 | 中国科学院声学研究所 | Broadband sound source depth estimation method using single-vector hydrophone |
CN112415467B (en) * | 2020-11-06 | 2022-10-25 | 中国海洋大学 | Single-vector subsurface buoy target positioning implementation method based on neural network |
CN113009417B (en) * | 2021-02-05 | 2022-09-20 | 中国人民解放军国防科技大学 | Submarine acoustic array formation estimation method utilizing sound field interference characteristics |
CN113009419B (en) * | 2021-02-25 | 2021-11-09 | 中国科学院声学研究所 | Target depth estimation method based on frequency domain cross-correlation matching |
CN114910901B (en) * | 2022-05-30 | 2024-07-12 | 深圳市大族机器人有限公司 | High-precision multi-sensor fusion ranging system of cooperative robot |
CN116068494B (en) * | 2023-03-31 | 2023-07-18 | 中国人民解放军96901部队 | Wide-area secondary sound source positioning method based on secondary sound propagation model |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170353789A1 (en) * | 2016-06-01 | 2017-12-07 | Google Inc. | Sound source estimation using neural networks |
Family Cites Families (13)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US6804167B2 (en) * | 2003-02-25 | 2004-10-12 | Lockheed Martin Corporation | Bi-directional temporal correlation SONAR |
CN101539597B (en) * | 2009-04-29 | 2011-01-26 | 哈尔滨工程大学 | Multi-target method for distinguishing radiation noises having same frequency band |
CN202383285U (en) * | 2010-12-30 | 2012-08-15 | 中国科学院声学研究所 | Networked underwater sound positioning node system |
CN102708374B (en) * | 2012-01-06 | 2014-06-25 | 香港理工大学 | Method and device for classifying remote images by integrating edge information and support vector machine |
CN102981146B (en) * | 2012-11-19 | 2014-10-22 | 哈尔滨工程大学 | Passive positioning method of single-vector hydrophone |
CN103267966B (en) * | 2013-05-20 | 2014-12-31 | 西北工业大学 | Orientation estimation method for single two-dimensional differential pressure type vector hydrophone |
CN103927401B (en) * | 2014-04-16 | 2018-02-16 | 公安部第一研究所 | The high-speed signal acquisition and processing system of a kind of multi-channel sampling |
CN103954346B (en) * | 2014-05-16 | 2016-06-22 | 中国电子科技集团公司第四十九研究所 | There is the magnetic complex three-dimensional vector hydrophone of target location and discrimination function and the target location of this hydrophone and method of discrimination |
CN104848900B (en) * | 2015-05-26 | 2017-03-22 | 国家海洋技术中心 | Array-type ocean acoustical signal measurement system |
CN106249244B (en) * | 2016-07-07 | 2018-10-26 | 哈尔滨工程大学 | A kind of acoustic vector circle battle array wide-band coherent signal source direction estimation method |
CN108337002A (en) * | 2018-02-09 | 2018-07-27 | 黄淮学院 | A kind of multiple channel acousto signal reconfiguring method |
CN109492580B (en) * | 2018-11-08 | 2020-08-07 | 北方工业大学 | Multi-size aerial image positioning method based on neighborhood significance reference of full convolution network |
CN110146846B (en) * | 2019-06-06 | 2021-04-13 | 青岛理工大学 | Sound source position estimation method, readable storage medium and computer equipment |
-
2019
- 2019-06-06 CN CN201910493374.4A patent/CN110146846B/en active Active
-
2020
- 2020-05-13 WO PCT/CN2020/089894 patent/WO2020244359A1/en active Application Filing
- 2020-05-13 AU AU2020287677A patent/AU2020287677B2/en not_active Ceased
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20170353789A1 (en) * | 2016-06-01 | 2017-12-07 | Google Inc. | Sound source estimation using neural networks |
Non-Patent Citations (1)
Title |
---|
HUANG, Z. et al., 'A deep Neural network based method of source localization in a shallow water environment', 2018 International Conference on Acoustics, Speech and Signal Processing, Calgary, Canada, 15-20 April 2018, pp. 3499-3503. * |
Also Published As
Publication number | Publication date |
---|---|
CN110146846A (en) | 2019-08-20 |
WO2020244359A1 (en) | 2020-12-10 |
AU2020287677A1 (en) | 2021-03-04 |
CN110146846B (en) | 2021-04-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
AU2020287677B2 (en) | Sound source position estimation method, readable storage medium and computer device | |
CN104076355B (en) | Tracking before Dim targets detection in strong clutter environment based on dynamic programming | |
Vo et al. | Multi-sensor joint detection and tracking with the Bernoulli filter | |
US9099096B2 (en) | Source separation by independent component analysis with moving constraint | |
CN103439688B (en) | Sound source positioning system and method used for distributed microphone arrays | |
US20200342890A1 (en) | Target voice detection method and apparatus | |
CN103901395B (en) | Coherent signal direction of arrival Dynamic Tracking under a kind of impulsive noise environment | |
CN109272989A (en) | Voice awakening method, device and computer readable storage medium | |
CN105068048A (en) | Distributed microphone array sound source positioning method based on space sparsity | |
Ravela | Quantifying uncertainty for coherent structures | |
CN109116293A (en) | A kind of Wave arrival direction estimating method based on sparse Bayesian out of place | |
CN113640768B (en) | Low-resolution radar target identification method based on wavelet transformation | |
Feng et al. | DAMUN: A domain adaptive human activity recognition network based on multimodal feature fusion | |
Yang et al. | Deep-learning-based mask-cut method for InSAR phase unwrapping | |
Zhong et al. | Particle filtering for 2-D direction of arrival tracking using an acoustic vector sensor | |
Liu et al. | Virtual-network-reconstruction-based underwater acoustic localization framework: Theory and experiment validation | |
He et al. | Clustering stochastic weather scenarios using influence model-based distance measures | |
Zhong et al. | Particle filtering with enhanced likelihood model for underwater acoustic source DOA tracking | |
CN113421281A (en) | Pedestrian micromotion part separation method based on segmentation theory | |
Jun et al. | Research on non-stationary blind separation method for unfixed signal based on extended joint diagonalization | |
Zhang et al. | Underdetermined mixing matrix estimation based on single source detection | |
Sun et al. | Improved method for drone sound event detection system aiming at the impact of background noise and angle deviation | |
Lu et al. | EM-Based TDOA Estimation of a Speech Source via Gaussian Mixture Models in Noisy and Anechoic Environments | |
Sun et al. | Method for Estimating the Location of A Low-frequency Target in A Shallow Sea Based on A Single Vector Hydrophone | |
Guo et al. | Indoor UWB positioning technology research based on the NLOS identification |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
FGA | Letters patent sealed or granted (standard patent) | ||
MK14 | Patent ceased section 143(a) (annual fees not paid) or expired |