CN115903022A - Deep learning chip suitable for real-time seismic data processing - Google Patents
Deep learning chip suitable for real-time seismic data processing Download PDFInfo
- Publication number
- CN115903022A CN115903022A CN202211556940.XA CN202211556940A CN115903022A CN 115903022 A CN115903022 A CN 115903022A CN 202211556940 A CN202211556940 A CN 202211556940A CN 115903022 A CN115903022 A CN 115903022A
- Authority
- CN
- China
- Prior art keywords
- wave
- unit
- layer
- input end
- subsystem
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000013135 deep learning Methods 0.000 title claims abstract description 17
- 230000006698 induction Effects 0.000 claims abstract description 69
- 238000000605 extraction Methods 0.000 claims abstract description 34
- 238000001514 detection method Methods 0.000 claims abstract description 11
- 238000000034 method Methods 0.000 claims abstract description 6
- 238000011176 pooling Methods 0.000 claims description 45
- 230000006870 function Effects 0.000 claims description 26
- 230000004913 activation Effects 0.000 claims description 5
- 230000008569 process Effects 0.000 claims description 3
- 230000033228 biological regulation Effects 0.000 claims 1
- 238000013527 convolutional neural network Methods 0.000 description 13
- 230000008859 change Effects 0.000 description 8
- 230000009467 reduction Effects 0.000 description 4
- 230000009286 beneficial effect Effects 0.000 description 3
- 238000010586 diagram Methods 0.000 description 3
- 239000000284 extract Substances 0.000 description 3
- 238000005065 mining Methods 0.000 description 3
- 239000011435 rock Substances 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 2
- 238000013528 artificial neural network Methods 0.000 description 2
- 230000001276 controlling effect Effects 0.000 description 2
- 230000000694 effects Effects 0.000 description 2
- 230000001105 regulatory effect Effects 0.000 description 2
- 230000003213 activating effect Effects 0.000 description 1
- 238000009412 basement excavation Methods 0.000 description 1
- 239000002775 capsule Substances 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
Images
Classifications
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/30—Assessment of water resources
Landscapes
- Management, Administration, Business Operations System, And Electronic Commerce (AREA)
Abstract
The invention discloses a deep learning chip suitable for real-time seismic data processing, which comprises: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem; the characteristic extraction subsystem is used for extracting microseism detection data to obtain microseism characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the micro seismic characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time; the method solves the problem that the seismic source distance of the micro earthquake can not be measured only by extracting P waves or P wave first arrivals in the prior art.
Description
Technical Field
The invention relates to the technical field of seismic induction, in particular to a deep learning chip suitable for real-time seismic data processing.
Background
A micro-earthquake is a small earthquake. Rock fracture and seismic activity, often unavoidable phenomena, occur during deep mining in underground mines. The seismic activity induced by mining, micro-earthquakes are generally defined as those earthquakes caused by rock failure due to changes in stress fields within the rock mass in the vicinity of the mining excavation.
The existing documents of 'a method and a system for picking up a microseism P wave first arrival based on a capsule neural network' and 'a method and a system for identifying a microseism P wave based on a deep convolution neural network' are to identify the P wave or extract the P wave first arrival, but only extract the P wave or the P wave first arrival cannot measure the seismic source distance of the microseism.
Disclosure of Invention
Aiming at the defects in the prior art, the deep learning chip suitable for real-time seismic data processing provided by the invention solves the problem that the seismic source distance of a micro earthquake cannot be measured only by extracting P waves or P wave first arrivals in the prior art.
In order to achieve the purpose of the invention, the invention adopts the technical scheme that: a deep learning chip adapted for real-time seismic data processing, comprising: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting micro-seismic detection data to obtain micro-seismic characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time.
Further, the feature extraction subsystem includes: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the characteristic extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
Further, the CNN unit includes: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
Further, the P-wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilSTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P-wave first arrival induction subsystem.
Further, the S-wave first-arrival induction subsystem comprises: a fourth BilSTM unit, a third global attention unit and a second full-connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
Further, the input and output relationship of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P-wave first arrival induction subsystem or the S-wave first arrival induction subsystem is as follows:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )⊙h t )
y t =σ[W o ·(y t-1 ,x t ,C t-1 ,C t )+b o ]⊙tanh[C t ]
wherein f is t The output of the forgetting gate at time t, σ [, ]]For sigmoid activation functions, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Output of cells at time t-1, x t For input of cells at time t, C t-1 The state of the cells at time t-1, i t Input the output of the gate at time t, tanh 2]As hyperbolic tangent activation function, W i As the weight of the input gate, b i For input of the offset of the gate, h t Output of candidate gate at time t, W h As a weight of the candidate gate, b h As a candidate for the offset of the gate, C t The state of the cell at time t, < is a Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
The beneficial effects of the above further scheme are: taking into account the state C of the last moment in the LSTM module t-1 Input x at the current time t Output y at the previous time t-1 The relationship between the state, input and output is taken into full account when calculating the cell.
Further, the global attention unit in the feature extraction subsystem, the P-wave first-motion induction subsystem or the S-wave first-motion induction subsystem includes: a sixth convolutional layer, a Softmax layer, a multiplier, a seventh convolutional layer, a ReLU layer, an eighth convolutional layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
Further, the characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem adopt microseism detection data and data labels to form a training data set, the training data set is adopted to train the characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem to obtain the trained characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem, and the trained characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem are arranged in the processor.
Further, the weight updating formula of the training process is as follows:
wherein, w i+1 Is the weight of the i +1 th iteration, w i Is the weight, η, of the ith iteration i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 Loss for the i-1 th iterationThe loss function, γ is the proportionality coefficient and ζ is the tuning constant.
The beneficial effects of the above further scheme are: the weighting formula of the past second derivative and the current second derivative of the loss function is designed, the larger the second derivative of the loss function is, the larger the change rate of the gradient of the loss function is, the weight is used for regulating and controlling the weight after the change rate of the loss function and the past gradient change rate are weighted and accumulated to be smooth and filtered, so that the step length of weight iteration is hooked with the gradient change rate, the overshoot of the weight iteration is prevented, and the slowness of the weight iteration is avoided; in the design of the learning rate parameter, the degree of the loss function reduction is considered, and when the degree of the loss function reduction is large, J i-1 -J i Is large, so that the learning rate eta is large i The variable speed is increased, the step length adjustment force of weight updating iteration is increased, the reduction degree of the loss function is smaller, J i-1 -J i Is small, so that the learning rate eta is small i The variable speed is reduced, the step length adjustment strength of the weight updating iteration is reduced, and finally the weight can be adaptive to the rapid stable iteration and reach the optimal value rapidly.
In conclusion, the beneficial effects of the invention are as follows: the microseism characteristic data in the microseism detection data are extracted through the characteristic extraction subsystem, on one hand, the quantity is reduced, on the other hand, the data characteristics are reserved, the P wave first arrival time and the S wave first arrival time are extracted through the P wave first arrival induction subsystem and the S wave first arrival induction subsystem respectively, and the microseism seismic source position is obtained through estimation of the microseism estimation subsystem.
Drawings
FIG. 1 is a system diagram of a deep learning chip suitable for real-time seismic data processing;
FIG. 2 is a system block diagram of a CNN unit;
FIG. 3 is a system block diagram of a global attention unit.
Detailed Description
The following description of the embodiments of the present invention is provided to facilitate the understanding of the present invention by those skilled in the art, but it should be understood that the present invention is not limited to the scope of the embodiments, and it will be apparent to those skilled in the art that various changes may be made without departing from the spirit and scope of the invention as defined and defined in the appended claims, and all matters produced by the invention using the inventive concept are protected.
As shown in fig. 1, a deep learning chip suitable for real-time seismic data processing includes: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting microseism detection data to obtain microseism characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the micro-seismic estimation subsystem is used for estimating a micro-seismic source according to the P wave first arrival time and the S wave first arrival time.
The feature extraction subsystem includes: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the feature extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
As shown in fig. 2, the CNN unit includes: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
The P wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P-wave first arrival induction subsystem.
The S-wave first arrival induction subsystem comprises: a fourth BilSt unit, a third global attention unit and a second full connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
The input and output relations of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P wave first arrival induction subsystem or the S wave first arrival induction subsystem are as follows:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )Qh t )
y t =σ[W o ·(y t-1 ,x t ,C t-1 ,C t )+b o ]⊙tanh[C t ]
wherein f is t The output of the forgetting gate at time t, σ [ ]]Activating a function for sigmoid, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Is the output of the cell at time t-1, x t Input of cells at time t, C t-1 The state of the cells at time t-1, i t The output of the input gate at time t, tanh [ 2 ]]Is a hyperbolic tangent activation function, W i As the weight of the input gate, b i For the biasing of the input gate, h t Output of candidate gate at time t, W h As the weight of the candidate door, b h As a candidate for the offset of the gate, C t The state of the cell at time t, <' > is Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
Taking into account the state C of the last moment in the LSTM module t-1 Input x at the current time t Output y of the previous time t-1 The cell is operated with the relationship among the state, input, and output taken into full account.
As shown in fig. 3, the global attention unit in the feature extraction subsystem, the P-wave first-motion induction subsystem or the S-wave first-motion induction subsystem includes: a sixth convolution layer, a Softmax layer, a multiplier, a seventh convolution layer, a ReLU layer, an eighth convolution layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
The characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem are arranged in the processor.
The data tags are P-wave first arrival and S-wave first arrival times.
The weight updating formula of the training process is as follows:
wherein, w i+1 Is the weight of the i +1 th iteration, w i Is the weight, η, of the ith iteration i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 And gamma is a proportionality coefficient and zeta is an adjusting constant for the loss function of the i-1 iteration.
The weighting formula of the past second derivative and the current second derivative of the loss function is designed, the larger the second derivative of the loss function is, the larger the change rate of the gradient of the loss function is, the weight is used for regulating and controlling the weight after the change rate of the loss function and the change rate of the past gradient are weighted and accumulated to be smooth and filtered, so that the step length of weight iteration is hooked with the change rate of the gradient, the overshoot of the weight iteration is prevented, and the slow of the weight iteration is avoided; in the design of the learning rate parameter, the degree of the loss function is considered, and when the degree of the loss function is largerWhen, J i-1 -J i Is large, so that the learning rate eta is large i The variable speed is increased, the step length adjustment strength of weight updating iteration is increased, the reduction degree of the loss function is smaller, J i-1 -J i Is small, so that the learning rate eta is small i The variable speed is reduced, the step length adjustment strength of the weight updating iteration is reduced, and finally the weight can be adaptive to the rapid stable iteration and reach the optimal value rapidly.
The invention extracts the micro-seismic characteristic data in the micro-seismic detection data through the characteristic extraction subsystem, on one hand, the quantity is reduced, on the other hand, the data characteristics are kept, and then the P-wave first-arrival time and the S-wave first-arrival time are extracted through the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem respectively, the speed of the P-wave and the speed of the S-wave can be measured through the sensor, and the micro-seismic source position can be estimated through the time difference between the P-wave first-arrival time and the S-wave first-arrival time and the speed of the P-wave and the S-wave.
Claims (9)
1. A deep learning chip adapted for real-time seismic data processing, comprising: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting micro-seismic detection data to obtain micro-seismic characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time.
2. The deep learning chip suitable for real-time seismic data processing according to claim 1, wherein the feature extraction subsystem comprises: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the characteristic extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
3. The deep learning chip suitable for real-time seismic data processing according to claim 2, wherein the CNN unit comprises: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
4. The deep learning chip for real-time seismic data processing according to claim 1, wherein the P-wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilSTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P wave first arrival induction subsystem.
5. The deep learning chip for real-time seismic data processing according to claim 1, wherein the S-wave first arrival induction subsystem comprises: a fourth BilSt unit, a third global attention unit and a second full connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
6. The deep learning chip for real-time seismic data processing according to any one of claims 2, 4, and 5, wherein the input/output relationship of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P-wave first arrival induction subsystem, or the S-wave first arrival induction subsystem is:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )⊙h t )
y t =[W o ·(y t-1 ,x t ,C t-1 ,C t )+ o ]⊙h[ t ]
wherein f is t The output of the forgetting gate at time t, σ [ ]]For sigmoid activation functions, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Is the output of the cell at time t-1, x t Input of cells at time t, C t-1 The state of the cells at time t-1, i t Input the output of the gate at time t, tanh 2]As hyperbolic tangent activation function, W i As the weight of the input gate, b i For the biasing of the input gate, h t For the output of the candidate gate at time t, W h As the weight of the candidate door, b h As a candidate for the offset of the gate, C t The state of the cell at time t, <' > is Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
7. The deep learning chip suitable for real-time seismic data processing according to any one of claims 2, 4, and 5, wherein the global attention unit in the feature extraction subsystem, the P-wave first-arrival induction subsystem, or the S-wave first-arrival induction subsystem comprises: a sixth convolution layer, a Softmax layer, a multiplier, a seventh convolution layer, a ReLU layer, an eighth convolution layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
8. The deep learning chip for real-time seismic data processing according to claim 1, wherein the feature extraction subsystem, the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem use micro-seismic detection data and data labels to form a training data set, the feature extraction subsystem, the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem are trained by using the training data set to obtain a trained feature extraction subsystem, a trained P-wave first-arrival induction subsystem and a trained S-wave first-arrival induction subsystem, and the trained feature extraction subsystem, the trained P-wave first-arrival induction subsystem and the trained S-wave first-arrival induction subsystem are arranged in the processor.
9. The deep learning chip for real-time seismic data processing according to claim 8, wherein the weight update formula of the training process is:
wherein, w i+1 Weight of the i +1 th iteration, w i Is the weight of the ith iteration, η i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 And gamma is a proportionality coefficient and zeta is a regulation constant for the loss function of the i-1 th iteration.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211556940.XA CN115903022B (en) | 2022-12-06 | 2022-12-06 | Deep learning chip suitable for real-time seismic data processing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202211556940.XA CN115903022B (en) | 2022-12-06 | 2022-12-06 | Deep learning chip suitable for real-time seismic data processing |
Publications (2)
Publication Number | Publication Date |
---|---|
CN115903022A true CN115903022A (en) | 2023-04-04 |
CN115903022B CN115903022B (en) | 2023-10-31 |
Family
ID=86489479
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202211556940.XA Active CN115903022B (en) | 2022-12-06 | 2022-12-06 | Deep learning chip suitable for real-time seismic data processing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN115903022B (en) |
Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112017289A (en) * | 2020-08-31 | 2020-12-01 | 电子科技大学 | Well-seismic combined initial lithology model construction method based on deep learning |
CN112068195A (en) * | 2019-06-10 | 2020-12-11 | 中国石油化工股份有限公司 | Automatic first arrival picking method for microseism P & S wave matching event and computer storage medium |
US20210117737A1 (en) * | 2019-10-18 | 2021-04-22 | Korea University Research And Business Foundation | Earthquake event classification method using attention-based convolutional neural network, recording medium and device for performing the method |
CN113158792A (en) * | 2021-03-15 | 2021-07-23 | 辽宁大学 | Microseismic event identification method based on improved model transfer learning |
CN114660656A (en) * | 2022-03-17 | 2022-06-24 | 中国科学院地质与地球物理研究所 | Seismic data first arrival picking method and system |
-
2022
- 2022-12-06 CN CN202211556940.XA patent/CN115903022B/en active Active
Patent Citations (5)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112068195A (en) * | 2019-06-10 | 2020-12-11 | 中国石油化工股份有限公司 | Automatic first arrival picking method for microseism P & S wave matching event and computer storage medium |
US20210117737A1 (en) * | 2019-10-18 | 2021-04-22 | Korea University Research And Business Foundation | Earthquake event classification method using attention-based convolutional neural network, recording medium and device for performing the method |
CN112017289A (en) * | 2020-08-31 | 2020-12-01 | 电子科技大学 | Well-seismic combined initial lithology model construction method based on deep learning |
CN113158792A (en) * | 2021-03-15 | 2021-07-23 | 辽宁大学 | Microseismic event identification method based on improved model transfer learning |
CN114660656A (en) * | 2022-03-17 | 2022-06-24 | 中国科学院地质与地球物理研究所 | Seismic data first arrival picking method and system |
Also Published As
Publication number | Publication date |
---|---|
CN115903022B (en) | 2023-10-31 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111723329B (en) | Seismic phase feature recognition waveform inversion method based on full convolution neural network | |
CN107544904B (en) | Software reliability prediction method based on deep CG-LSTM neural network | |
CN108596327B (en) | Seismic velocity spectrum artificial intelligence picking method based on deep learning | |
CN101794528B (en) | Gesture language-voice bidirectional translation system | |
CN106646587B (en) | Object detection and recognition method and system based on acoustic vibration signal | |
CN111144542A (en) | Oil well productivity prediction method, device and equipment | |
CN108830421B (en) | Method and device for predicting gas content of tight sandstone reservoir | |
CN111580151B (en) | SSNet model-based earthquake event time-of-arrival identification method | |
CN108875836B (en) | Simple-complex activity collaborative recognition method based on deep multitask learning | |
CN110379418A (en) | A kind of voice confrontation sample generating method | |
CN106384122A (en) | Device fault mode identification method based on improved CS-LSSVM | |
CN111695413A (en) | Signal first arrival pickup method and device combining U-Net and Temporal encoding | |
CN112052551B (en) | Fan surge operation fault identification method and system | |
CN110223342B (en) | Space target size estimation method based on deep neural network | |
CN113156492B (en) | Real-time intelligent early warning method applied to TBM tunnel rockburst disasters | |
CN114818579A (en) | Analog circuit fault diagnosis method based on one-dimensional convolution long-short term memory network | |
CN114862015A (en) | Typhoon wind speed intelligent prediction method based on FEN-ConvLSTM model | |
Xu et al. | An automatic P-wave onset time picking method for mining-induced microseismic data based on long short-term memory deep neural network | |
CN115903022A (en) | Deep learning chip suitable for real-time seismic data processing | |
CN112132096B (en) | Behavior modal identification method of random configuration network for dynamically updating output weight | |
CN106842172A (en) | A kind of submarine target structural sparse feature extracting method | |
CN115424351A (en) | Dynamic image identification method based on space-time fusion attention impulse neural network | |
CN112433249A (en) | Horizon tracking method and device, computer equipment and computer readable storage medium | |
CN111487679B (en) | Transverse wave velocity prediction method, device and equipment | |
CN116561528B (en) | RUL prediction method of rotary machine |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |