CN115903022A - Deep learning chip suitable for real-time seismic data processing - Google Patents

Deep learning chip suitable for real-time seismic data processing Download PDF

Info

Publication number
CN115903022A
CN115903022A CN202211556940.XA CN202211556940A CN115903022A CN 115903022 A CN115903022 A CN 115903022A CN 202211556940 A CN202211556940 A CN 202211556940A CN 115903022 A CN115903022 A CN 115903022A
Authority
CN
China
Prior art keywords
wave
unit
layer
input end
subsystem
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202211556940.XA
Other languages
Chinese (zh)
Other versions
CN115903022B (en
Inventor
薛清峰
王一博
郑忆康
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Institute of Geology and Geophysics of CAS
Original Assignee
Institute of Geology and Geophysics of CAS
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Institute of Geology and Geophysics of CAS filed Critical Institute of Geology and Geophysics of CAS
Priority to CN202211556940.XA priority Critical patent/CN115903022B/en
Publication of CN115903022A publication Critical patent/CN115903022A/en
Application granted granted Critical
Publication of CN115903022B publication Critical patent/CN115903022B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/30Assessment of water resources

Landscapes

  • Management, Administration, Business Operations System, And Electronic Commerce (AREA)

Abstract

The invention discloses a deep learning chip suitable for real-time seismic data processing, which comprises: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem; the characteristic extraction subsystem is used for extracting microseism detection data to obtain microseism characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the micro seismic characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time; the method solves the problem that the seismic source distance of the micro earthquake can not be measured only by extracting P waves or P wave first arrivals in the prior art.

Description

Deep learning chip suitable for real-time seismic data processing
Technical Field
The invention relates to the technical field of seismic induction, in particular to a deep learning chip suitable for real-time seismic data processing.
Background
A micro-earthquake is a small earthquake. Rock fracture and seismic activity, often unavoidable phenomena, occur during deep mining in underground mines. The seismic activity induced by mining, micro-earthquakes are generally defined as those earthquakes caused by rock failure due to changes in stress fields within the rock mass in the vicinity of the mining excavation.
The existing documents of 'a method and a system for picking up a microseism P wave first arrival based on a capsule neural network' and 'a method and a system for identifying a microseism P wave based on a deep convolution neural network' are to identify the P wave or extract the P wave first arrival, but only extract the P wave or the P wave first arrival cannot measure the seismic source distance of the microseism.
Disclosure of Invention
Aiming at the defects in the prior art, the deep learning chip suitable for real-time seismic data processing provided by the invention solves the problem that the seismic source distance of a micro earthquake cannot be measured only by extracting P waves or P wave first arrivals in the prior art.
In order to achieve the purpose of the invention, the invention adopts the technical scheme that: a deep learning chip adapted for real-time seismic data processing, comprising: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting micro-seismic detection data to obtain micro-seismic characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time.
Further, the feature extraction subsystem includes: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the characteristic extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
Further, the CNN unit includes: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
Further, the P-wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilSTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P-wave first arrival induction subsystem.
Further, the S-wave first-arrival induction subsystem comprises: a fourth BilSTM unit, a third global attention unit and a second full-connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
Further, the input and output relationship of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P-wave first arrival induction subsystem or the S-wave first arrival induction subsystem is as follows:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )⊙h t )
y t =σ[W o ·(y t-1 ,x t ,C t-1 ,C t )+b o ]⊙tanh[C t ]
wherein f is t The output of the forgetting gate at time t, σ [, ]]For sigmoid activation functions, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Output of cells at time t-1, x t For input of cells at time t, C t-1 The state of the cells at time t-1, i t Input the output of the gate at time t, tanh 2]As hyperbolic tangent activation function, W i As the weight of the input gate, b i For input of the offset of the gate, h t Output of candidate gate at time t, W h As a weight of the candidate gate, b h As a candidate for the offset of the gate, C t The state of the cell at time t, < is a Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
The beneficial effects of the above further scheme are: taking into account the state C of the last moment in the LSTM module t-1 Input x at the current time t Output y at the previous time t-1 The relationship between the state, input and output is taken into full account when calculating the cell.
Further, the global attention unit in the feature extraction subsystem, the P-wave first-motion induction subsystem or the S-wave first-motion induction subsystem includes: a sixth convolutional layer, a Softmax layer, a multiplier, a seventh convolutional layer, a ReLU layer, an eighth convolutional layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
Further, the characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem adopt microseism detection data and data labels to form a training data set, the training data set is adopted to train the characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem to obtain the trained characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem, and the trained characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem are arranged in the processor.
Further, the weight updating formula of the training process is as follows:
Figure BDA0003983723940000041
Figure BDA0003983723940000042
wherein, w i+1 Is the weight of the i +1 th iteration, w i Is the weight, η, of the ith iteration i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 Loss for the i-1 th iterationThe loss function, γ is the proportionality coefficient and ζ is the tuning constant.
The beneficial effects of the above further scheme are: the weighting formula of the past second derivative and the current second derivative of the loss function is designed, the larger the second derivative of the loss function is, the larger the change rate of the gradient of the loss function is, the weight is used for regulating and controlling the weight after the change rate of the loss function and the past gradient change rate are weighted and accumulated to be smooth and filtered, so that the step length of weight iteration is hooked with the gradient change rate, the overshoot of the weight iteration is prevented, and the slowness of the weight iteration is avoided; in the design of the learning rate parameter, the degree of the loss function reduction is considered, and when the degree of the loss function reduction is large, J i-1 -J i Is large, so that the learning rate eta is large i The variable speed is increased, the step length adjustment force of weight updating iteration is increased, the reduction degree of the loss function is smaller, J i-1 -J i Is small, so that the learning rate eta is small i The variable speed is reduced, the step length adjustment strength of the weight updating iteration is reduced, and finally the weight can be adaptive to the rapid stable iteration and reach the optimal value rapidly.
In conclusion, the beneficial effects of the invention are as follows: the microseism characteristic data in the microseism detection data are extracted through the characteristic extraction subsystem, on one hand, the quantity is reduced, on the other hand, the data characteristics are reserved, the P wave first arrival time and the S wave first arrival time are extracted through the P wave first arrival induction subsystem and the S wave first arrival induction subsystem respectively, and the microseism seismic source position is obtained through estimation of the microseism estimation subsystem.
Drawings
FIG. 1 is a system diagram of a deep learning chip suitable for real-time seismic data processing;
FIG. 2 is a system block diagram of a CNN unit;
FIG. 3 is a system block diagram of a global attention unit.
Detailed Description
The following description of the embodiments of the present invention is provided to facilitate the understanding of the present invention by those skilled in the art, but it should be understood that the present invention is not limited to the scope of the embodiments, and it will be apparent to those skilled in the art that various changes may be made without departing from the spirit and scope of the invention as defined and defined in the appended claims, and all matters produced by the invention using the inventive concept are protected.
As shown in fig. 1, a deep learning chip suitable for real-time seismic data processing includes: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting microseism detection data to obtain microseism characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the micro-seismic estimation subsystem is used for estimating a micro-seismic source according to the P wave first arrival time and the S wave first arrival time.
The feature extraction subsystem includes: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the feature extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
As shown in fig. 2, the CNN unit includes: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
The P wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P-wave first arrival induction subsystem.
The S-wave first arrival induction subsystem comprises: a fourth BilSt unit, a third global attention unit and a second full connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
The input and output relations of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P wave first arrival induction subsystem or the S wave first arrival induction subsystem are as follows:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )Qh t )
y t =σ[W o ·(y t-1 ,x t ,C t-1 ,C t )+b o ]⊙tanh[C t ]
wherein f is t The output of the forgetting gate at time t, σ [ ]]Activating a function for sigmoid, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Is the output of the cell at time t-1, x t Input of cells at time t, C t-1 The state of the cells at time t-1, i t The output of the input gate at time t, tanh [ 2 ]]Is a hyperbolic tangent activation function, W i As the weight of the input gate, b i For the biasing of the input gate, h t Output of candidate gate at time t, W h As the weight of the candidate door, b h As a candidate for the offset of the gate, C t The state of the cell at time t, <' > is Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
Taking into account the state C of the last moment in the LSTM module t-1 Input x at the current time t Output y of the previous time t-1 The cell is operated with the relationship among the state, input, and output taken into full account.
As shown in fig. 3, the global attention unit in the feature extraction subsystem, the P-wave first-motion induction subsystem or the S-wave first-motion induction subsystem includes: a sixth convolution layer, a Softmax layer, a multiplier, a seventh convolution layer, a ReLU layer, an eighth convolution layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
The characteristic extraction subsystem, the P wave first arrival induction subsystem and the S wave first arrival induction subsystem are arranged in the processor.
The data tags are P-wave first arrival and S-wave first arrival times.
The weight updating formula of the training process is as follows:
Figure BDA0003983723940000081
Figure BDA0003983723940000082
wherein, w i+1 Is the weight of the i +1 th iteration, w i Is the weight, η, of the ith iteration i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 And gamma is a proportionality coefficient and zeta is an adjusting constant for the loss function of the i-1 iteration.
The weighting formula of the past second derivative and the current second derivative of the loss function is designed, the larger the second derivative of the loss function is, the larger the change rate of the gradient of the loss function is, the weight is used for regulating and controlling the weight after the change rate of the loss function and the change rate of the past gradient are weighted and accumulated to be smooth and filtered, so that the step length of weight iteration is hooked with the change rate of the gradient, the overshoot of the weight iteration is prevented, and the slow of the weight iteration is avoided; in the design of the learning rate parameter, the degree of the loss function is considered, and when the degree of the loss function is largerWhen, J i-1 -J i Is large, so that the learning rate eta is large i The variable speed is increased, the step length adjustment strength of weight updating iteration is increased, the reduction degree of the loss function is smaller, J i-1 -J i Is small, so that the learning rate eta is small i The variable speed is reduced, the step length adjustment strength of the weight updating iteration is reduced, and finally the weight can be adaptive to the rapid stable iteration and reach the optimal value rapidly.
The invention extracts the micro-seismic characteristic data in the micro-seismic detection data through the characteristic extraction subsystem, on one hand, the quantity is reduced, on the other hand, the data characteristics are kept, and then the P-wave first-arrival time and the S-wave first-arrival time are extracted through the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem respectively, the speed of the P-wave and the speed of the S-wave can be measured through the sensor, and the micro-seismic source position can be estimated through the time difference between the P-wave first-arrival time and the S-wave first-arrival time and the speed of the P-wave and the S-wave.

Claims (9)

1. A deep learning chip adapted for real-time seismic data processing, comprising: the system comprises a characteristic extraction subsystem, a P wave first arrival induction subsystem, an S wave first arrival induction subsystem and a microseism estimation subsystem;
the characteristic extraction subsystem is used for extracting micro-seismic detection data to obtain micro-seismic characteristic data; the P wave first arrival induction subsystem is used for extracting P wave first arrival time in the microseism characteristic data; the S wave first arrival induction subsystem is used for extracting S wave first arrival time in the microseism characteristic data; the microseism estimation subsystem is used for estimating a microseism seismic source according to P wave first arrival time and S wave first arrival time.
2. The deep learning chip suitable for real-time seismic data processing according to claim 1, wherein the feature extraction subsystem comprises: a CNN unit, a first BilSTM unit, a second BilSTM unit and a first global attention unit;
the input end of the CNN unit is used as the input end of the characteristic extraction subsystem and used for inputting microseism detection data; the input end of the first BilSTM unit is connected with the output end of the CNN unit, and the output end of the first BilSTM unit is connected with the input end of the second BilSTM unit; and the input end of the first global attention unit is connected with the output end of the second BilSTM unit, and the output end of the first global attention unit is used as the output end of the feature extraction subsystem.
3. The deep learning chip suitable for real-time seismic data processing according to claim 2, wherein the CNN unit comprises: a first convolution layer, a first maximum pooling layer, a second convolution layer, a second maximum pooling layer, a third convolution layer, a third maximum pooling layer, a fourth convolution layer, a fourth maximum pooling layer, a fifth convolution layer and a fifth maximum pooling layer;
the input end of the first convolution layer is used as the input end of the CNN unit, and the output end of the first convolution layer is connected with the input end of the first maximum pooling layer; the input end of the second convolutional layer is connected with the output end of the first maximum pooling layer, and the output end of the second convolutional layer is connected with the input end of the second maximum pooling layer; the input end of the third convolutional layer is connected with the output end of the second largest pooling layer, and the output end of the third convolutional layer is connected with the input end of the third largest pooling layer; the input end of the fourth convolutional layer is connected with the output end of the third largest pooling layer, and the output end of the fourth convolutional layer is connected with the input end of the fourth largest pooling layer; the input end of the fifth convolutional layer is connected with the output end of the fourth largest pooling layer, and the output end of the fifth convolutional layer is connected with the input end of the fifth largest pooling layer; and the output end of the fifth maximum pooling layer is used as the output end of the CNN unit.
4. The deep learning chip for real-time seismic data processing according to claim 1, wherein the P-wave first arrival induction subsystem comprises: a third BilSTM unit, a second global attention unit and a first full connection layer unit;
the input end of the third BilSTM unit is used as the input end of the P-wave first arrival induction subsystem; the input end of the second global attention unit is connected with the output end of the third BilSTM unit, and the output end of the second global attention unit is connected with the input end of the first full-connection layer unit; and the output end of the first full connection layer unit is used as the output end of the P wave first arrival induction subsystem.
5. The deep learning chip for real-time seismic data processing according to claim 1, wherein the S-wave first arrival induction subsystem comprises: a fourth BilSt unit, a third global attention unit and a second full connection layer unit;
the input end of the fourth BilSTM unit is connected as the input end of the S-wave first arrival induction subsystem; the input end of the third global attention unit is connected with the output end of the fourth BilSTM unit, and the output end of the third global attention unit is connected with the input end of the second full-connection layer unit; and the output end of the second full connection layer unit is used as the output end of the S-wave first arrival induction subsystem.
6. The deep learning chip for real-time seismic data processing according to any one of claims 2, 4, and 5, wherein the input/output relationship of the cells in the LSTM module of the BiLSTM unit in the feature extraction subsystem, the P-wave first arrival induction subsystem, or the S-wave first arrival induction subsystem is:
f t =σ[(W f ·(y t-1 ,x t ,C t-1 )+b f ]
i t =tanh[W i ·(y t-1 ,x t ,C t-1 )+b i ]
h t =σ[W h ·(y t-1 ,x t ,C t-1 )+b h ]
C t =(C t-1 ⊙f t +(1-f t )⊙i t )⊙((1-i t )⊙h t )
y t =[W o ·(y t-1 ,x t ,C t-1 ,C t )+ o ]⊙h[ t ]
wherein f is t The output of the forgetting gate at time t, σ [ ]]For sigmoid activation functions, W f Weight of forgetting gate, b f To forget the biasing of the door, y t-1 Is the output of the cell at time t-1, x t Input of cells at time t, C t-1 The state of the cells at time t-1, i t Input the output of the gate at time t, tanh 2]As hyperbolic tangent activation function, W i As the weight of the input gate, b i For the biasing of the input gate, h t For the output of the candidate gate at time t, W h As the weight of the candidate door, b h As a candidate for the offset of the gate, C t The state of the cell at time t, <' > is Hadamard product, y t Output of the gate for time t, W o As weights of output gates, b o Is the biasing of the output gate.
7. The deep learning chip suitable for real-time seismic data processing according to any one of claims 2, 4, and 5, wherein the global attention unit in the feature extraction subsystem, the P-wave first-arrival induction subsystem, or the S-wave first-arrival induction subsystem comprises: a sixth convolution layer, a Softmax layer, a multiplier, a seventh convolution layer, a ReLU layer, an eighth convolution layer, and an adder;
the input end of the sixth convolutional layer is respectively connected with the first input end of the multiplier and the first input end of the adder and is used as the input end of the global attention unit; the input end of the Softmax layer is connected with the output end of the sixth convolutional layer, and the output end of the Softmax layer is connected with the second input end of the multiplier; the input end of the seventh convolution layer is connected with the output end of the multiplier, and the output end of the seventh convolution layer is connected with the input end of the ReLU layer; the input end of the eighth convolution layer is connected with the output end of the ReLU layer, and the output end of the eighth convolution layer is connected with the second input end of the adder; the output of the adder serves as the output of the global attention unit.
8. The deep learning chip for real-time seismic data processing according to claim 1, wherein the feature extraction subsystem, the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem use micro-seismic detection data and data labels to form a training data set, the feature extraction subsystem, the P-wave first-arrival induction subsystem and the S-wave first-arrival induction subsystem are trained by using the training data set to obtain a trained feature extraction subsystem, a trained P-wave first-arrival induction subsystem and a trained S-wave first-arrival induction subsystem, and the trained feature extraction subsystem, the trained P-wave first-arrival induction subsystem and the trained S-wave first-arrival induction subsystem are arranged in the processor.
9. The deep learning chip for real-time seismic data processing according to claim 8, wherein the weight update formula of the training process is:
Figure FDA0003983723930000041
Figure FDA0003983723930000042
wherein, w i+1 Weight of the i +1 th iteration, w i Is the weight of the ith iteration, η i Learning rate, η, for the ith iteration i-1 Learning rate for the i-1 st iteration, J i As a loss function for the ith iteration, J i-1 And gamma is a proportionality coefficient and zeta is a regulation constant for the loss function of the i-1 th iteration.
CN202211556940.XA 2022-12-06 2022-12-06 Deep learning chip suitable for real-time seismic data processing Active CN115903022B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202211556940.XA CN115903022B (en) 2022-12-06 2022-12-06 Deep learning chip suitable for real-time seismic data processing

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202211556940.XA CN115903022B (en) 2022-12-06 2022-12-06 Deep learning chip suitable for real-time seismic data processing

Publications (2)

Publication Number Publication Date
CN115903022A true CN115903022A (en) 2023-04-04
CN115903022B CN115903022B (en) 2023-10-31

Family

ID=86489479

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202211556940.XA Active CN115903022B (en) 2022-12-06 2022-12-06 Deep learning chip suitable for real-time seismic data processing

Country Status (1)

Country Link
CN (1) CN115903022B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112017289A (en) * 2020-08-31 2020-12-01 电子科技大学 Well-seismic combined initial lithology model construction method based on deep learning
CN112068195A (en) * 2019-06-10 2020-12-11 中国石油化工股份有限公司 Automatic first arrival picking method for microseism P & S wave matching event and computer storage medium
US20210117737A1 (en) * 2019-10-18 2021-04-22 Korea University Research And Business Foundation Earthquake event classification method using attention-based convolutional neural network, recording medium and device for performing the method
CN113158792A (en) * 2021-03-15 2021-07-23 辽宁大学 Microseismic event identification method based on improved model transfer learning
CN114660656A (en) * 2022-03-17 2022-06-24 中国科学院地质与地球物理研究所 Seismic data first arrival picking method and system

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112068195A (en) * 2019-06-10 2020-12-11 中国石油化工股份有限公司 Automatic first arrival picking method for microseism P & S wave matching event and computer storage medium
US20210117737A1 (en) * 2019-10-18 2021-04-22 Korea University Research And Business Foundation Earthquake event classification method using attention-based convolutional neural network, recording medium and device for performing the method
CN112017289A (en) * 2020-08-31 2020-12-01 电子科技大学 Well-seismic combined initial lithology model construction method based on deep learning
CN113158792A (en) * 2021-03-15 2021-07-23 辽宁大学 Microseismic event identification method based on improved model transfer learning
CN114660656A (en) * 2022-03-17 2022-06-24 中国科学院地质与地球物理研究所 Seismic data first arrival picking method and system

Also Published As

Publication number Publication date
CN115903022B (en) 2023-10-31

Similar Documents

Publication Publication Date Title
CN111723329B (en) Seismic phase feature recognition waveform inversion method based on full convolution neural network
CN107544904B (en) Software reliability prediction method based on deep CG-LSTM neural network
CN108596327B (en) Seismic velocity spectrum artificial intelligence picking method based on deep learning
CN101794528B (en) Gesture language-voice bidirectional translation system
CN106646587B (en) Object detection and recognition method and system based on acoustic vibration signal
CN111144542A (en) Oil well productivity prediction method, device and equipment
CN108830421B (en) Method and device for predicting gas content of tight sandstone reservoir
CN111580151B (en) SSNet model-based earthquake event time-of-arrival identification method
CN108875836B (en) Simple-complex activity collaborative recognition method based on deep multitask learning
CN110379418A (en) A kind of voice confrontation sample generating method
CN106384122A (en) Device fault mode identification method based on improved CS-LSSVM
CN111695413A (en) Signal first arrival pickup method and device combining U-Net and Temporal encoding
CN112052551B (en) Fan surge operation fault identification method and system
CN110223342B (en) Space target size estimation method based on deep neural network
CN113156492B (en) Real-time intelligent early warning method applied to TBM tunnel rockburst disasters
CN114818579A (en) Analog circuit fault diagnosis method based on one-dimensional convolution long-short term memory network
CN114862015A (en) Typhoon wind speed intelligent prediction method based on FEN-ConvLSTM model
Xu et al. An automatic P-wave onset time picking method for mining-induced microseismic data based on long short-term memory deep neural network
CN115903022A (en) Deep learning chip suitable for real-time seismic data processing
CN112132096B (en) Behavior modal identification method of random configuration network for dynamically updating output weight
CN106842172A (en) A kind of submarine target structural sparse feature extracting method
CN115424351A (en) Dynamic image identification method based on space-time fusion attention impulse neural network
CN112433249A (en) Horizon tracking method and device, computer equipment and computer readable storage medium
CN111487679B (en) Transverse wave velocity prediction method, device and equipment
CN116561528B (en) RUL prediction method of rotary machine

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant