CN106886023B - A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks - Google Patents
A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks Download PDFInfo
- Publication number
- CN106886023B CN106886023B CN201710110183.6A CN201710110183A CN106886023B CN 106886023 B CN106886023 B CN 106886023B CN 201710110183 A CN201710110183 A CN 201710110183A CN 106886023 B CN106886023 B CN 106886023B
- Authority
- CN
- China
- Prior art keywords
- layer
- image
- error term
- network
- layers
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Classifications
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/95—Radar or analogous systems specially adapted for specific applications for meteorological use
-
- G—PHYSICS
- G01—MEASURING; TESTING
- G01S—RADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
- G01S13/00—Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
- G01S13/88—Radar or analogous systems specially adapted for specific applications
- G01S13/89—Radar or analogous systems specially adapted for specific applications for mapping or imaging
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02A—TECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
- Y02A90/00—Technologies having an indirect contribution to adaptation to climate change
- Y02A90/10—Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation
Abstract
The invention discloses a kind of Radar Echo Extrapolation methods based on dynamic convolutional neural networks, it include: offline convolution neural metwork training: to given training image collection, training sample set is obtained by data prediction, initialize dynamic convolution neural network model, and using training sample set training dynamic convolutional neural networks, output valve is calculated by network propagated forward, the process of back-propagating update network parameter restrains dynamic convolutional neural networks.Online Radar Echo Extrapolation: test sample collection is converted for test chart image set by data prediction, trained dynamic convolutional neural networks are tested using test sample collection, by in input image sequence last width radar return image and network propagated forward in the probability vector phase convolution that obtains, the Radar Echo Extrapolation image predicted.
Description
Technical field
The invention belongs to surface weather observation technical fields in Atmospheric Survey, more particularly to one kind to be based on dynamic convolutional Neural
The Radar Echo Extrapolation method of network.
Background technique
Nowcasting refers mainly to the weather forecast of 0~3 hour high-spatial and temporal resolution, and Main Prediction object includes strong drop
The diastrous weathers such as water, strong wind, hail.Currently, many forecast systems all use Numerical Prediction Models, but due to numerical forecast
In the presence of return slow (spin-up) has been forecast, Nowcasting ability is limited.New Generation Doppler Weather Radar has very high
Sensitivity and resolution ratio, the spatial resolution of data information can reach 200~1000m, and temporal resolution can reach 2
~15min.In addition, Doppler radar also has reasonable operating mode, comprehensive condition monitoring and fault warning, advanced
Real-time Calibration System and Radar meteorology product algorithm abundant, the reliability of Nowcasting can be greatly improved.Nowadays,
New Generation Doppler Weather Radar has become one of most effective tool of nowcasting, is faced using Doppler radar
Nearly forecast is based primarily upon Radar Echo Extrapolation technology, i.e., according to current time radar observation result, thus it is speculated that radar return future
Position and intensity, to realize the track prediction to strong convection system.
Traditional Radar Echo Extrapolation method is centroid tracking method and the cross-correlation technique based on maximum correlation coefficient
(Tracking Radar Echoes by Correlation, TREC), but all there is certain deficiency, mass center in conventional method
Tracing is only applicable to echo compared with strong, the lesser storm monomer of range, unreliable for the forecast of a wide range of precipitation;TREC is general
Echo is considered as linear change, and echo variation is increasingly complex in reality, while such method is vulnerable in vector field
Unordered vector disturbance.In addition, existing method is low to the utilization rate of Radar Data, and history Radar Data includes local weather system
The important feature of system variation, has very high researching value.
For improve Radar Echo Extrapolation timeliness, and from a large amount of history Radar Data study radar return variation
Machine learning method is introduced into Radar Echo Extrapolation by rule.Convolutional neural networks (Convolutional Neural
Network, CNN) important branch as deep learning, it is widely used in image procossing, the fields such as pattern-recognition.The network is most
Big feature is using part connection, weight is shared, down-sampling method, has to the deformation of input picture, translation and overturning
There is stronger adaptability.For strong temporal correlation existing between radar return image, the dynamic convolution based on input is designed
Neural network, which can dynamically change weighting parameter according to the radar echo map of input, and then predict extrapolated image.Benefit
With history Radar Data training dynamic convolutional neural networks, network is made more fully to extract echo character, study echo variation
Rule, for improving Radar Echo Extrapolation accuracy, optimization nowcasting effect is of great significance.
Summary of the invention
Goal of the invention: when the technical problem to be solved by the present invention is to be directed to the extrapolation of existing Radar Echo Extrapolation method
Imitate it is short, it is insufficient to Radar Data utilization rate, propose a kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks, it is real
Now to the high plane such as radar echo intensity show CAPPI (Constant AltitudePlan Position Indicator,
CAPPI) the outside forecast of image, comprising the following steps:
Step 1, the offline convolutional neural networks of training: input training image collection carries out data prediction to training image collection,
Training sample set is obtained, designs dynamic convolution neural network structure, and initialize network training parameter;It is assembled for training using training sample
Practice dynamic convolutional neural networks, the orderly image sequence of input obtains width prediction by dynamic convolutional neural networks propagated forward
Image calculates forecast image and compares the error between label, updates the weighting parameter of network by backpropagation and biasing is joined
Number repeats this process until reaching trained termination condition, obtains convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, obtains
Test sample collection, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by before network to biography
Calculating probability vector is broadcast, and last width radar return image in input image sequence is mutually rolled up with obtained probability vector
Product, the Radar Echo Extrapolation image predicted.
Step 1 of the present invention the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image standardize
Change processing, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image
Set is divided, and construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generate
The sub-network of probability vector reconstructs the probabilistic forecasting layer for image extrapolation, provides for the offline neural metwork training stage dynamic
State convolutional neural networks initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, training stage
The sample size BatchSize=10 inputted every time, most large quantities of frequency of training of training sample setCurrently crowd frequency of training BatchNum=1, the maximum number of iterations of network training
IterationMax=40, current iteration number IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, training the training sample obtained from step 1-1 every time
It concentrates and reads BatchSize group training sample, every group of training sample is { x1,x2,x3,x4, y }, it altogether include 5 width images, wherein
{x1,x2,x3,x4It is used as input image sequence, y is corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level
Probability vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence according to
It is secondary with VPV, HPV phase convolution, obtain the output forecast image of propagated forward;
Backpropagation: step 1-6 reversely acquires the error term of probability vector, according to probability vector in probabilistic forecasting layer
Error term from rear to the preceding layer-by-layer error term for calculating each network layer in subnet network layers, and then calculate error in each network layer
Item utilizes the parameter of obtained gradient updating dynamic convolutional neural networks to the gradient of weighting parameter and offset parameter;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following
Three kinds of situations:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then step is returned
Rapid 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current net
Network the number of iterations is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, returns
Step 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network changes
Generation number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates offline neural metwork training rank
Section, obtains trained dynamic convolution neural network model.
Step 1-1 data prediction of the present invention the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, when
Between between be divided into 6 minutes, altogether include NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus, [N] indicates the maximum integer for being not more than N, acquires
After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes
Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains
The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 1-1-3 constructs training sample set:
Training sample set is constructed using the floating number image set that step 1-1-2 is obtained, by every four in floating number image set
Adjacent image, i.e. { 4N+1,4N+2,4N+3,4N+4 } width image are as one group of list entries, [4 × (N+1)+1] width figure
It is cut as passing through, the part that the central resolution ratio of reservation is 240 × 240 is as the control label for corresponding to sample, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0,
TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting
Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Image conversion: step 1-1-2-1 for the computation complexity for reducing convolutional neural networks training, step 1-1-1 is adopted
The image that sample obtains is converted into gray level image, retains the part that original image center resolution ratio is 560 × 560 by cutting, will
Image resolution ratio boil down to 280 × 280 after cutting obtains the grayscale image that resolution ratio is 280 × 280;
Step 1-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel
Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
Step 1-2 of the present invention the following steps are included:
Step 1-2-1 constructs sub-network:
Sub-network is made of 10 network layers, be followed successively by from front to back convolutional layer C1, down-sampling layer S1, convolutional layer C2, under
Sample level S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer:
In probabilistic forecasting layer construct dynamic convolutional layer DC1 and dynamic convolutional layer DC2, by sub-network output vertical probability to
Measure convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer,
Convolution kernel k and offset parameter bias.For convolution kernel, it is thus necessary to determine that width KernelSize (point of convolution kernel of convolution kernel
Resolution is KernelSize × KernelSize), the quantity KernelNumber of convolution kernel (value be the convolutional layer input with it is defeated
The product of characteristic pattern quantity out), and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity with should
The output characteristic pattern quantity of layer is identical.For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern width of this layer
ForValue codetermined by its input feature vector figure resolution ratio and the size of convolution kernel, i.e.,Indicate convolutional layer lCUpper one layer of convolutional layer
Output characteristic pattern width;
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1=12, export the width of characteristic pattern
OutputSizeC1=272, convolution kernel width KernelSizeC1=9, offset parameter biasC1It is initialized as zero, C1 layers of volume
Product core kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, C2 layers of output characteristic pattern quantity OutputMaps are enabledC2=32, export the width of characteristic pattern
OutputSizeC2=128, convolution kernel width KernelSizeC2=9, offset parameter is initialized as zero, C2 layers of convolution kernel
kC2Quantity KernelNumberC2=384, the initial value of each parameter is in convolution kernel
For convolutional layer C3, C3 layers of output characteristic pattern quantity OutputMaps are enabledC3=32, export the width of characteristic pattern
OutputSizeC3=56, convolution kernel width KernelSizeC3=9, offset parameter is initialized as zero, C3 layers of convolution kernel kC3
Quantity KernelNumberC3=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C4, C4 layers of output characteristic pattern quantity OutputMaps are enabledC4=32, export the width of characteristic pattern
OutputSizeC4=20, convolution kernel width KernelSizeC4=9, offset parameter is initialized as zero, C4 layers of convolution kernel kC4
Quantity KernelNumberC4=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C5, this layer of OutputMaps is enabledC5=32, OutputSizeC5=4, KernelSizeC5=7, partially
Set the convolution kernel k that parameter is initialized as zero, C5 layersC5Quantity KernelNumberC5=1024, each ginseng in convolution kernel
Several initial values are
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1,
The sampling core of S2, S3 and S4 are initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output
Characteristic pattern quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, output characteristic pattern is wide
DegreeIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, F1 layers of weighting parameter
For horizontal weighting parameter matrix W H and vertical weighting parameter matrix W V, size is 41 × 512, is enabled every in weighting parameter matrix
The initial value of one parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV,
It is initialized as 41 × 1 one-dimensional null vector.
Step 1-5 of the present invention the following steps are included:
Step 1-5-1, sub-network calculate probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network
Extract input image sequence characteristic, handled in classifier layer by Softmax function, obtain level probability vector HPV with
Vertical probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting
The convolution kernel of layer obtains the defeated of propagated forward by the last piece image in input image sequence successively with VPV, HPV phase convolution
Forecast image out.
Step 1-5-1 of the present invention the following steps are included:
Step 1-5-1-1, judges network layer type: indicating the network layer in the sub-network being presently in, l initial value with l
For C1, the type of network layer l is judged, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l
∈ { S1, S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCLayer
J-th of output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are asked
L is added with, summed resultCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveIt calculates public
Formula is as follows:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n is current
The output characteristic pattern number of the previous down-sampling layer of convolutional layer, Indicate lCI-th of input of layer
Characteristic pattern, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input
Layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and returns to step
Rapid 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained
The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lSLayer
Output characteristic patternCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolution of current down-sampling layer
Layer,Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, more by l
It is newly l+1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 is become by matrix
Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1
Export feature vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1's
Apposition sums calculated result with Horizontal offset parameter BH, vertical off setting parameter BV respectively, after the processing of Softmax function
To level probability vector HPV and vertical probability vector VPV, specific formula for calculation is as follows:
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to
VPV phase convolution is measured, the DC1 layer that resolution ratio is 240 × 280 is obtained and exports characteristic pattern aDC1;
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1With vertical probability vector HPV phase
Convolution, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
Step 1-6 of the present invention the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the instruction of the step 1-5-2-2 forecast image obtained and input
The control label practiced in sample asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term of level probability vector
δHPVWith the error term δ of vertical probability vectorVPV;
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith vertical probability vector
Error term δVPV, from it is rear to preceding successively calculate classification layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2,
The error term of S1, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the mistake of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained
Gradient value of the poor item to this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the ladder of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter
Angle value is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, will be former
Weighting parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
Step 1-6-1 of the present invention the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image and the group that step 1-5-2-2 is obtained
The control label of sample asks poor, obtains the error term matrix delta that size is 240 × 240DC2;
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2
Expanding is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability after overturning
Vector phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates that zero extended function, rot180 () indicate that angle is 180 ° of rotation letter
2 × 2 matrix zero is extended for 4 × 4 matrix by number, the matrix after zero expansion, the region and former square that central resolution ratio is 2 × 2
Consistent, zero pixel filling of remaining position of battle array;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers
Export characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term of HPV
δHPV, formula is as follows:
δHPV=aDC1*δDC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Mutually roll up
It is long-pending, 41 × 1 column vector is obtained after convolution, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term of the step 1-6-1-3 probability vector obtained
δVPVAnd δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by square
The apposition of battle array is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, the F1 layer that will be obtained in step 1-6-2-1
Error term δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term δC5,Table
Show that transformed 32nd resolution ratio is 4 × 4 matrix;Matrixing in this time matrixing and step 1-5-1-6 is each other
Inverse transformation,
Step 1-6-2-3, judges network layer type: network layer (the l initial value in the sub-network being presently in is indicated with l
For S4), the type of network layer l is judged, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l
∈ { C4, C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2,
S3, S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrix
It expands to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer
J-th of error term, and haveAgain by corresponding convolution kernel180 degree is rotated, after then expanding
Matrix and overturning after convolution nuclear phase convolution, and convolution results are summed, obtain lSI-th of error term matrix of layerIt is public
Formula is as follows:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and returns to step
Rapid 1-6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3,
C4, C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCThe i-th of layer
A error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-samplingIn each element error entry value average mark into sampling area, obtaining resolution ratio is
Up-sampling matrix, then calculate activation primitive in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure,
Obtain lCI-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, the corresponding up-sampling of each of original image pixel after up-sampling
Region in each of original pixel value mean allocation to sampling area pixel, successively calculates all error term matrixes, obtains
To lCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates a upper network layer
Error term;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC
∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersC- 1 layer
I-th of output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the ladder of corresponding convolution kernel
Angle valueFormula is as follows:
In above formula,WithRespectively indicate lCThe output characteristic pattern number and l of layerC-
1 layer of output characteristic pattern number, ▽ k are gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC
∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of layer
Error term matrixIn all elements sum, obtain j-th of this layer biasing gradient valueThe following institute of formula
Show:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculate separately level probability vector with it is vertical
The error term δ of probability vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH,
The gradient value of WV, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is error term to vertical weighting parameter
Gradient value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability to
The error term δ of amountHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV,
Formula is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is error term to vertical off setting parameter
Gradient value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term pair that step 1-6-3-1 is obtained
The gradient of convolution kernel is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and be somebody's turn to do
Correction term asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term pair that step 1-6-3-2 is obtained
The gradient of biasing is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former bias term and be somebody's turn to do
Correction term asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layers of weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained is to weighting parameter
The gradient value of WH and WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then former weight is joined
Number WH and WV asks poor with the correction term acquired respectively, the WH and WV updated, and formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layers of offset parameter: the F1 layer error term that step 1-6-3-4 is obtained is to offset parameter
The gradient value of BH and BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then original biasing is joined
Number BH and BV asks poor with the correction term acquired respectively, the BH and BV updated, and formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
Step 2 of the present invention the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image standardize
Change processing, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, constructs
Test sample collection comprising TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is by training
Dynamic convolutional neural networks in;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector
HPVtestWith vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with
VPVtest、HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
Step 2-1 of the present invention the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, when
Between between be divided into 6 minutes, altogether include NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, is adopted
Last image is concentrated to meet the requirements amount of images by deleting test image when sample;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains
The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 2-1-3 constructs test sample collection: constructing test sample using the floating number image set that step 2-1-2 is obtained
Collection, by four adjacent images every in floating number image set, i.e. { 4M+1,4M+2,4M+3,4M+4 } width image is defeated as one group
Enter sequence, for [4 × (M+1)+1] width image by cutting, retaining central resolution ratio is 240 × 240 part as corresponding sample
This control label, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test specimens
This test sample collection;
Step 2-1-2 the following steps are included:
Step 2-1-2-1, image conversion: to reduce the computation complexity that convolutional neural networks are trained, by colored echo
Intensity CAPPI image is converted into gray level image, then retains the part that original image center resolution ratio is 560 × 560 by cutting,
By the image resolution ratio boil down to 280 × 280 after cutting, the grayscale image that resolution ratio is 280 × 280 is obtained;
Step 2-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel
Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network
The image sequence characteristic for extracting input, is then handled by Softmax function in classifier layer, obtains level probability vector
HPVtestWith vertical probability vector VPVtest;
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probability
The convolution kernel of prediction interval, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, is moved
The final extrapolated image of state convolutional neural networks;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network
The type of layer p, p initial value is C1, if p ∈ { C1, C2, C3, C4, C5 }, then p is convolutional layer, step 2-3-1-2 is executed, if p ∈
{ S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCLayer
V-th of output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then by convolution knot
Fruit summation, summed result add pCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCThe v of layer
A output characteristic patternCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m are to work as
The output characteristic pattern number of the previous down-sampling layer of preceding convolutional layer,Indicate pCU-th of layer
Input feature vector figure, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is
Input layer.
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and returns to step
Rapid 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained
The output characteristic pattern of the convolutional layer arrived then with step-length is respectively 2 to be sampled with phase convolution, and sampling obtains pS
The output characteristic pattern calculation formula of layer is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolution of current down-sampling layer
Layer,Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, by p
It is updated to p+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 is become by matrix
Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1
Export feature vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, by calculate tie
Fruit is summed with Horizontal offset parameter BH, vertical off setting parameter BV respectively, and summed result obtains level after the processing of Softmax function
Probability vector HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to
Measure VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector
HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240, by extrapolated image
It compares with the control label of corresponding test sample, determines that dynamic convolutional neural networks carry out the accuracy of Radar Echo Extrapolation.
The utility model has the advantages that being based on dynamic convolutional neural networks, trained on a large amount of radar return image set, and utilize
Trained network carries out Radar Echo Extrapolation, effectively increases the accuracy of Radar Echo Extrapolation.
Specifically the present invention has the advantage that the utilization rate of 1. Radar Datas is high compared with existing method, with traditional thunder
It is only compared up to echo extrapolation technique using a small amount of Radar Data, present invention training on a large amount of radar return image set, number
It is higher according to utilization rate;2. timeliness of extrapolating is long, the time interval between sample can be concentrated outer to improve by adjusting training sample
Push away timeliness.
Detailed description of the invention
The present invention is done with reference to the accompanying drawings and detailed description and is further illustrated, it is of the invention above-mentioned or
Otherwise advantage will become apparent.
Fig. 1 is flow chart of the present invention.
Fig. 2 is dynamic convolutional neural networks initialization model structure chart.
Fig. 3 is sub-network structural map.
Fig. 4 is probabilistic forecasting layer structural map.
Fig. 5 is that matrix zero expands schematic diagram.
Fig. 6 is the process schematic up-sampled to 2 × 2 matrix.
Specific embodiment
The present invention will be further described with reference to the accompanying drawings and embodiments.
As shown in Figure 1, the present invention the following steps are included:
Step 1, as shown in Fig. 2, the offline convolutional neural networks of training: input training image collection carries out training image collection
Data prediction obtains training sample set, designs dynamic convolution neural network structure, and initialize network training parameter;It utilizes
Training sample set trains dynamic convolutional neural networks, and the orderly image sequence of input passes through dynamic convolutional neural networks propagated forward
A width forecast image is obtained, forecast image is calculated and compares the error between label, the weight of network is updated by backpropagation
Parameter and offset parameter repeat this process until reaching trained termination condition, obtain convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, obtains
Test sample collection, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by before network to biography
Calculating probability vector is broadcast, and last width radar return image in input image sequence is mutually rolled up with obtained probability vector
Product, the Radar Echo Extrapolation image predicted.
Step 1 of the present invention the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image standardize
Change processing, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image
Set is divided, and construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generate
The sub-network of probability vector reconstructs the probabilistic forecasting layer for image extrapolation, provides for the offline neural metwork training stage dynamic
State convolutional neural networks initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, training stage
The sample size BatchSize=10 inputted every time, most large quantities of frequency of training of training sample setCurrently crowd frequency of training BatchNum=1, the maximum number of iterations of network training
IterationMax=40, current iteration number IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, training the training sample obtained from step 1-1 every time
It concentrates and reads BatchSize group training sample, every group of training sample is { x1,x2,x3,x4, y }, it altogether include 5 width images, wherein
{x1,x2,x3,x4It is used as input image sequence, x1Indicate that piece image, y are corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level
Probability vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence according to
It is secondary with VPV, HPV phase convolution, obtain the output forecast image of propagated forward;
Backpropagation: step 1-6 reversely acquires the error term of probability vector, according to probability vector in probabilistic forecasting layer
Error term from rear to the preceding layer-by-layer error term for calculating each network layer in subnet network layers, and then calculate error in each network layer
Item utilizes the parameter of obtained gradient updating dynamic convolutional neural networks to the gradient of weighting parameter and offset parameter;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following
Three kinds of situations:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then step is returned
Rapid 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current net
Network the number of iterations is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, returns
Step 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network changes
Generation number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates offline neural metwork training rank
Section, obtains trained dynamic convolution neural network model.
Step 1-1 data prediction of the present invention the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, when
Between between be divided into 6 minutes, altogether include NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus, [N] indicates the maximum integer for being not more than N, acquires
After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes
Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains
The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 1-1-3 constructs training sample set:
Training sample set is constructed using the floating number image set that step 1-1-2 is obtained, by every four in floating number image set
Adjacent image, i.e. { 4N+1,4N+2,4N+3,4N+4 } width image are as one group of list entries, [4 × (N+1)+1] width figure
It is cut as passing through, the part that the central resolution ratio of reservation is 240 × 240 is as the control label for corresponding to sample, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0,
TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting
Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Image conversion: step 1-1-2-1 for the computation complexity for reducing convolutional neural networks training, step 1-1-1 is adopted
The image that sample obtains is converted into gray level image, retains the part that original image center resolution ratio is 560 × 560 by cutting, will
Image resolution ratio boil down to 280 × 280 after cutting obtains the grayscale image that resolution ratio is 280 × 280;
Step 1-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel
Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
Step 1-2 of the present invention the following steps are included:
Step 1-2-1 constructs sub-network: sub-network is configured to structure as shown in Figure 3:
Sub-network is made of 10 network layers, be followed successively by from front to back convolutional layer C1, down-sampling layer S1, convolutional layer C2, under
Sample level S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer: probabilistic forecasting layer is configured to structure as shown in Figure 4:
In probabilistic forecasting layer construct dynamic convolutional layer DC1 and dynamic convolutional layer DC2, by sub-network output vertical probability to
Measure convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer,
Convolution kernel k and offset parameter bias.For convolution kernel, it is thus necessary to determine that width KernelSize (point of convolution kernel of convolution kernel
Resolution is KernelSize × KernelSize), the quantity KernelNumber of convolution kernel (value be the convolutional layer input with it is defeated
The product of characteristic pattern quantity out), and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity with should
The output characteristic pattern quantity of layer is identical.For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern width of this layer
ForValue codetermined by its input feature vector figure resolution ratio and the size of convolution kernel, i.e.,
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1=12, export the width of characteristic pattern
OutputSizeC1=272, convolution kernel width KernelSizeC1=9, offset parameter biasC1It is initialized as zero, C1 layers of volume
Product core kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, this layer of OutputMaps is enabledC2=32, OutputSizeC2=128, KernelSizeC2=9,
Offset parameter is initialized as zero, C2 layers of convolution kernel kC2Quantity KernelNumberC2=384, each ginseng in convolution kernel
Several initial values are
For convolutional layer C3, this layer of OutputMaps is enabledC3=32, OutputSizeC3=56, KernelSizeC3=9, partially
Set the convolution kernel k that parameter is initialized as zero, C3 layersC3Quantity KernelNumberC3=1024, each ginseng in convolution kernel
Several initial values are
For convolutional layer C4, this layer of OutputMaps is enabledC4=32, OutputSizeC4=20, KernelSizeC4=9, partially
Set the convolution kernel k that parameter is initialized as zero, C4 layersC4Quantity KernelNumberC4=1024, each ginseng in convolution kernel
Several initial values are
For convolutional layer C5, this layer of OutputMaps is enabledC5=32, OutputSizeC5=4, KernelSizeC5=7, partially
Set the convolution kernel k that parameter is initialized as zero, C5 layersC5Quantity KernelNumberC5=1024, each ginseng in convolution kernel
Several initial values are
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1,
The sampling core of S2, S3 and S4 are initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output
Characteristic pattern quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, output characteristic pattern is wide
DegreeIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, F1 layers of weighting parameter
For horizontal weighting parameter matrix W H and vertical weighting parameter matrix W V, size is 41 × 512, is enabled every in weighting parameter matrix
The initial value of one parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV,
It is initialized as 41 × 1 one-dimensional null vector.
Step 1-5 of the present invention the following steps are included:
Step 1-5-1, sub-network calculate probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network
Extract input image sequence characteristic, handled in classifier layer by Softmax function, obtain level probability vector HPV with
Vertical probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting
The convolution kernel of layer obtains the defeated of propagated forward by the last piece image in input image sequence successively with VPV, HPV phase convolution
Forecast image out.
Step 1-5-1 of the present invention the following steps are included:
Step 1-5-1-1, judges network layer type: indicating the network layer in the sub-network being presently in, l initial value with l
For C1, the type of network layer l is judged, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l
∈ { S1, S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCLayer
J-th of output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are asked
L is added with, summed resultCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveIt calculates public
Formula is as follows:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n is current
The output characteristic pattern number of the previous down-sampling layer of convolutional layer,Indicate lCI-th of input of layer
Characteristic pattern, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input
Layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and returns to step
Rapid 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained
The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lS
The output characteristic pattern of layerCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolution of current down-sampling layer
Layer,Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, more by l
It is newly l+1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 is become by matrix
Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1
Export feature vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1's
Apposition sums calculated result with Horizontal offset parameter BH, vertical off setting parameter BV respectively, after the processing of Softmax function
To level probability vector HPV and vertical probability vector VPV, specific formula for calculation is as follows:
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to
VPV phase convolution is measured, the DC1 layer that resolution ratio is 240 × 280 is obtained and exports characteristic pattern aDC1;
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1With vertical probability vector HPV phase
Convolution, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
Step 1-6 of the present invention the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the instruction of the step 1-5-2-2 forecast image obtained and input
The control label practiced in sample asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term of level probability vector
δHPVWith the error term δ of vertical probability vectorVPV;
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith vertical probability vector
Error term δVPV, from it is rear to preceding successively calculate classification layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2,
The error term of S1, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the mistake of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained
Gradient value of the poor item to this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the ladder of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter
Angle value is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, will be former
Weighting parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
Step 1-6-1 of the present invention the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image and the group that step 1-5-2-2 is obtained
The control label of sample asks poor, obtains the error term matrix delta that size is 240 × 240DC2;
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2
Expanding is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability after overturning
Vector phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates that zero extended function, rot180 () indicate that angle is 180 ° of rotation letter
Number, by 2 × 2 matrix zero be extended for 4 × 4 process as shown in figure 5, zero expand after matrix, central resolution ratio is 2 × 2
Region is consistent with original matrix, zero pixel filling of remaining position;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers
Export characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term of HPV
δHPV, formula is as follows:
δHPV=aDC1*δDC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Mutually roll up
It is long-pending, 41 × 1 column vector is obtained after convolution, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term of the step 1-6-1-3 probability vector obtained
δVPVAnd δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by square
The apposition of battle array is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, the F1 layer that will be obtained in step 1-6-2-1
Error term δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term
Indicate that transformed 32nd resolution ratio is 4 × 4 matrix;This time matrixing and the matrixing in step 1-5-1-6 are mutual
For inverse transformation,
Step 1-6-2-3, judges network layer type: network layer (the l initial value in the sub-network being presently in is indicated with l
For S4), the type of network layer l is judged, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l
∈ { C4, C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2,
S3, S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrix
It expands to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer
J-th of error term, and haveAgain by corresponding convolution kernel180 degree is rotated, after then expanding
Matrix and overturning after convolution nuclear phase convolution, and convolution results are summed, obtain lSI-th of error term matrix of layerIt is public
Formula is as follows:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and returns to step
Rapid 1-6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3,
C4, C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCThe i-th of layer
A error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-samplingIn each element error entry value average mark into sampling area, obtaining resolution ratio is
Up-sampling matrix, then calculate activation primitive in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure,
Obtain lCI-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, and Fig. 6 is the process that is up-sampled of matrix to 2 × 2, after up-sampling
Each of original image pixel corresponds to a up-sampling region, each of original pixel value mean allocation to sampling area picture
In vegetarian refreshments, all error term matrixes are successively calculated, l is obtainedCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates a upper network layer
Error term;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC
∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersC- 1 layer
I-th of output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the ladder of corresponding convolution kernel
Angle valueFormula is as follows:
In above formula,WithRespectively indicate lCLayer and lC- 1 layer of output characteristic pattern
Number, ▽ k are gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC
∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of layer
Error term matrixIn all elements sum, obtain j-th of this layer biasing gradient valueThe following institute of formula
Show:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculate separately level probability vector with it is vertical
The error term δ of probability vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH,
The gradient value of WV, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is error term to vertical weighting parameter
Gradient value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability to
The error term δ of amountHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV,
Formula is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is error term to vertical off setting parameter
Gradient value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term pair that step 1-6-3-1 is obtained
The gradient of convolution kernel is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and be somebody's turn to do
Correction term asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term pair that step 1-6-3-2 is obtained
The gradient of biasing is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former bias term and be somebody's turn to do
Correction term asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layers of weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained is to weighting parameter
The gradient value of WH and WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then former weight is joined
Number WH and WV asks poor with the correction term acquired respectively, the WH and WV updated, and formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layers of offset parameter: the F1 layer error term that step 1-6-3-4 is obtained is to offset parameter
The gradient value of BH and BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then original biasing is joined
Number BH and BV asks poor with the correction term acquired respectively, the BH and BV updated, and formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
Step 2 of the present invention the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image standardize
Change processing, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, constructs
Test sample collection comprising TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is by training
Dynamic convolutional neural networks in;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector
HPVtestWith vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with
VPVtest、HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
Step 2-1 of the present invention the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, when
Between between be divided into 6 minutes, altogether include NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, is adopted
Last image is concentrated to meet the requirements amount of images by deleting test image when sample;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains
The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 2-1-3 constructs test sample collection: constructing test sample using the floating number image set that step 2-1-2 is obtained
Collection, by four adjacent images every in floating number image set, i.e. { 4M+1,4M+2,4M+3,4M+4 } width image is defeated as one group
Enter sequence, for [4 × (M+1)+1] width image by cutting, retaining central resolution ratio is 240 × 240 part as corresponding sample
This control label, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test specimens
This test sample collection;
Step 2-1-2 the following steps are included:
Step 2-1-2-1, image conversion: to reduce the computation complexity that convolutional neural networks are trained, by colored echo
Intensity CAPPI image is converted into gray level image, then retains the part that original image center resolution ratio is 560 × 560 by cutting,
By the image resolution ratio boil down to 280 × 280 after cutting, the grayscale image that resolution ratio is 280 × 280 is obtained;
Step 2-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel
Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network
The image sequence characteristic for extracting input, is then handled by Softmax function in classifier layer, obtains level probability vector
HPVtestWith vertical probability vector VPVtest;
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probability
The convolution kernel of prediction interval, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, is moved
The final extrapolated image of state convolutional neural networks;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network
The type of layer p, p initial value is C1, if p ∈ { C1, C2, C3, C4, C5 }, then p is convolutional layer, step 2-3-1-2 is executed, if p ∈
{ S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCLayer
V-th of output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then by convolution results
Summation, summed result add pCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCV-th of layer
Export characteristic patternCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m is current
The output characteristic pattern number of the previous down-sampling layer of convolutional layer,Indicate pCU-th of layer is defeated
Enter characteristic pattern, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is defeated
Enter layer.
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and returns to step
Rapid 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained
The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains pS
The output characteristic pattern of layerCalculation formula is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolution of current down-sampling layer
Layer,Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, by p
It is updated to p+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 is become by matrix
Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1
Export feature vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, will calculate
As a result it sums respectively with Horizontal offset parameter BH, vertical off setting parameter BV, summed result obtains water after the processing of Softmax function
Flat probability vector HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to
Measure VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector
HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240, by extrapolated image
It compares with the control label of corresponding test sample, determines that dynamic convolutional neural networks carry out the accuracy of Radar Echo Extrapolation.
The present invention provides a kind of Radar Echo Extrapolation methods based on dynamic convolutional neural networks, implement the technology
There are many method and approach of scheme, the above is only a preferred embodiment of the present invention, it is noted that for the art
Those of ordinary skill for, various improvements and modifications may be made without departing from the principle of the present invention, these change
It also should be regarded as protection scope of the present invention into retouching.The available prior art of each component part being not known in the present embodiment adds
To realize.
Claims (9)
1. a kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks, which comprises the following steps:
Step 1, the offline convolutional neural networks of training: input training image collection carries out data prediction to training image collection, obtains
Training sample set designs dynamic convolution neural network structure, and initializes network training parameter;It is dynamic using training sample set training
The orderly image sequence of state convolutional neural networks, input obtains a width prognostic chart by dynamic convolutional neural networks propagated forward
Picture calculates forecast image and compares the error between label, the weighting parameter and offset parameter of network are updated by backpropagation,
This process is repeated until reaching trained termination condition, obtains convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, is tested
Sample set, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by network propagated forward meter
Calculate probability vector, and by input image sequence last width radar return image and obtained probability vector phase convolution, obtain
To the Radar Echo Extrapolation image of prediction;
Step 1 the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image carry out at standardization
Reason, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image collection
It is divided, construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generating probability
The sub-network of vector reconstructs the probabilistic forecasting layer for image extrapolation, provides dynamic volume for the offline neural metwork training stage
Product neural network initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, the training stage is each
The sample size BatchSize=10 of input, most large quantities of frequency of training of training sample set
Currently crowd frequency of training BatchNum=1, the maximum number of iterations IterationMax=40 of network training, current iteration number
IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, the training sample that training is obtained from step 1-1 every time is concentrated
BatchSize group training sample is read, every group of training sample is { x1,x2,x3,x4, y }, it include altogether 5 width images, wherein { x1,x2,
x3,x4It is used as input image sequence, y is corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level probability
Vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence successively with
VPV, HPV phase convolution obtain the output forecast image of propagated forward;
Step 1-6, backpropagation: reversely acquiring the error term of probability vector in probabilistic forecasting layer, according to the mistake of probability vector
The error term of poor Xiang Conghou each network layer into preceding layer-by-layer calculating subnet network layers, and then calculate error term pair in each network layer
The gradient of weighting parameter and offset parameter utilizes the parameter of obtained gradient updating dynamic convolutional neural networks;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following three kinds
Situation:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then return step 1-4
Continue to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current network changes
Generation number is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, return step
1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network iteration time
Number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates the offline neural metwork training stage, obtain
To trained dynamic convolution neural network model.
2. the method according to claim 1, wherein step 1-1 data prediction the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, between the time
It is divided into 6 minutes, altogether includes NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus,Expression is not more thanMaximum integer, acquire
After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes
Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original resolution to the image that sampling obtains
The floating number image that resolution ratio is 280 × 280 is converted into for 2000 × 2000 color image;
Step 1-1-3 constructs training sample set: training sample set is constructed using the floating number image set that step 1-1-2 is obtained, it will
Every four adjacent images in floating number image set, i.e., { 4N+1,4N+2,4N+3,4N+4 } width image is as one group of input sequence
Column, [4 × (N+1)+1] width image is by cutting, and the part that the central resolution ratio of reservation is 240 × 240 is as corresponding sample
Label is compareed, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0,
TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting
Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Step 1-1-2-1, image conversion: converting gray level image for the step 1-1-1 image sampled, is retained by cutting
The image resolution ratio boil down to 280 × 280 after cutting is divided in the part that original image center resolution ratio is 560 × 560
The grayscale image that resolution is 280 × 280;
Step 1-1-2-2, data normalization: by the value of each of the grayscale image obtained in step 1-1-2-1 pixel from
[0~255] is mapped to [0~1], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
3. according to the method described in claim 2, it is characterized in that, step 1-2 the following steps are included:
Step 1-2-1 constructs sub-network:
Sub-network is made of 10 network layers, is followed successively by convolutional layer C1, down-sampling layer S1, convolutional layer C2, down-sampling from front to back
Layer S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer:
Dynamic convolutional layer DC1 and dynamic convolutional layer DC2 is constructed in probabilistic forecasting layer, the vertical probability vector that sub-network is exported
Convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer, convolution
Core k and offset parameter bias, for convolution kernel, it is thus necessary to determine that the width KernelSize of convolution kernel, the quantity of convolution kernel
KernelNumber, the quantity of convolution kernel are the product that the convolutional layer inputs and exports characteristic pattern quantity, the resolution ratio of convolution kernel
For KernelSize × KernelSize, and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity
It is identical as the output characteristic pattern quantity of this layer;For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern of this layer
Width is Value by convolutional layer lCInput feature vector figure resolution ratio and convolution kernel widthIt codetermines, i.e.,
Indicate convolutional layer lCIt is upper one layer volume
The output characteristic pattern width of lamination;
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1The width of=12, C1 layers of output characteristic pattern
OutputSizeC1=272, C1 layers of convolution kernel width KernelSizeC1=9, C1 layers of offset parameter biasC1It is initialized as zero,
C1 layers of convolution kernel kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, C2 layers of output characteristic pattern quantity OutputMaps are enabledC2The width of=32, C2 layers of output characteristic pattern
OutputSizeC2=128, C2 layers of convolution kernel width KernelSizeC2=9, C2 layers of offset parameter are initialized as zero, C2 layers
Convolution kernel kC2Quantity KernelNumberC2=384, the initial value of each parameter is in convolution kernel
For convolutional layer C3, C3 layers of output characteristic pattern quantity OutputMaps are enabledC3The width of=32, C3 layers of output characteristic pattern
OutputSizeC3=56, C3 layers of convolution kernel width KernelSizeC3=9, C3 layers of offset parameter are initialized as zero, C3 layers
Convolution kernel kC3Quantity KernelNumberC3=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C4, C4 layers of output characteristic pattern quantity OutputMaps are enabledC4The width of=32, C4 layers of output characteristic pattern
OutputSizeC4=20, C4 layers of convolution kernel width KernelSizeC4=9, C4 layers of offset parameter are initialized as zero, C4 layers
Convolution kernel kC4Quantity KernelNumberC4=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C5, C5 layers of output characteristic pattern quantity OutputMaps are enabledC5The width of=32, C5 layers of output characteristic pattern
OutputSizeC5=4, C5 layers of convolution kernel width KernelSizeC5=7, C5 layers of offset parameter are initialized as zero, C5 layers
Convolution kernel kC5Quantity KernelNumberC5=1024, the initial value of each parameter is in convolution kernel
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1, S2,
The sampling core of S3 and S4 is initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output are special
Levy figure quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, exports characteristic pattern widthIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, and F1 layers of weighting parameter is water
Equal rights value parameter matrix W H and vertical weighting parameter matrix W V, size are 41 × 512, enable each of weighting parameter matrix
The initial value of parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV, just
Beginning turns to 41 × 1 one-dimensional null vector.
4. according to the method described in claim 3, it is characterized in that, step 1-5 the following steps are included:
Step 1-5-1, sub-network calculate probability vector: being extracted in a sub-network by the alternate treatment of convolutional layer and down-sampling layer
The image sequence characteristic of input is handled in classifier layer by Softmax function, and level probability vector HPV and vertical is obtained
Probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting layer
Convolution kernel, by the last piece image in input image sequence successively with VPV, HPV phase convolution, the output for obtaining propagated forward is pre-
Altimetric image.
5. according to the method described in claim 4, it is characterized in that, step 1-5-1 the following steps are included:
Step 1-5-1-1, judges network layer type: the network layer in the sub-network being presently in is indicated with l, l initial value is C1,
Judge the type of network layer l, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l ∈ S1,
S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCThe jth of layer
A output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are summed,
Summed result adds lCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveCalculation formula is such as
Shown in lower:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n are current convolution
The output characteristic pattern number of the previous down-sampling layer of layer, Indicate lCI-th of input feature vector of layer
Figure, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and return step 1-
5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained
The output characteristic pattern of convolutional layer respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lSLayer
Export characteristic patternCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolutional layer of current down-sampling layer,
Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, l is updated to l
+ 1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 will by matrixing
For the output characteristic pattern that the 32 width resolution ratio of C5 are 4 × 4 with column sequential deployment, the output for obtaining the F1 layer that resolution ratio is 512 × 1 is special
Levy vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1Apposition,
Calculated result is summed with Horizontal offset parameter BH, vertical off setting parameter BV respectively, obtains level after the processing of Softmax function
Probability vector HPV and vertical probability vector VPV, specific formula for calculation are as follows:
HPV=Softmax (WH × aF1+BH)
VPV=Softmax (WV × aF1+ BV),
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability vector VPV
Phase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern aDC1;
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1It is rolled up with level probability vector HPV phase
Product, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
6. according to the method described in claim 5, it is characterized in that, step 1-6 the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the training sample of the step 1-5-2-2 forecast image obtained and input
Control label in this asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term δ of level probability vectorHPVWith
The error term δ of vertical probability vectorVPV;
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith the mistake of vertical probability vector
Poor item δVPV, from rear to preceding classification the layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2, S1 of successively calculating
Error term, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the error term of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained
To the gradient value of this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the gradient value of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter
It is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, by former weight
Parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
7. according to the method described in claim 6, it is characterized in that, step 1-6-1 the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image that step 1-5-2-2 is obtained and this group of sample
Control label ask poor, obtain size be 240 × 240 error term matrix deltaDC2;
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2It expands
It is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability vector after overturning
Phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates zero extended function, and rot180 () indicates that angle is 180 ° of rotation function, by 2
× 2 matrix zero is extended for 4 × 4 matrix, the matrix after zero expansion, the region and original matrix phase one that central resolution ratio is 2 × 2
It causes, zero pixel filling of remaining position;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers of output
Characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term δ of HPVHPV, public
Formula is as follows:
δHPV=aDC1*δDC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Phase convolution, volume
41 × 1 column vector is obtained after product, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term δ of the step 1-6-1-3 probability vector obtainedVPVWith
δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by matrix
Apposition is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, by the mistake of the F1 layer obtained in step 1-6-2-1
Poor item δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term δC5,It indicates to become
The matrix that the 32nd resolution ratio after changing is 4 × 4;
Step 1-6-2-3, judges network layer type: the network layer in the sub-network being presently in is indicated with l, l initial value is S4,
Judge the type of network layer l, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l ∈ C4,
C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2,S3,
S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrixIt opens up
It opens up to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer of jth
A error term, and haveAgain by corresponding convolution kernel180 degree is rotated, then by the square after expansion
Battle array and the convolution nuclear phase convolution after overturning, and convolution results are summed, obtain lSI-th of error term matrix of layerFormula is such as
Shown in lower:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and return step 1-
6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3,C4,
C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCI-th of layer
Error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-sampling
In each element error entry value average mark into sampling area, obtaining resolution ratio isIt is upper
Sampling matrix, then activation primitive is calculated in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure, obtains lC
I-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, the corresponding up-sampling area of each of original image pixel after up-sampling
Domain in each of original pixel value mean allocation to sampling area pixel, successively calculates all error term matrixes, obtains
lCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates the mistake of a upper network layer
Poor item;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC∈
{ C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersCThe i-th of -1 layer
A output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the gradient of corresponding convolution kernel
ValueFormula is as follows:
lC∈{C1,C2,C3,C4,C5},
In above formula,WithRespectively indicate lCThe output characteristic pattern number and l of layerC- 1 layer
Characteristic pattern number is exported, ▽ k is gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC∈
{ C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of mistake of layer
Poor item matrixIn all elements sum, obtain j-th of this layer biasing gradient valueFormula is as follows:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculating separately level probability vector and vertical probability
The error term δ of vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH, WV
Gradient value, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is ladder of the error term to vertical weighting parameter
Angle value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability vector
Error term δHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV, formula
It is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is gradient of the error term to vertical off setting parameter
Value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term that step 1-6-3-1 is obtained is to convolution
The gradient of core is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and the amendment
Item asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term that step 1-6-3-2 is obtained is to biasing
Gradient be multiplied by the learning rates of dynamic convolutional neural networks, obtain the correction term of offset parameter, then by former bias term and the amendment
Item asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layer weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained to weighting parameter WH with
The gradient value of WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then by former weighting parameter WH
Ask poor with the correction term acquired respectively with WV, the WH and WV updated, formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layer offset parameter: the F1 layer error term that step 1-6-3-4 is obtained to offset parameter BH with
The gradient value of BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former offset parameter BH
Ask poor with the correction term acquired respectively with BV, the BH and BV updated, formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
8. the method according to the description of claim 7 is characterized in that it is characterized in that, step 2 the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image carry out at standardization
Reason, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, construction includes
The test sample collection of TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is trained dynamic
In state convolutional neural networks;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector HPVtestWith
Vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with VPVtest、
HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
9. according to the method described in claim 8, it is characterized in that, step 2-1 the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, between the time
It is divided into 6 minutes, altogether includes NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, when sampling
Last image is concentrated to meet the requirements amount of images by deleting test image;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original resolution to the image that sampling obtains
The floating number image that resolution ratio is 280 × 280 is converted into for 2000 × 2000 color image;
Step 2-1-3 constructs test sample collection: test sample collection is constructed using the floating number image set that step 2-1-2 is obtained, it will
Every four adjacent images in floating number image set, i.e., { 4M+1,4M+2,4M+3,4M+4 } width image is as one group of input sequence
Column, [4 × (M+1)+1] width image is by cutting, and the part that the central resolution ratio of reservation is 240 × 240 is as corresponding sample
Label is compareed, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test sample
Test sample collection;
Step 2-1-2 the following steps are included:
Image conversion: step 2-1-2-1 converts gray level image for colored echo strength CAPPI image, then is protected by cutting
Staying original image center resolution ratio is that 560 × 560 part obtains the image resolution ratio boil down to 280 × 280 after cutting
The grayscale image that resolution ratio is 280 × 280;
Step 2-1-2-2, data normalization: by the value of each of the grayscale image obtained in step 1-1-2-1 pixel from
[0~255] is mapped to [0~1], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: being extracted in a sub-network by the alternate treatment of convolutional layer and down-sampling layer
Then the image sequence characteristic of input is handled in classifier layer by Softmax function, level probability vector HPV is obtainedtest
With vertical probability vector VPVtest;
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probabilistic forecasting
Layer convolution kernel, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, obtains dynamic volume
The final extrapolated image of product neural network;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network layer p
Type, p initial value be C1, if p ∈ { C1, C2, C3, C4, C5 }, then p be convolutional layer, execute step 2-3-1-2, if p ∈
{ S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCThe v of layer
A output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then convolution results are asked
P is added with, summed resultCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCV-th of layer is defeated
Characteristic pattern outCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m are current convolution
The output characteristic pattern number of the previous down-sampling layer of layer, Indicate pCU-th of input feature vector of layer
Figure, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is input layer;
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and return step 2-
3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained
The output characteristic pattern of convolutional layer respectively withPhase convolution, then sampled with step-length for 2, sampling obtains pSLayer
Export characteristic patternCalculation formula is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolutional layer of current down-sampling layer,
Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, p is updated to
P+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 will by matrixing
For the output characteristic pattern that the 32 width resolution ratio of C5 are 4 × 4 with column sequential deployment, the output for obtaining the F1 layer that resolution ratio is 512 × 1 is special
Levy vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, calculated result is distinguished
With Horizontal offset parameter BH, vertical off setting parameter BV sum, summed result through Softmax function processing after obtain level probability to
Measure HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability vector
VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector HPVtest
Phase convolution, obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710110183.6A CN106886023B (en) | 2017-02-27 | 2017-02-27 | A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201710110183.6A CN106886023B (en) | 2017-02-27 | 2017-02-27 | A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks |
Publications (2)
Publication Number | Publication Date |
---|---|
CN106886023A CN106886023A (en) | 2017-06-23 |
CN106886023B true CN106886023B (en) | 2019-04-02 |
Family
ID=59180072
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201710110183.6A Active CN106886023B (en) | 2017-02-27 | 2017-02-27 | A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106886023B (en) |
Families Citing this family (32)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107748942B (en) * | 2017-11-24 | 2018-12-11 | 清华大学 | Radar Echo Extrapolation prediction technique and system based on velocity field sensing network |
CN109919295B (en) * | 2017-12-12 | 2022-10-28 | 北京大学深圳研究生院 | Embedded audio event detection method based on lightweight convolutional neural network |
CN108229404B (en) * | 2018-01-09 | 2022-03-08 | 东南大学 | Radar echo signal target identification method based on deep learning |
CN109754357B (en) | 2018-01-26 | 2021-09-21 | 京东方科技集团股份有限公司 | Image processing method, processing device and processing equipment |
CN108508505B (en) * | 2018-02-05 | 2020-12-15 | 南京云思创智信息科技有限公司 | Heavy rainfall and thunderstorm forecasting method and system based on multi-scale convolutional neural network |
CN108872993A (en) * | 2018-04-28 | 2018-11-23 | 中国人民解放军国防科技大学 | Radar echo extrapolation method based on cyclic dynamic convolution neural network |
CN108846409A (en) * | 2018-04-28 | 2018-11-20 | 中国人民解放军国防科技大学 | Radar echo extrapolation model training method based on cyclic dynamic convolution neural network |
CN108734357A (en) * | 2018-05-29 | 2018-11-02 | 北京佳格天地科技有限公司 | Weather prognosis system and method |
CN108427989B (en) * | 2018-06-12 | 2019-10-11 | 中国人民解放军国防科技大学 | Deep space-time prediction neural network training method for radar echo extrapolation |
CN109001736B (en) * | 2018-06-12 | 2022-04-05 | 中国人民解放军国防科技大学 | Radar echo extrapolation method based on deep space-time prediction neural network |
CN108648457B (en) * | 2018-06-28 | 2021-07-13 | 苏州大学 | Method, device and computer readable storage medium for speed prediction |
CN108732550B (en) * | 2018-08-01 | 2021-06-29 | 北京百度网讯科技有限公司 | Method and apparatus for predicting radar echo |
CN109283505B (en) * | 2018-09-03 | 2022-06-07 | 南京信息工程大学 | Method for correcting divergence phenomenon of radar echo extrapolated image |
CN109444863A (en) * | 2018-10-23 | 2019-03-08 | 广西民族大学 | A kind of estimation method of the narrowband ultrasonic echo number based on convolutional neural networks |
WO2020124948A1 (en) * | 2018-12-21 | 2020-06-25 | 中科寒武纪科技股份有限公司 | Network offline model processing method, artificial intelligence processing device, and related product |
CN111694617B (en) | 2018-12-29 | 2023-05-02 | 中科寒武纪科技股份有限公司 | Processing method of network offline model, artificial intelligence processing device and related products |
CN109884625B (en) * | 2019-02-22 | 2020-01-14 | 中国人民解放军军事科学院国防科技创新研究院 | Radar correlation imaging method based on convolutional neural network |
CN109871829B (en) * | 2019-03-15 | 2021-06-04 | 北京行易道科技有限公司 | Detection model training method and device based on deep learning |
CN110009150B (en) * | 2019-04-02 | 2019-11-19 | 中国石油大学(华东) | The compact reservoir physical parameter intelligent Forecasting of data-driven |
CN112116060B (en) * | 2019-06-21 | 2023-07-25 | 杭州海康威视数字技术股份有限公司 | Network configuration implementation method and device |
CN110568442B (en) * | 2019-10-15 | 2021-08-20 | 中国人民解放军国防科技大学 | Radar echo extrapolation method based on confrontation extrapolation neural network |
CN111062410B (en) * | 2019-11-05 | 2023-05-30 | 复旦大学 | Star information bridge weather prediction method based on deep learning |
CN111142109A (en) * | 2019-12-30 | 2020-05-12 | 上海眼控科技股份有限公司 | Marking method, marking device, computer equipment and storage medium |
CN111488887B (en) * | 2020-04-09 | 2023-04-18 | 腾讯科技(深圳)有限公司 | Image processing method and device based on artificial intelligence |
CN111507474B (en) * | 2020-06-18 | 2022-07-01 | 四川大学 | Neural network distributed training method for dynamically adjusting Batch-size |
CN111736157B (en) * | 2020-08-26 | 2021-01-05 | 蔻斯科技(上海)有限公司 | PPI data-based prediction method and device for nowcasting |
CN112232361B (en) * | 2020-10-13 | 2021-09-21 | 国网电子商务有限公司 | Image processing method and device, electronic equipment and computer readable storage medium |
CN113239722B (en) * | 2021-03-31 | 2022-08-30 | 成都信息工程大学 | Deep learning based strong convection extrapolation method and system under multi-scale |
CN113657477B (en) * | 2021-08-10 | 2022-04-08 | 南宁五加五科技有限公司 | Method, device and system for forecasting short-term rainfall |
CN113640769B (en) * | 2021-08-27 | 2023-06-13 | 南京信息工程大学 | Weather radar basic reflectivity prediction method based on deep neural network |
CN115390164B (en) * | 2022-10-27 | 2023-01-31 | 南京信息工程大学 | Radar echo extrapolation forecasting method and system |
CN115755227B (en) * | 2023-01-10 | 2023-04-14 | 南京信大气象科学技术研究院有限公司 | Three-dimensional radar extrapolation method based on deep neural network model |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104537387A (en) * | 2014-12-16 | 2015-04-22 | 广州中国科学院先进技术研究所 | Method and system for classifying automobile types based on neural network |
CN104811276A (en) * | 2015-05-04 | 2015-07-29 | 东南大学 | DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication |
CN106127725A (en) * | 2016-05-16 | 2016-11-16 | 北京工业大学 | A kind of millimetre-wave radar cloud atlas dividing method based on multiresolution CNN |
CN106228201A (en) * | 2016-06-20 | 2016-12-14 | 电子科技大学 | A kind of anti-Deceiving interference method of synthetic aperture radar based on shade characteristic |
Family Cites Families (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US8026840B2 (en) * | 2005-10-28 | 2011-09-27 | Raytheon Company | Biometric radar system and method for identifying persons and positional states of persons |
-
2017
- 2017-02-27 CN CN201710110183.6A patent/CN106886023B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN104537387A (en) * | 2014-12-16 | 2015-04-22 | 广州中国科学院先进技术研究所 | Method and system for classifying automobile types based on neural network |
CN104811276A (en) * | 2015-05-04 | 2015-07-29 | 东南大学 | DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication |
CN106127725A (en) * | 2016-05-16 | 2016-11-16 | 北京工业大学 | A kind of millimetre-wave radar cloud atlas dividing method based on multiresolution CNN |
CN106228201A (en) * | 2016-06-20 | 2016-12-14 | 电子科技大学 | A kind of anti-Deceiving interference method of synthetic aperture radar based on shade characteristic |
Also Published As
Publication number | Publication date |
---|---|
CN106886023A (en) | 2017-06-23 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106886023B (en) | A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks | |
Zhou et al. | Forecasting different types of convective weather: A deep learning approach | |
CN108846409A (en) | Radar echo extrapolation model training method based on cyclic dynamic convolution neural network | |
CN106203430B (en) | A kind of conspicuousness object detecting method based on foreground focused degree and background priori | |
Franz et al. | Ocean eddy identification and tracking using neural networks | |
CN106355151B (en) | A kind of three-dimensional S AR images steganalysis method based on depth confidence network | |
CN107527352A (en) | Remote sensing Ship Target contours segmentation and detection method based on deep learning FCN networks | |
CN108254741A (en) | Targetpath Forecasting Methodology based on Recognition with Recurrent Neural Network | |
CN107016357A (en) | A kind of video pedestrian detection method based on time-domain convolutional neural networks | |
CN107392130A (en) | Classification of Multispectral Images method based on threshold adaptive and convolutional neural networks | |
CN110163836A (en) | Based on deep learning for the excavator detection method under the inspection of high-altitude | |
CN113496104B (en) | Precipitation prediction correction method and system based on deep learning | |
CN103258324B (en) | Based on the method for detecting change of remote sensing image that controlled kernel regression and super-pixel are split | |
CN105741309B (en) | A kind of remote sensing image variation detection method based on the conversion of card side and samples selection | |
CN104680151B (en) | A kind of panchromatic remote sensing image variation detection method of high-resolution for taking snow covering influence into account | |
CN109657616A (en) | A kind of remote sensing image land cover pattern automatic classification method | |
CN103366371B (en) | Based on K distribution and the SAR image segmentation method of textural characteristics | |
CN107967474A (en) | A kind of sea-surface target conspicuousness detection method based on convolutional neural networks | |
CN112949407B (en) | Remote sensing image building vectorization method based on deep learning and point set optimization | |
CN102938073A (en) | Method for classifying remote sensing images | |
CN107607942A (en) | Based on the large scale electromagnetic scattering of deep learning model and the Forecasting Methodology of back scattering | |
CN109598711A (en) | A kind of thermal image defect extracting method based on feature mining and neural network | |
CN111383273B (en) | High-speed rail contact net part positioning method based on improved structure reasoning network | |
CN106845343A (en) | A kind of remote sensing image offshore platform automatic testing method | |
CN106407975B (en) | Multiple dimensioned layering object detection method based on space-optical spectrum structural constraint |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |