CN106886023B - A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks - Google Patents

A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks Download PDF

Info

Publication number
CN106886023B
CN106886023B CN201710110183.6A CN201710110183A CN106886023B CN 106886023 B CN106886023 B CN 106886023B CN 201710110183 A CN201710110183 A CN 201710110183A CN 106886023 B CN106886023 B CN 106886023B
Authority
CN
China
Prior art keywords
layer
image
error term
network
layers
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201710110183.6A
Other languages
Chinese (zh)
Other versions
CN106886023A (en
Inventor
李骞
施恩
顾大权
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
PLA University of Science and Technology
Original Assignee
PLA University of Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by PLA University of Science and Technology filed Critical PLA University of Science and Technology
Priority to CN201710110183.6A priority Critical patent/CN106886023B/en
Publication of CN106886023A publication Critical patent/CN106886023A/en
Application granted granted Critical
Publication of CN106886023B publication Critical patent/CN106886023B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/95Radar or analogous systems specially adapted for specific applications for meteorological use
    • GPHYSICS
    • G01MEASURING; TESTING
    • G01SRADIO DIRECTION-FINDING; RADIO NAVIGATION; DETERMINING DISTANCE OR VELOCITY BY USE OF RADIO WAVES; LOCATING OR PRESENCE-DETECTING BY USE OF THE REFLECTION OR RERADIATION OF RADIO WAVES; ANALOGOUS ARRANGEMENTS USING OTHER WAVES
    • G01S13/00Systems using the reflection or reradiation of radio waves, e.g. radar systems; Analogous systems using reflection or reradiation of waves whose nature or wavelength is irrelevant or unspecified
    • G01S13/88Radar or analogous systems specially adapted for specific applications
    • G01S13/89Radar or analogous systems specially adapted for specific applications for mapping or imaging
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Abstract

The invention discloses a kind of Radar Echo Extrapolation methods based on dynamic convolutional neural networks, it include: offline convolution neural metwork training: to given training image collection, training sample set is obtained by data prediction, initialize dynamic convolution neural network model, and using training sample set training dynamic convolutional neural networks, output valve is calculated by network propagated forward, the process of back-propagating update network parameter restrains dynamic convolutional neural networks.Online Radar Echo Extrapolation: test sample collection is converted for test chart image set by data prediction, trained dynamic convolutional neural networks are tested using test sample collection, by in input image sequence last width radar return image and network propagated forward in the probability vector phase convolution that obtains, the Radar Echo Extrapolation image predicted.

Description

A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks
Technical field
The invention belongs to surface weather observation technical fields in Atmospheric Survey, more particularly to one kind to be based on dynamic convolutional Neural The Radar Echo Extrapolation method of network.
Background technique
Nowcasting refers mainly to the weather forecast of 0~3 hour high-spatial and temporal resolution, and Main Prediction object includes strong drop The diastrous weathers such as water, strong wind, hail.Currently, many forecast systems all use Numerical Prediction Models, but due to numerical forecast In the presence of return slow (spin-up) has been forecast, Nowcasting ability is limited.New Generation Doppler Weather Radar has very high Sensitivity and resolution ratio, the spatial resolution of data information can reach 200~1000m, and temporal resolution can reach 2 ~15min.In addition, Doppler radar also has reasonable operating mode, comprehensive condition monitoring and fault warning, advanced Real-time Calibration System and Radar meteorology product algorithm abundant, the reliability of Nowcasting can be greatly improved.Nowadays, New Generation Doppler Weather Radar has become one of most effective tool of nowcasting, is faced using Doppler radar Nearly forecast is based primarily upon Radar Echo Extrapolation technology, i.e., according to current time radar observation result, thus it is speculated that radar return future Position and intensity, to realize the track prediction to strong convection system.
Traditional Radar Echo Extrapolation method is centroid tracking method and the cross-correlation technique based on maximum correlation coefficient (Tracking Radar Echoes by Correlation, TREC), but all there is certain deficiency, mass center in conventional method Tracing is only applicable to echo compared with strong, the lesser storm monomer of range, unreliable for the forecast of a wide range of precipitation;TREC is general Echo is considered as linear change, and echo variation is increasingly complex in reality, while such method is vulnerable in vector field Unordered vector disturbance.In addition, existing method is low to the utilization rate of Radar Data, and history Radar Data includes local weather system The important feature of system variation, has very high researching value.
For improve Radar Echo Extrapolation timeliness, and from a large amount of history Radar Data study radar return variation Machine learning method is introduced into Radar Echo Extrapolation by rule.Convolutional neural networks (Convolutional Neural Network, CNN) important branch as deep learning, it is widely used in image procossing, the fields such as pattern-recognition.The network is most Big feature is using part connection, weight is shared, down-sampling method, has to the deformation of input picture, translation and overturning There is stronger adaptability.For strong temporal correlation existing between radar return image, the dynamic convolution based on input is designed Neural network, which can dynamically change weighting parameter according to the radar echo map of input, and then predict extrapolated image.Benefit With history Radar Data training dynamic convolutional neural networks, network is made more fully to extract echo character, study echo variation Rule, for improving Radar Echo Extrapolation accuracy, optimization nowcasting effect is of great significance.
Summary of the invention
Goal of the invention: when the technical problem to be solved by the present invention is to be directed to the extrapolation of existing Radar Echo Extrapolation method Imitate it is short, it is insufficient to Radar Data utilization rate, propose a kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks, it is real Now to the high plane such as radar echo intensity show CAPPI (Constant AltitudePlan Position Indicator, CAPPI) the outside forecast of image, comprising the following steps:
Step 1, the offline convolutional neural networks of training: input training image collection carries out data prediction to training image collection, Training sample set is obtained, designs dynamic convolution neural network structure, and initialize network training parameter;It is assembled for training using training sample Practice dynamic convolutional neural networks, the orderly image sequence of input obtains width prediction by dynamic convolutional neural networks propagated forward Image calculates forecast image and compares the error between label, updates the weighting parameter of network by backpropagation and biasing is joined Number repeats this process until reaching trained termination condition, obtains convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, obtains Test sample collection, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by before network to biography Calculating probability vector is broadcast, and last width radar return image in input image sequence is mutually rolled up with obtained probability vector Product, the Radar Echo Extrapolation image predicted.
Step 1 of the present invention the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image standardize Change processing, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image Set is divided, and construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generate The sub-network of probability vector reconstructs the probabilistic forecasting layer for image extrapolation, provides for the offline neural metwork training stage dynamic State convolutional neural networks initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, training stage The sample size BatchSize=10 inputted every time, most large quantities of frequency of training of training sample setCurrently crowd frequency of training BatchNum=1, the maximum number of iterations of network training IterationMax=40, current iteration number IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, training the training sample obtained from step 1-1 every time It concentrates and reads BatchSize group training sample, every group of training sample is { x1,x2,x3,x4, y }, it altogether include 5 width images, wherein {x1,x2,x3,x4It is used as input image sequence, y is corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level Probability vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence according to It is secondary with VPV, HPV phase convolution, obtain the output forecast image of propagated forward;
Backpropagation: step 1-6 reversely acquires the error term of probability vector, according to probability vector in probabilistic forecasting layer Error term from rear to the preceding layer-by-layer error term for calculating each network layer in subnet network layers, and then calculate error in each network layer Item utilizes the parameter of obtained gradient updating dynamic convolutional neural networks to the gradient of weighting parameter and offset parameter;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following Three kinds of situations:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then step is returned Rapid 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current net Network the number of iterations is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, returns Step 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network changes Generation number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates offline neural metwork training rank Section, obtains trained dynamic convolution neural network model.
Step 1-1 data prediction of the present invention the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, when Between between be divided into 6 minutes, altogether include NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus, [N] indicates the maximum integer for being not more than N, acquires After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 1-1-3 constructs training sample set:
Training sample set is constructed using the floating number image set that step 1-1-2 is obtained, by every four in floating number image set Adjacent image, i.e. { 4N+1,4N+2,4N+3,4N+4 } width image are as one group of list entries, [4 × (N+1)+1] width figure It is cut as passing through, the part that the central resolution ratio of reservation is 240 × 240 is as the control label for corresponding to sample, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0, TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Image conversion: step 1-1-2-1 for the computation complexity for reducing convolutional neural networks training, step 1-1-1 is adopted The image that sample obtains is converted into gray level image, retains the part that original image center resolution ratio is 560 × 560 by cutting, will Image resolution ratio boil down to 280 × 280 after cutting obtains the grayscale image that resolution ratio is 280 × 280;
Step 1-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
Step 1-2 of the present invention the following steps are included:
Step 1-2-1 constructs sub-network:
Sub-network is made of 10 network layers, be followed successively by from front to back convolutional layer C1, down-sampling layer S1, convolutional layer C2, under Sample level S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer:
In probabilistic forecasting layer construct dynamic convolutional layer DC1 and dynamic convolutional layer DC2, by sub-network output vertical probability to Measure convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer, Convolution kernel k and offset parameter bias.For convolution kernel, it is thus necessary to determine that width KernelSize (point of convolution kernel of convolution kernel Resolution is KernelSize × KernelSize), the quantity KernelNumber of convolution kernel (value be the convolutional layer input with it is defeated The product of characteristic pattern quantity out), and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity with should The output characteristic pattern quantity of layer is identical.For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern width of this layer ForValue codetermined by its input feature vector figure resolution ratio and the size of convolution kernel, i.e.,Indicate convolutional layer lCUpper one layer of convolutional layer Output characteristic pattern width;
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1=12, export the width of characteristic pattern OutputSizeC1=272, convolution kernel width KernelSizeC1=9, offset parameter biasC1It is initialized as zero, C1 layers of volume Product core kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, C2 layers of output characteristic pattern quantity OutputMaps are enabledC2=32, export the width of characteristic pattern OutputSizeC2=128, convolution kernel width KernelSizeC2=9, offset parameter is initialized as zero, C2 layers of convolution kernel kC2Quantity KernelNumberC2=384, the initial value of each parameter is in convolution kernel
For convolutional layer C3, C3 layers of output characteristic pattern quantity OutputMaps are enabledC3=32, export the width of characteristic pattern OutputSizeC3=56, convolution kernel width KernelSizeC3=9, offset parameter is initialized as zero, C3 layers of convolution kernel kC3 Quantity KernelNumberC3=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C4, C4 layers of output characteristic pattern quantity OutputMaps are enabledC4=32, export the width of characteristic pattern OutputSizeC4=20, convolution kernel width KernelSizeC4=9, offset parameter is initialized as zero, C4 layers of convolution kernel kC4 Quantity KernelNumberC4=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C5, this layer of OutputMaps is enabledC5=32, OutputSizeC5=4, KernelSizeC5=7, partially Set the convolution kernel k that parameter is initialized as zero, C5 layersC5Quantity KernelNumberC5=1024, each ginseng in convolution kernel Several initial values are
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1, The sampling core of S2, S3 and S4 are initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output Characteristic pattern quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, output characteristic pattern is wide DegreeIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, F1 layers of weighting parameter For horizontal weighting parameter matrix W H and vertical weighting parameter matrix W V, size is 41 × 512, is enabled every in weighting parameter matrix The initial value of one parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV, It is initialized as 41 × 1 one-dimensional null vector.
Step 1-5 of the present invention the following steps are included:
Step 1-5-1, sub-network calculate probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network Extract input image sequence characteristic, handled in classifier layer by Softmax function, obtain level probability vector HPV with Vertical probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting The convolution kernel of layer obtains the defeated of propagated forward by the last piece image in input image sequence successively with VPV, HPV phase convolution Forecast image out.
Step 1-5-1 of the present invention the following steps are included:
Step 1-5-1-1, judges network layer type: indicating the network layer in the sub-network being presently in, l initial value with l For C1, the type of network layer l is judged, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l ∈ { S1, S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCLayer J-th of output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are asked L is added with, summed resultCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveIt calculates public Formula is as follows:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n is current The output characteristic pattern number of the previous down-sampling layer of convolutional layer, Indicate lCI-th of input of layer Characteristic pattern, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input Layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and returns to step Rapid 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lSLayer Output characteristic patternCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolution of current down-sampling layer Layer,Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, more by l It is newly l+1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 is become by matrix Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1 Export feature vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1's Apposition sums calculated result with Horizontal offset parameter BH, vertical off setting parameter BV respectively, after the processing of Softmax function To level probability vector HPV and vertical probability vector VPV, specific formula for calculation is as follows:
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to VPV phase convolution is measured, the DC1 layer that resolution ratio is 240 × 280 is obtained and exports characteristic pattern aDC1
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1With vertical probability vector HPV phase Convolution, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
Step 1-6 of the present invention the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the instruction of the step 1-5-2-2 forecast image obtained and input The control label practiced in sample asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term of level probability vector δHPVWith the error term δ of vertical probability vectorVPV
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith vertical probability vector Error term δVPV, from it is rear to preceding successively calculate classification layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2, The error term of S1, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the mistake of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained Gradient value of the poor item to this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the ladder of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter Angle value is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, will be former Weighting parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
Step 1-6-1 of the present invention the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image and the group that step 1-5-2-2 is obtained The control label of sample asks poor, obtains the error term matrix delta that size is 240 × 240DC2
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2 Expanding is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability after overturning Vector phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates that zero extended function, rot180 () indicate that angle is 180 ° of rotation letter 2 × 2 matrix zero is extended for 4 × 4 matrix by number, the matrix after zero expansion, the region and former square that central resolution ratio is 2 × 2 Consistent, zero pixel filling of remaining position of battle array;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers Export characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term of HPV δHPV, formula is as follows:
δHPV=aDC1DC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Mutually roll up It is long-pending, 41 × 1 column vector is obtained after convolution, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term of the step 1-6-1-3 probability vector obtained δVPVAnd δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by square The apposition of battle array is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, the F1 layer that will be obtained in step 1-6-2-1 Error term δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term δC5,Table Show that transformed 32nd resolution ratio is 4 × 4 matrix;Matrixing in this time matrixing and step 1-5-1-6 is each other Inverse transformation,
Step 1-6-2-3, judges network layer type: network layer (the l initial value in the sub-network being presently in is indicated with l For S4), the type of network layer l is judged, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l ∈ { C4, C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2, S3, S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrix It expands to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer J-th of error term, and haveAgain by corresponding convolution kernel180 degree is rotated, after then expanding Matrix and overturning after convolution nuclear phase convolution, and convolution results are summed, obtain lSI-th of error term matrix of layerIt is public Formula is as follows:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and returns to step Rapid 1-6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3, C4, C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCThe i-th of layer A error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-samplingIn each element error entry value average mark into sampling area, obtaining resolution ratio is Up-sampling matrix, then calculate activation primitive in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure, Obtain lCI-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, the corresponding up-sampling of each of original image pixel after up-sampling Region in each of original pixel value mean allocation to sampling area pixel, successively calculates all error term matrixes, obtains To lCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates a upper network layer Error term;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC ∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersC- 1 layer I-th of output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the ladder of corresponding convolution kernel Angle valueFormula is as follows:
In above formula,WithRespectively indicate lCThe output characteristic pattern number and l of layerC- 1 layer of output characteristic pattern number, ▽ k are gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC ∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of layer Error term matrixIn all elements sum, obtain j-th of this layer biasing gradient valueThe following institute of formula Show:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculate separately level probability vector with it is vertical The error term δ of probability vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH, The gradient value of WV, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is error term to vertical weighting parameter Gradient value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability to The error term δ of amountHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV, Formula is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is error term to vertical off setting parameter Gradient value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term pair that step 1-6-3-1 is obtained The gradient of convolution kernel is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and be somebody's turn to do Correction term asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term pair that step 1-6-3-2 is obtained The gradient of biasing is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former bias term and be somebody's turn to do Correction term asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layers of weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained is to weighting parameter The gradient value of WH and WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then former weight is joined Number WH and WV asks poor with the correction term acquired respectively, the WH and WV updated, and formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layers of offset parameter: the F1 layer error term that step 1-6-3-4 is obtained is to offset parameter The gradient value of BH and BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then original biasing is joined Number BH and BV asks poor with the correction term acquired respectively, the BH and BV updated, and formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
Step 2 of the present invention the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image standardize Change processing, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, constructs Test sample collection comprising TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is by training Dynamic convolutional neural networks in;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector HPVtestWith vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with VPVtest、HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
Step 2-1 of the present invention the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, when Between between be divided into 6 minutes, altogether include NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, is adopted Last image is concentrated to meet the requirements amount of images by deleting test image when sample;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 2-1-3 constructs test sample collection: constructing test sample using the floating number image set that step 2-1-2 is obtained Collection, by four adjacent images every in floating number image set, i.e. { 4M+1,4M+2,4M+3,4M+4 } width image is defeated as one group Enter sequence, for [4 × (M+1)+1] width image by cutting, retaining central resolution ratio is 240 × 240 part as corresponding sample This control label, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test specimens This test sample collection;
Step 2-1-2 the following steps are included:
Step 2-1-2-1, image conversion: to reduce the computation complexity that convolutional neural networks are trained, by colored echo Intensity CAPPI image is converted into gray level image, then retains the part that original image center resolution ratio is 560 × 560 by cutting, By the image resolution ratio boil down to 280 × 280 after cutting, the grayscale image that resolution ratio is 280 × 280 is obtained;
Step 2-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network The image sequence characteristic for extracting input, is then handled by Softmax function in classifier layer, obtains level probability vector HPVtestWith vertical probability vector VPVtest
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probability The convolution kernel of prediction interval, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, is moved The final extrapolated image of state convolutional neural networks;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network The type of layer p, p initial value is C1, if p ∈ { C1, C2, C3, C4, C5 }, then p is convolutional layer, step 2-3-1-2 is executed, if p ∈ { S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCLayer V-th of output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then by convolution knot Fruit summation, summed result add pCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCThe v of layer A output characteristic patternCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m are to work as The output characteristic pattern number of the previous down-sampling layer of preceding convolutional layer,Indicate pCU-th of layer Input feature vector figure, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is Input layer.
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and returns to step Rapid 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained The output characteristic pattern of the convolutional layer arrived then with step-length is respectively 2 to be sampled with phase convolution, and sampling obtains pS The output characteristic pattern calculation formula of layer is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolution of current down-sampling layer Layer,Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, by p It is updated to p+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 is become by matrix Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1 Export feature vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, by calculate tie Fruit is summed with Horizontal offset parameter BH, vertical off setting parameter BV respectively, and summed result obtains level after the processing of Softmax function Probability vector HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to Measure VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240, by extrapolated image It compares with the control label of corresponding test sample, determines that dynamic convolutional neural networks carry out the accuracy of Radar Echo Extrapolation.
The utility model has the advantages that being based on dynamic convolutional neural networks, trained on a large amount of radar return image set, and utilize Trained network carries out Radar Echo Extrapolation, effectively increases the accuracy of Radar Echo Extrapolation.
Specifically the present invention has the advantage that the utilization rate of 1. Radar Datas is high compared with existing method, with traditional thunder It is only compared up to echo extrapolation technique using a small amount of Radar Data, present invention training on a large amount of radar return image set, number It is higher according to utilization rate;2. timeliness of extrapolating is long, the time interval between sample can be concentrated outer to improve by adjusting training sample Push away timeliness.
Detailed description of the invention
The present invention is done with reference to the accompanying drawings and detailed description and is further illustrated, it is of the invention above-mentioned or Otherwise advantage will become apparent.
Fig. 1 is flow chart of the present invention.
Fig. 2 is dynamic convolutional neural networks initialization model structure chart.
Fig. 3 is sub-network structural map.
Fig. 4 is probabilistic forecasting layer structural map.
Fig. 5 is that matrix zero expands schematic diagram.
Fig. 6 is the process schematic up-sampled to 2 × 2 matrix.
Specific embodiment
The present invention will be further described with reference to the accompanying drawings and embodiments.
As shown in Figure 1, the present invention the following steps are included:
Step 1, as shown in Fig. 2, the offline convolutional neural networks of training: input training image collection carries out training image collection Data prediction obtains training sample set, designs dynamic convolution neural network structure, and initialize network training parameter;It utilizes Training sample set trains dynamic convolutional neural networks, and the orderly image sequence of input passes through dynamic convolutional neural networks propagated forward A width forecast image is obtained, forecast image is calculated and compares the error between label, the weight of network is updated by backpropagation Parameter and offset parameter repeat this process until reaching trained termination condition, obtain convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, obtains Test sample collection, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by before network to biography Calculating probability vector is broadcast, and last width radar return image in input image sequence is mutually rolled up with obtained probability vector Product, the Radar Echo Extrapolation image predicted.
Step 1 of the present invention the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image standardize Change processing, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image Set is divided, and construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generate The sub-network of probability vector reconstructs the probabilistic forecasting layer for image extrapolation, provides for the offline neural metwork training stage dynamic State convolutional neural networks initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, training stage The sample size BatchSize=10 inputted every time, most large quantities of frequency of training of training sample setCurrently crowd frequency of training BatchNum=1, the maximum number of iterations of network training IterationMax=40, current iteration number IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, training the training sample obtained from step 1-1 every time It concentrates and reads BatchSize group training sample, every group of training sample is { x1,x2,x3,x4, y }, it altogether include 5 width images, wherein {x1,x2,x3,x4It is used as input image sequence, x1Indicate that piece image, y are corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level Probability vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence according to It is secondary with VPV, HPV phase convolution, obtain the output forecast image of propagated forward;
Backpropagation: step 1-6 reversely acquires the error term of probability vector, according to probability vector in probabilistic forecasting layer Error term from rear to the preceding layer-by-layer error term for calculating each network layer in subnet network layers, and then calculate error in each network layer Item utilizes the parameter of obtained gradient updating dynamic convolutional neural networks to the gradient of weighting parameter and offset parameter;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following Three kinds of situations:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then step is returned Rapid 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current net Network the number of iterations is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, returns Step 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network changes Generation number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates offline neural metwork training rank Section, obtains trained dynamic convolution neural network model.
Step 1-1 data prediction of the present invention the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, when Between between be divided into 6 minutes, altogether include NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus, [N] indicates the maximum integer for being not more than N, acquires After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 1-1-3 constructs training sample set:
Training sample set is constructed using the floating number image set that step 1-1-2 is obtained, by every four in floating number image set Adjacent image, i.e. { 4N+1,4N+2,4N+3,4N+4 } width image are as one group of list entries, [4 × (N+1)+1] width figure It is cut as passing through, the part that the central resolution ratio of reservation is 240 × 240 is as the control label for corresponding to sample, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0, TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Image conversion: step 1-1-2-1 for the computation complexity for reducing convolutional neural networks training, step 1-1-1 is adopted The image that sample obtains is converted into gray level image, retains the part that original image center resolution ratio is 560 × 560 by cutting, will Image resolution ratio boil down to 280 × 280 after cutting obtains the grayscale image that resolution ratio is 280 × 280;
Step 1-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
Step 1-2 of the present invention the following steps are included:
Step 1-2-1 constructs sub-network: sub-network is configured to structure as shown in Figure 3:
Sub-network is made of 10 network layers, be followed successively by from front to back convolutional layer C1, down-sampling layer S1, convolutional layer C2, under Sample level S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer: probabilistic forecasting layer is configured to structure as shown in Figure 4:
In probabilistic forecasting layer construct dynamic convolutional layer DC1 and dynamic convolutional layer DC2, by sub-network output vertical probability to Measure convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer, Convolution kernel k and offset parameter bias.For convolution kernel, it is thus necessary to determine that width KernelSize (point of convolution kernel of convolution kernel Resolution is KernelSize × KernelSize), the quantity KernelNumber of convolution kernel (value be the convolutional layer input with it is defeated The product of characteristic pattern quantity out), and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity with should The output characteristic pattern quantity of layer is identical.For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern width of this layer ForValue codetermined by its input feature vector figure resolution ratio and the size of convolution kernel, i.e.,
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1=12, export the width of characteristic pattern OutputSizeC1=272, convolution kernel width KernelSizeC1=9, offset parameter biasC1It is initialized as zero, C1 layers of volume Product core kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, this layer of OutputMaps is enabledC2=32, OutputSizeC2=128, KernelSizeC2=9, Offset parameter is initialized as zero, C2 layers of convolution kernel kC2Quantity KernelNumberC2=384, each ginseng in convolution kernel Several initial values are
For convolutional layer C3, this layer of OutputMaps is enabledC3=32, OutputSizeC3=56, KernelSizeC3=9, partially Set the convolution kernel k that parameter is initialized as zero, C3 layersC3Quantity KernelNumberC3=1024, each ginseng in convolution kernel Several initial values are
For convolutional layer C4, this layer of OutputMaps is enabledC4=32, OutputSizeC4=20, KernelSizeC4=9, partially Set the convolution kernel k that parameter is initialized as zero, C4 layersC4Quantity KernelNumberC4=1024, each ginseng in convolution kernel Several initial values are
For convolutional layer C5, this layer of OutputMaps is enabledC5=32, OutputSizeC5=4, KernelSizeC5=7, partially Set the convolution kernel k that parameter is initialized as zero, C5 layersC5Quantity KernelNumberC5=1024, each ginseng in convolution kernel Several initial values are
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1, The sampling core of S2, S3 and S4 are initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output Characteristic pattern quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, output characteristic pattern is wide DegreeIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, F1 layers of weighting parameter For horizontal weighting parameter matrix W H and vertical weighting parameter matrix W V, size is 41 × 512, is enabled every in weighting parameter matrix The initial value of one parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV, It is initialized as 41 × 1 one-dimensional null vector.
Step 1-5 of the present invention the following steps are included:
Step 1-5-1, sub-network calculate probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network Extract input image sequence characteristic, handled in classifier layer by Softmax function, obtain level probability vector HPV with Vertical probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting The convolution kernel of layer obtains the defeated of propagated forward by the last piece image in input image sequence successively with VPV, HPV phase convolution Forecast image out.
Step 1-5-1 of the present invention the following steps are included:
Step 1-5-1-1, judges network layer type: indicating the network layer in the sub-network being presently in, l initial value with l For C1, the type of network layer l is judged, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l ∈ { S1, S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCLayer J-th of output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are asked L is added with, summed resultCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveIt calculates public Formula is as follows:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n is current The output characteristic pattern number of the previous down-sampling layer of convolutional layer,Indicate lCI-th of input of layer Characteristic pattern, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input Layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and returns to step Rapid 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lS The output characteristic pattern of layerCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolution of current down-sampling layer Layer,Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, more by l It is newly l+1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 is become by matrix Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1 Export feature vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1's Apposition sums calculated result with Horizontal offset parameter BH, vertical off setting parameter BV respectively, after the processing of Softmax function To level probability vector HPV and vertical probability vector VPV, specific formula for calculation is as follows:
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to VPV phase convolution is measured, the DC1 layer that resolution ratio is 240 × 280 is obtained and exports characteristic pattern aDC1
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1With vertical probability vector HPV phase Convolution, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
Step 1-6 of the present invention the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the instruction of the step 1-5-2-2 forecast image obtained and input The control label practiced in sample asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term of level probability vector δHPVWith the error term δ of vertical probability vectorVPV
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith vertical probability vector Error term δVPV, from it is rear to preceding successively calculate classification layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2, The error term of S1, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the mistake of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained Gradient value of the poor item to this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the ladder of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter Angle value is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, will be former Weighting parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
Step 1-6-1 of the present invention the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image and the group that step 1-5-2-2 is obtained The control label of sample asks poor, obtains the error term matrix delta that size is 240 × 240DC2
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2 Expanding is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability after overturning Vector phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates that zero extended function, rot180 () indicate that angle is 180 ° of rotation letter Number, by 2 × 2 matrix zero be extended for 4 × 4 process as shown in figure 5, zero expand after matrix, central resolution ratio is 2 × 2 Region is consistent with original matrix, zero pixel filling of remaining position;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers Export characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term of HPV δHPV, formula is as follows:
δHPV=aDC1DC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Mutually roll up It is long-pending, 41 × 1 column vector is obtained after convolution, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term of the step 1-6-1-3 probability vector obtained δVPVAnd δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by square The apposition of battle array is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, the F1 layer that will be obtained in step 1-6-2-1 Error term δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term Indicate that transformed 32nd resolution ratio is 4 × 4 matrix;This time matrixing and the matrixing in step 1-5-1-6 are mutual For inverse transformation,
Step 1-6-2-3, judges network layer type: network layer (the l initial value in the sub-network being presently in is indicated with l For S4), the type of network layer l is judged, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l ∈ { C4, C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2, S3, S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrix It expands to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer J-th of error term, and haveAgain by corresponding convolution kernel180 degree is rotated, after then expanding Matrix and overturning after convolution nuclear phase convolution, and convolution results are summed, obtain lSI-th of error term matrix of layerIt is public Formula is as follows:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and returns to step Rapid 1-6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3, C4, C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCThe i-th of layer A error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-samplingIn each element error entry value average mark into sampling area, obtaining resolution ratio is Up-sampling matrix, then calculate activation primitive in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure, Obtain lCI-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, and Fig. 6 is the process that is up-sampled of matrix to 2 × 2, after up-sampling Each of original image pixel corresponds to a up-sampling region, each of original pixel value mean allocation to sampling area picture In vegetarian refreshments, all error term matrixes are successively calculated, l is obtainedCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates a upper network layer Error term;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC ∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersC- 1 layer I-th of output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the ladder of corresponding convolution kernel Angle valueFormula is as follows:
In above formula,WithRespectively indicate lCLayer and lC- 1 layer of output characteristic pattern Number, ▽ k are gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC ∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of layer Error term matrixIn all elements sum, obtain j-th of this layer biasing gradient valueThe following institute of formula Show:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculate separately level probability vector with it is vertical The error term δ of probability vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH, The gradient value of WV, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is error term to vertical weighting parameter Gradient value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability to The error term δ of amountHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV, Formula is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is error term to vertical off setting parameter Gradient value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term pair that step 1-6-3-1 is obtained The gradient of convolution kernel is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and be somebody's turn to do Correction term asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term pair that step 1-6-3-2 is obtained The gradient of biasing is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former bias term and be somebody's turn to do Correction term asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layers of weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained is to weighting parameter The gradient value of WH and WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then former weight is joined Number WH and WV asks poor with the correction term acquired respectively, the WH and WV updated, and formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layers of offset parameter: the F1 layer error term that step 1-6-3-4 is obtained is to offset parameter The gradient value of BH and BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then original biasing is joined Number BH and BV asks poor with the correction term acquired respectively, the BH and BV updated, and formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
Step 2 of the present invention the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image standardize Change processing, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, constructs Test sample collection comprising TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is by training Dynamic convolutional neural networks in;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector HPVtestWith vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with VPVtest、HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
Step 2-1 of the present invention the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, when Between between be divided into 6 minutes, altogether include NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, is adopted Last image is concentrated to meet the requirements amount of images by deleting test image when sample;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original point to the image that sampling obtains The color image that resolution is 2000 × 2000 is converted into the floating number image that resolution ratio is 280 × 280;
Step 2-1-3 constructs test sample collection: constructing test sample using the floating number image set that step 2-1-2 is obtained Collection, by four adjacent images every in floating number image set, i.e. { 4M+1,4M+2,4M+3,4M+4 } width image is defeated as one group Enter sequence, for [4 × (M+1)+1] width image by cutting, retaining central resolution ratio is 240 × 240 part as corresponding sample This control label, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test specimens This test sample collection;
Step 2-1-2 the following steps are included:
Step 2-1-2-1, image conversion: to reduce the computation complexity that convolutional neural networks are trained, by colored echo Intensity CAPPI image is converted into gray level image, then retains the part that original image center resolution ratio is 560 × 560 by cutting, By the image resolution ratio boil down to 280 × 280 after cutting, the grayscale image that resolution ratio is 280 × 280 is obtained;
Step 2-1-2-2, data normalization: by each of the grayscale image obtained in step 1-1-2-1 pixel Value is mapped to [0~1] from [0~255], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: passing through the alternate treatment of convolutional layer and down-sampling layer in a sub-network The image sequence characteristic for extracting input, is then handled by Softmax function in classifier layer, obtains level probability vector HPVtestWith vertical probability vector VPVtest
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probability The convolution kernel of prediction interval, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, is moved The final extrapolated image of state convolutional neural networks;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network The type of layer p, p initial value is C1, if p ∈ { C1, C2, C3, C4, C5 }, then p is convolutional layer, step 2-3-1-2 is executed, if p ∈ { S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCLayer V-th of output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then by convolution results Summation, summed result add pCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCV-th of layer Export characteristic patternCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m is current The output characteristic pattern number of the previous down-sampling layer of convolutional layer,Indicate pCU-th of layer is defeated Enter characteristic pattern, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is defeated Enter layer.
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and returns to step Rapid 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained The output characteristic pattern of the convolutional layer arrived respectively withPhase convolution, then sampled with step-length for 2, sampling obtains pS The output characteristic pattern of layerCalculation formula is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolution of current down-sampling layer Layer,Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, by p It is updated to p+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 is become by matrix Change, by the 32 width resolution ratio of C5 be 4 × 4 output characteristic pattern with column sequential deployment, obtain the F1 layer that resolution ratio is 512 × 1 Export feature vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, will calculate As a result it sums respectively with Horizontal offset parameter BH, vertical off setting parameter BV, summed result obtains water after the processing of Softmax function Flat probability vector HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability to Measure VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240, by extrapolated image It compares with the control label of corresponding test sample, determines that dynamic convolutional neural networks carry out the accuracy of Radar Echo Extrapolation.
The present invention provides a kind of Radar Echo Extrapolation methods based on dynamic convolutional neural networks, implement the technology There are many method and approach of scheme, the above is only a preferred embodiment of the present invention, it is noted that for the art Those of ordinary skill for, various improvements and modifications may be made without departing from the principle of the present invention, these change It also should be regarded as protection scope of the present invention into retouching.The available prior art of each component part being not known in the present embodiment adds To realize.

Claims (9)

1. a kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks, which comprises the following steps:
Step 1, the offline convolutional neural networks of training: input training image collection carries out data prediction to training image collection, obtains Training sample set designs dynamic convolution neural network structure, and initializes network training parameter;It is dynamic using training sample set training The orderly image sequence of state convolutional neural networks, input obtains a width prognostic chart by dynamic convolutional neural networks propagated forward Picture calculates forecast image and compares the error between label, the weighting parameter and offset parameter of network are updated by backpropagation, This process is repeated until reaching trained termination condition, obtains convergent dynamic convolutional neural networks;
Step 2, online Radar Echo Extrapolation: input test image set carries out data prediction to test chart image set, is tested Sample set, then by the dynamic convolutional neural networks obtained in test sample collection input step 1, by network propagated forward meter Calculate probability vector, and by input image sequence last width radar return image and obtained probability vector phase convolution, obtain To the Radar Echo Extrapolation image of prediction;
Step 1 the following steps are included:
Step 1-1, data prediction: input training image collection, the every piece image concentrated to training image carry out at standardization Reason, converts every piece image to 280 × 280 floating number image, obtains floating number image collection, to floating number image collection It is divided, construction includes the training sample set of TrainsetSize group sample;
Step 1-2 initializes dynamic convolutional neural networks: design dynamic convolution neural network structure is configured to generating probability The sub-network of vector reconstructs the probabilistic forecasting layer for image extrapolation, provides dynamic volume for the offline neural metwork training stage Product neural network initialization model;
Step 1-3 initializes dynamic convolution Neural Network Training Parameter: enabling e-learning rate λ=0.0001, the training stage is each The sample size BatchSize=10 of input, most large quantities of frequency of training of training sample set Currently crowd frequency of training BatchNum=1, the maximum number of iterations IterationMax=40 of network training, current iteration number IterationNum=1;
Step 1-4 reads training sample: by the way of batch training, the training sample that training is obtained from step 1-1 every time is concentrated BatchSize group training sample is read, every group of training sample is { x1,x2,x3,x4, y }, it include altogether 5 width images, wherein { x1,x2, x3,x4It is used as input image sequence, y is corresponding control label;
Step 1-5, propagated forward: the input image sequence feature that extraction step 1-4 is obtained in a sub-network obtains level probability Vector HPV and vertical probability vector VPV;In probabilistic forecasting layer, by the last piece image in input image sequence successively with VPV, HPV phase convolution obtain the output forecast image of propagated forward;
Step 1-6, backpropagation: reversely acquiring the error term of probability vector in probabilistic forecasting layer, according to the mistake of probability vector The error term of poor Xiang Conghou each network layer into preceding layer-by-layer calculating subnet network layers, and then calculate error term pair in each network layer The gradient of weighting parameter and offset parameter utilizes the parameter of obtained gradient updating dynamic convolutional neural networks;
Off-line training step control: step 1-7 carries out whole control to the offline neural metwork training stage, is divided into following three kinds Situation:
If training sample is concentrated there are still original training sample, i.e. BatchNum < BatchMax, then return step 1-4 Continue to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and current network changes Generation number is less than maximum number of iterations, i.e. IterationNum < IterationMax then enables BatchNum=1, return step 1-4 continues to read BatchSize group training sample, carries out network training;
If training sample, which is concentrated, is not present original training sample, i.e. BatchNum=BatchMax, and network iteration time Number reaches maximum number of iterations, i.e. IterationNum=IterationMax, then terminates the offline neural metwork training stage, obtain To trained dynamic convolution neural network model.
2. the method according to claim 1, wherein step 1-1 data prediction the following steps are included:
Step 1-1-1, sampling: the image that training image is concentrated is sequentially arranged, and constant duration is distributed, between the time It is divided into 6 minutes, altogether includes NTrainWidth image determines TrainsetSize by following formula:
Wherein, Mod (NTrain, 4) and indicate NTrainTo 4 modulus,Expression is not more thanMaximum integer, acquire After TrainsetSize, training image is retained by sampling and concentrates preceding 4 × TrainsetSize+1 width image, when sampling, which passes through, deletes Except training image concentrates last image to meet the requirements amount of images;
Normalized images: step 1-1-2 carries out image transformation, normalization operation, by original resolution to the image that sampling obtains The floating number image that resolution ratio is 280 × 280 is converted into for 2000 × 2000 color image;
Step 1-1-3 constructs training sample set: training sample set is constructed using the floating number image set that step 1-1-2 is obtained, it will Every four adjacent images in floating number image set, i.e., { 4N+1,4N+2,4N+3,4N+4 } width image is as one group of input sequence Column, [4 × (N+1)+1] width image is by cutting, and the part that the central resolution ratio of reservation is 240 × 240 is as corresponding sample Label is compareed, for N group sampleIts make is as follows:
In above formula, G4N+1Indicate floating number image set in 4N+1 width image, N is positive integer, and have N ∈ [0, TrainsetSize-1], Crop () indicates trimming operation, and the portion that original image center size is 240 × 240 is retained after cutting Point, finally obtain the training sample set comprising TrainsetSize group training sample;
Wherein, step 1-1-2 the following steps are included:
Step 1-1-2-1, image conversion: converting gray level image for the step 1-1-1 image sampled, is retained by cutting The image resolution ratio boil down to 280 × 280 after cutting is divided in the part that original image center resolution ratio is 560 × 560 The grayscale image that resolution is 280 × 280;
Step 1-1-2-2, data normalization: by the value of each of the grayscale image obtained in step 1-1-2-1 pixel from [0~255] is mapped to [0~1], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization.
3. according to the method described in claim 2, it is characterized in that, step 1-2 the following steps are included:
Step 1-2-1 constructs sub-network:
Sub-network is made of 10 network layers, is followed successively by convolutional layer C1, down-sampling layer S1, convolutional layer C2, down-sampling from front to back Layer S2, convolutional layer C3, down-sampling layer S3, convolutional layer C4, down-sampling layer S5, convolutional layer C5 and classifier layer F1;
Step 1-2-2 constructs probabilistic forecasting layer:
Dynamic convolutional layer DC1 and dynamic convolutional layer DC2 is constructed in probabilistic forecasting layer, the vertical probability vector that sub-network is exported Convolution kernel of the VPV as dynamic convolutional layer DC1, convolution kernel of the level probability vector HPV as dynamic convolutional layer DC2;
Wherein, step 1-2-1 the following steps are included:
Step 1-2-1-1 constructs convolutional layer: determining the following contents: the output characteristic pattern quantity OutputMaps of convolutional layer, convolution Core k and offset parameter bias, for convolution kernel, it is thus necessary to determine that the width KernelSize of convolution kernel, the quantity of convolution kernel KernelNumber, the quantity of convolution kernel are the product that the convolutional layer inputs and exports characteristic pattern quantity, the resolution ratio of convolution kernel For KernelSize × KernelSize, and convolution kernel is constructed according to Xavier initial method;For offset parameter, quantity It is identical as the output characteristic pattern quantity of this layer;For convolutional layer lC,lC∈ { C1, C2, C3, C4, C5 }, the output characteristic pattern of this layer Width is Value by convolutional layer lCInput feature vector figure resolution ratio and convolution kernel widthIt codetermines, i.e.,
Indicate convolutional layer lCIt is upper one layer volume The output characteristic pattern width of lamination;
For convolutional layer C1, C1 layers of output characteristic pattern quantity OutputMaps is enabledC1The width of=12, C1 layers of output characteristic pattern OutputSizeC1=272, C1 layers of convolution kernel width KernelSizeC1=9, C1 layers of offset parameter biasC1It is initialized as zero, C1 layers of convolution kernel kC1Quantity KernelNumberC1=48, the initial value of each parameter is in convolution kernelRand () is for generating random number;
For convolutional layer C2, C2 layers of output characteristic pattern quantity OutputMaps are enabledC2The width of=32, C2 layers of output characteristic pattern OutputSizeC2=128, C2 layers of convolution kernel width KernelSizeC2=9, C2 layers of offset parameter are initialized as zero, C2 layers Convolution kernel kC2Quantity KernelNumberC2=384, the initial value of each parameter is in convolution kernel
For convolutional layer C3, C3 layers of output characteristic pattern quantity OutputMaps are enabledC3The width of=32, C3 layers of output characteristic pattern OutputSizeC3=56, C3 layers of convolution kernel width KernelSizeC3=9, C3 layers of offset parameter are initialized as zero, C3 layers Convolution kernel kC3Quantity KernelNumberC3=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C4, C4 layers of output characteristic pattern quantity OutputMaps are enabledC4The width of=32, C4 layers of output characteristic pattern OutputSizeC4=20, C4 layers of convolution kernel width KernelSizeC4=9, C4 layers of offset parameter are initialized as zero, C4 layers Convolution kernel kC4Quantity KernelNumberC4=1024, the initial value of each parameter is in convolution kernel
For convolutional layer C5, C5 layers of output characteristic pattern quantity OutputMaps are enabledC5The width of=32, C5 layers of output characteristic pattern OutputSizeC5=4, C5 layers of convolution kernel width KernelSizeC5=7, C5 layers of offset parameter are initialized as zero, C5 layers Convolution kernel kC5Quantity KernelNumberC5=1024, the initial value of each parameter is in convolution kernel
Step 1-2-1-2, construct down-sampling layer: in down-sampling layer do not include need training parameter, by down-sampling layer S1, S2, The sampling core of S3 and S4 is initialized asFor down-sampling layer lS,lS∈ { S1, S2, S3, S4 }, output are special Levy figure quantityIt is consistent with the output characteristic pattern quantity of one layer of convolutional layer thereon, exports characteristic pattern widthIt is the 1/2 of the output characteristic pattern width of one layer of convolutional layer thereon, formula is expressed as follows:
Step 1-2-1-3, structural classification device layer: classifier layer is made of a full articulamentum F1, and F1 layers of weighting parameter is water Equal rights value parameter matrix W H and vertical weighting parameter matrix W V, size are 41 × 512, enable each of weighting parameter matrix The initial value of parameter isOffset parameter is Horizontal offset parameter BH and vertical off setting parameter BV, just Beginning turns to 41 × 1 one-dimensional null vector.
4. according to the method described in claim 3, it is characterized in that, step 1-5 the following steps are included:
Step 1-5-1, sub-network calculate probability vector: being extracted in a sub-network by the alternate treatment of convolutional layer and down-sampling layer The image sequence characteristic of input is handled in classifier layer by Softmax function, and level probability vector HPV and vertical is obtained Probability vector VPV;
Step 1-5-2, calculate probabilistic forecasting layer and export image: the HPV and VPV that step 1-5-1 is obtained are as probabilistic forecasting layer Convolution kernel, by the last piece image in input image sequence successively with VPV, HPV phase convolution, the output for obtaining propagated forward is pre- Altimetric image.
5. according to the method described in claim 4, it is characterized in that, step 1-5-1 the following steps are included:
Step 1-5-1-1, judges network layer type: the network layer in the sub-network being presently in is indicated with l, l initial value is C1, Judge the type of network layer l, if l ∈ { C1, C2, C3, C4, C5 }, then l is convolutional layer, step 1-5-1-2 is executed, if l ∈ S1, S2, S3, S4 }, then l is down-sampling layer, executes step 1-5-1-4;
Step 1-5-1-2 handles convolutional layer: having l=l at this timeC,lC∈ { C1, C2, C3, C4, C5 }, first calculating lCThe jth of layer A output characteristic patternBy lCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, convolution results are summed, Summed result adds lCJ-th of offset parameter of layerIt handles, obtains using ReLU activation primitiveCalculation formula is such as Shown in lower:
Wherein,For lCI-th of input feature vector figure convolution kernel corresponding with j-th of output characteristic pattern of layer, n are current convolution The output characteristic pattern number of the previous down-sampling layer of layer, Indicate lCI-th of input feature vector of layer Figure, while being also lC- 1 layer of i-th of output characteristic pattern, * representing matrix convolution, if lC=C1, then lC- 1 layer is input layer;
All output characteristic patterns are successively calculated, l is obtainedCThe output characteristic pattern of layerL is updated to l+1, and return step 1- 5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-3 handles down-sampling layer: having l=l at this timeS,lS∈ { S1, S2, S3, S4 }, step 1-5-1-2 is obtained The output characteristic pattern of convolutional layer respectively withPhase convolution, then sampled with step-length for 2, sampling obtains lSLayer Export characteristic patternCalculation formula is as follows:
In above formula, Sample () indicates that step-length is 2 sampling processing, lS- 1 indicates the previous convolutional layer of current down-sampling layer, Indicate lSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain lSThe output characteristic pattern of layerAfterwards, l is updated to l + 1, and return step 1-5-1-1 judges network type, carries out the operation of next network layer;
Step 1-5-1-4 calculates F1 layers of probability vector: if network layer l is classifier layer, i.e. l=F1 will by matrixing For the output characteristic pattern that the 32 width resolution ratio of C5 are 4 × 4 with column sequential deployment, the output for obtaining the F1 layer that resolution ratio is 512 × 1 is special Levy vector aF1, calculate separately horizontal weighting parameter matrix W H and aF1Apposition, vertical weighting parameter matrix W V and aF1Apposition, Calculated result is summed with Horizontal offset parameter BH, vertical off setting parameter BV respectively, obtains level after the processing of Softmax function Probability vector HPV and vertical probability vector VPV, specific formula for calculation are as follows:
HPV=Softmax (WH × aF1+BH)
VPV=Softmax (WV × aF1+ BV),
By its vertical probability vector VPV transposition, final vertical probability vector is obtained;
Step 1-5-2 the following steps are included:
Step 1-5-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability vector VPV Phase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern aDC1
Step 1-5-2-2 predicts DC2 layers of vertical direction: by DC1 layers of output characteristic pattern aDC1It is rolled up with level probability vector HPV phase Product, obtains the output forecast image of propagated forward, and resolution ratio is 240 × 240.
6. according to the method described in claim 5, it is characterized in that, step 1-6 the following steps are included:
Step 1-6-1 calculates probabilistic forecasting layer error term: by the training sample of the step 1-5-2-2 forecast image obtained and input Control label in this asks poor, calculates the error term of DC2 layers, DC1 layers, finally acquires the error term δ of level probability vectorHPVWith The error term δ of vertical probability vectorVPV
Step 1-6-2 calculates sub-network error term: according to the error term δ of level probability vectorHPVWith the mistake of vertical probability vector Poor item δVPV, from rear to preceding classification the layer F1, convolutional layer C5, C4, C3, C2, C1 and down-sampling layer S4, S3, S2, S1 of successively calculating Error term, the resolution ratio of any layer error term matrix acquired and the output resolution ratio of characteristic pattern of this layer are consistent;
Step 1-6-3 calculates gradient: the error term of each network layer of sub-network is calculated according to the error term that step 1-6-2 is obtained To the gradient value of this layer of weighting parameter and offset parameter;
Step 1-6-4, undated parameter: by the gradient value of the weighting parameter of the step 1-6-3 each network layer obtained and offset parameter It is multiplied by the learning rate of dynamic convolutional neural networks, obtains the update item of each network layer weighting parameter and offset parameter, by former weight Parameter and offset parameter ask poor with the update item respectively, obtain updated weighting parameter and offset parameter.
7. according to the method described in claim 6, it is characterized in that, step 1-6-1 the following steps are included:
Step 1-6-1-1 calculates dynamic convolutional layer DC2 error term: the forecast image that step 1-5-2-2 is obtained and this group of sample Control label ask poor, obtain size be 240 × 240 error term matrix deltaDC2
Step 1-6-1-2 calculates dynamic convolutional layer DC1 error term: by zero padding by DC2 layers of error term matrix deltaDC2It expands It is 240 × 320, by level probability Vector rotation 180 degree, by the error term matrix after expansion and the level probability vector after overturning Phase convolution obtains DC1 layers of error term matrix deltaDC1, size is 240 × 280, and formula is as follows:
δDC1=Expand_Zero (δDC2) * rot180 (HPV),
Wherein, Expand_Zero () indicates zero extended function, and rot180 () indicates that angle is 180 ° of rotation function, by 2 × 2 matrix zero is extended for 4 × 4 matrix, the matrix after zero expansion, the region and original matrix phase one that central resolution ratio is 2 × 2 It causes, zero pixel filling of remaining position;
Step 1-6-1-3 calculates probability vector error term: the error term of level probability vector HPV is calculated, by DC1 layers of output Characteristic pattern and error term matrix deltaDC2Phase convolution obtains 1 × 41 row vector after convolution, which is the error term δ of HPVHPV, public Formula is as follows:
δHPV=aDC1DC2,
The error term for calculating vertical probability vector VPV, by the input feature vector figure of input layer and error term matrix deltaDC1Phase convolution, volume 41 × 1 column vector is obtained after product, which is the error term δ of VPVVPV, formula is as follows:
Wherein,For the last piece image in the input image sequence of training sample;
Step 1-6-2 the following steps are included:
Step 1-6-2-1 calculates classifier layer F1 error term: by the error term δ of the step 1-6-1-3 probability vector obtainedVPVWith δHPVWeighting parameter matrix W V vertical with F1 layers and horizontal weighting parameter matrix W H carry out matrix multiple respectively, then by matrix Apposition is summed and is averaged, and F1 layers of error term δ is obtainedF1, formula is as follows:
Wherein, × representing matrix apposition, ()TThe transposition for representing matrix, obtained δF1Size be 512 × 1;
Step 1-6-2-2 calculates convolutional layer C5 error term: by matrixing, by the mistake of the F1 layer obtained in step 1-6-2-1 Poor item δF1It is transformed to the matrix that 32 resolution ratio are 4 × 4Obtain C5 layers of error term δC5,It indicates to become The matrix that the 32nd resolution ratio after changing is 4 × 4;
Step 1-6-2-3, judges network layer type: the network layer in the sub-network being presently in is indicated with l, l initial value is S4, Judge the type of network layer l, if l ∈ { S4, S3, S2, S1 }, then l is down-sampling layer, step 1-6-2-4 is executed, if l ∈ C4, C3, C2, C1 }, then l is convolutional layer, executes step 1-6-2-5;
Step 1-6-2-4 calculates down-sampling layer error term: if l layers are down-sampling layer, there is l=l at this timeS,lS∈{S1,S2,S3, S4 }, for lSI-th of error term matrix of layerBy zero padding respectively by lS+ 1 layer of each error term matrixIt opens up It opens up to width and isWherein lSThe latter convolutional layer of+1 layer of current down-sampling layer of expression, j indicate lS+ 1 layer of jth A error term, and haveAgain by corresponding convolution kernel180 degree is rotated, then by the square after expansion Battle array and the convolution nuclear phase convolution after overturning, and convolution results are summed, obtain lSI-th of error term matrix of layerFormula is such as Shown in lower:
Wherein,
All error term matrixes are successively calculated, l is obtainedSThe output characteristic pattern of layerL is updated to l-1, and return step 1- 6-2-3 judges network type, calculates the error term of a upper network layer;
Step 1-6-2-5 calculates convolutional layer error term: if l layers are convolutional layer, there is l=l at this timeC,lC∈{C1,C2,C3,C4, C5 }, it is not in l since the initial value of l in step 1-6-2-3 is S4CThe case where=C5, for lCI-th of layer Error term matrixFirst to lCCorresponding i-th of error term matrix in+1 layerIt is up-sampled, it will when up-sampling In each element error entry value average mark into sampling area, obtaining resolution ratio isIt is upper Sampling matrix, then activation primitive is calculated in lCThe inner product of derivative and the up-sampling matrix acquired at layer character pair figure, obtains lC I-th of error term matrix of layerFormula is as follows:
Wherein, representing matrix inner product, ReLU'() indicate the derivative of ReLU activation primitive, form is as follows:
UpSample () indicates up-sampling function, the corresponding up-sampling area of each of original image pixel after up-sampling Domain in each of original pixel value mean allocation to sampling area pixel, successively calculates all error term matrixes, obtains lCThe output characteristic pattern of layer
Step 1-6-2-6, l layers are convolutional layer, i.e. l=l at this timeC, it is divided into two kinds of situations later:
If l ≠ C1, l is updated to l-1, and return step 1-6-2-3 judges network type, calculates the mistake of a upper network layer Poor item;
If l=C1, the calculating of step 1-6-2 sub-network error term terminates;
Step 1-6-3 the following steps are included:
Step 1-6-3-1 calculates convolutional layer error term to the gradient of convolution kernel: using lCIndicate currently processed convolutional layer, lC∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of convolution kernel, by l since C1 layersCThe i-th of -1 layer A output characteristic patternWith lCJ-th of error term matrix of layerPhase convolution, convolution results are the gradient of corresponding convolution kernel ValueFormula is as follows:
lC∈{C1,C2,C3,C4,C5},
In above formula,WithRespectively indicate lCThe output characteristic pattern number and l of layerC- 1 layer Characteristic pattern number is exported, ▽ k is gradient value of the error term to convolution kernel;
Step 1-6-3-2 calculates each convolutional layer error term to the gradient of biasing: using lCIndicate currently processed convolutional layer, lC∈ { C1, C2, C3, C4, C5 } successively calculates each convolutional layer error term to the gradient of biasing, by l since C1 layersCJ-th of mistake of layer Poor item matrixIn all elements sum, obtain j-th of this layer biasing gradient valueFormula is as follows:
Wherein, Sum () expression sums to all elements of matrix;
Step 1-6-3-3 calculates F1 layers of error term to the gradient of weighting parameter: calculating separately level probability vector and vertical probability The error term δ of vectorHPV、δVPVWith F1 layers of error term δF1Inner product, calculated result be F1 layers of error term to weighting parameter WH, WV Gradient value, formula are as follows:
▽ WH=(δHPV)T×(δF1)T,
▽ WV=δVPV×(δF1)T,
In above formula, ▽ WH is gradient value of the error term to horizontal weighting parameter, and ▽ WV is ladder of the error term to vertical weighting parameter Angle value;
Step 1-6-3-4 calculates F1 layers of error term to the gradient of offset parameter: by level probability vector and vertical probability vector Error term δHPV、δVPVRespectively as F1 layers of error term to the gradient value of Horizontal offset parameter BH and vertical off setting parameter BV, formula It is as follows:
▽ BH=(δHPV)T,
▽ BV=δVPV,
Wherein, ▽ BH is gradient value of the error term to Horizontal offset parameter, and ▽ BV is gradient of the error term to vertical off setting parameter Value;
Step 1-6-4 the following steps are included:
Step 1-6-4-1 updates each convolutional layer weighting parameter: each convolutional layer error term that step 1-6-3-1 is obtained is to convolution The gradient of core is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of convolution kernel, then by former convolution kernel and the amendment Item asks poor, the convolution kernel updatedFormula is as follows:
Wherein, λ is the e-learning rate determined in step 1-3, λ=0.0001;
Step 1-6-4-2 updates each convolutional layer offset parameter: each convolutional layer error term that step 1-6-3-2 is obtained is to biasing Gradient be multiplied by the learning rates of dynamic convolutional neural networks, obtain the correction term of offset parameter, then by former bias term and the amendment Item asks poor, the bias term updatedFormula is as follows:
Step 1-6-4-3, update F1 layer weighting parameter: the F1 layer error term that step 1-6-3-3 is obtained to weighting parameter WH with The gradient value of WV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of weighting parameter, then by former weighting parameter WH Ask poor with the correction term acquired respectively with WV, the WH and WV updated, formula is as follows:
WH=WH- λ ▽ WH,
WV=WV- λ ▽ WV;
Step 1-6-4-4, update F1 layer offset parameter: the F1 layer error term that step 1-6-3-4 is obtained to offset parameter BH with The gradient value of BV is multiplied by the learning rate of dynamic convolutional neural networks, obtains the correction term of offset parameter, then by former offset parameter BH Ask poor with the correction term acquired respectively with BV, the BH and BV updated, formula is as follows:
BH=BH- λ ▽ BH,
BV=BV- λ ▽ BV.
8. the method according to the description of claim 7 is characterized in that it is characterized in that, step 2 the following steps are included:
Step 2-1, data prediction: input test image set, the every piece image concentrated to test image carry out at standardization Reason, converts every piece image to 280 × 280 floating number image, then divide to floating number image collection, construction includes The test sample collection of TestsetSize group sample;
Step 2-2, read test sample: the TestsetSize group test sample input that step 2-1 is obtained is trained dynamic In state convolutional neural networks;
Step 2-3, propagated forward: extracting the image sequence characteristic of input in a sub-network, obtains level probability vector HPVtestWith Vertical probability vector VPVtest;In probabilistic forecasting layer, by the last piece image in input image sequence successively with VPVtest、 HPVtestPhase convolution obtains the final extrapolated image of dynamic convolutional neural networks.
9. according to the method described in claim 8, it is characterized in that, step 2-1 the following steps are included:
Step 2-1-1, sampling: the image that test image is concentrated is sequentially arranged, and constant duration is distributed, between the time It is divided into 6 minutes, altogether includes NTestWidth image determines TestsetSize by following formula:
If Mod (NTest, 4)=0
If Mod (NTest, 4) ≠ 0
After acquiring TestsetSize, test image is retained by sampling and concentrates preceding 4 × TestsetSize+1 width image, when sampling Last image is concentrated to meet the requirements amount of images by deleting test image;
Image normalization: step 2-1-2 carries out image transformation, normalization operation, by original resolution to the image that sampling obtains The floating number image that resolution ratio is 280 × 280 is converted into for 2000 × 2000 color image;
Step 2-1-3 constructs test sample collection: test sample collection is constructed using the floating number image set that step 2-1-2 is obtained, it will Every four adjacent images in floating number image set, i.e., { 4M+1,4M+2,4M+3,4M+4 } width image is as one group of input sequence Column, [4 × (M+1)+1] width image is by cutting, and the part that the central resolution ratio of reservation is 240 × 240 is as corresponding sample Label is compareed, wherein for positive integer, and there is M ∈ [0, TestsetSize-1] to obtain comprising TestsetSize group test sample Test sample collection;
Step 2-1-2 the following steps are included:
Image conversion: step 2-1-2-1 converts gray level image for colored echo strength CAPPI image, then is protected by cutting Staying original image center resolution ratio is that 560 × 560 part obtains the image resolution ratio boil down to 280 × 280 after cutting The grayscale image that resolution ratio is 280 × 280;
Step 2-1-2-2, data normalization: by the value of each of the grayscale image obtained in step 1-1-2-1 pixel from [0~255] is mapped to [0~1], by obtaining the floating number image that resolution ratio is 280 × 280 after normalization;
Step 2-3 the following steps are included:
Step 2-3-1 calculates sub-network probability vector: being extracted in a sub-network by the alternate treatment of convolutional layer and down-sampling layer Then the image sequence characteristic of input is handled in classifier layer by Softmax function, level probability vector HPV is obtainedtest With vertical probability vector VPVtest
Step 2-3-2 calculates probabilistic forecasting layer and exports image: the VPV that step 2-3-1 is obtainedtestAnd HPVtestAs probabilistic forecasting Layer convolution kernel, by the last piece image in input image sequence successively with VPVtestAnd HPVtestPhase convolution, obtains dynamic volume The final extrapolated image of product neural network;
Step 2-3-1 the following steps are included:
Step 2-3-1-1, judges network layer type: indicating the network layer in the sub-network being presently in p, judges network layer p Type, p initial value be C1, if p ∈ { C1, C2, C3, C4, C5 }, then p be convolutional layer, execute step 2-3-1-2, if p ∈ { S1, S2, S3, S4 }, then p is down-sampling layer, executes step 2-3-1-4;
Step 2-3-1-2 handles convolutional layer: having p=p at this timeC,pC∈ { C1, C2, C3, C4, C5 }, first calculating pCThe v of layer A output characteristic patternBy pCThe input feature vector figure convolution nuclear phase convolution corresponding with this layer respectively of layer, then convolution results are asked P is added with, summed resultCV-th of biasing of layerIt is handled using ReLU activation primitive, obtains pCV-th of layer is defeated Characteristic pattern outCalculation formula is as follows:
Wherein,For pCU-th of input feature vector figure convolution kernel corresponding with v-th of output characteristic pattern of layer, m are current convolution The output characteristic pattern number of the previous down-sampling layer of layer, Indicate pCU-th of input feature vector of layer Figure, while being also pC- 1 layer of u-th of output characteristic pattern, * representing matrix convolution work as pCWhen=C1, then pC- 1 layer is input layer;
All output characteristic patterns are successively calculated, p is obtainedCThe output characteristic pattern of layerP is updated to p+1, and return step 2- 3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-3 handles down-sampling layer: having p=p at this timeS,pS∈ { S1, S2, S3, S4 }, step 2-3-1-2 is obtained The output characteristic pattern of convolutional layer respectively withPhase convolution, then sampled with step-length for 2, sampling obtains pSLayer Export characteristic patternCalculation formula is as follows:
Wherein, Sample () indicates that step-length is 2 sampling processing, pS- 1 indicates the previous convolutional layer of current down-sampling layer, Indicate pSThe output characteristic pattern of layerIn j-th of output characteristic pattern, obtain pSThe output characteristic pattern of layerAfterwards, p is updated to P+1, and return step 2-3-1-1 judges network type, carries out the operation of next network layer;
Step 2-3-1-4 calculates F1 layers of probability vector: if network layer p is classifier layer, i.e. p=F1 will by matrixing For the output characteristic pattern that the 32 width resolution ratio of C5 are 4 × 4 with column sequential deployment, the output for obtaining the F1 layer that resolution ratio is 512 × 1 is special Levy vectorThen calculate separately horizontal parameters matrix W H, Vertical Parameters matrix W V withApposition, calculated result is distinguished With Horizontal offset parameter BH, vertical off setting parameter BV sum, summed result through Softmax function processing after obtain level probability to Measure HPVtest, vertical probability vector VPVtest, calculation formula is as follows:
By its vertical probability vector VPVtestTransposition obtains final vertical probability vector;
Step 2-3-2 the following steps are included:
Step 2-3-2-1 predicts DC1 layers of vertical direction: by last width input picture of input layer and vertical probability vector VPVtestPhase convolution obtains the DC1 layer that resolution ratio is 240 × 280 and exports characteristic pattern
Step 2-3-2-2 predicts DC2 layers of vertical direction: step 2-3-2-1 is obtainedWith level probability vector HPVtest Phase convolution, obtains the final extrapolated image of dynamic convolutional neural networks, and resolution ratio is 240 × 240.
CN201710110183.6A 2017-02-27 2017-02-27 A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks Active CN106886023B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201710110183.6A CN106886023B (en) 2017-02-27 2017-02-27 A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201710110183.6A CN106886023B (en) 2017-02-27 2017-02-27 A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks

Publications (2)

Publication Number Publication Date
CN106886023A CN106886023A (en) 2017-06-23
CN106886023B true CN106886023B (en) 2019-04-02

Family

ID=59180072

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201710110183.6A Active CN106886023B (en) 2017-02-27 2017-02-27 A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks

Country Status (1)

Country Link
CN (1) CN106886023B (en)

Families Citing this family (32)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107748942B (en) * 2017-11-24 2018-12-11 清华大学 Radar Echo Extrapolation prediction technique and system based on velocity field sensing network
CN109919295B (en) * 2017-12-12 2022-10-28 北京大学深圳研究生院 Embedded audio event detection method based on lightweight convolutional neural network
CN108229404B (en) * 2018-01-09 2022-03-08 东南大学 Radar echo signal target identification method based on deep learning
CN109754357B (en) 2018-01-26 2021-09-21 京东方科技集团股份有限公司 Image processing method, processing device and processing equipment
CN108508505B (en) * 2018-02-05 2020-12-15 南京云思创智信息科技有限公司 Heavy rainfall and thunderstorm forecasting method and system based on multi-scale convolutional neural network
CN108872993A (en) * 2018-04-28 2018-11-23 中国人民解放军国防科技大学 Radar echo extrapolation method based on cyclic dynamic convolution neural network
CN108846409A (en) * 2018-04-28 2018-11-20 中国人民解放军国防科技大学 Radar echo extrapolation model training method based on cyclic dynamic convolution neural network
CN108734357A (en) * 2018-05-29 2018-11-02 北京佳格天地科技有限公司 Weather prognosis system and method
CN108427989B (en) * 2018-06-12 2019-10-11 中国人民解放军国防科技大学 Deep space-time prediction neural network training method for radar echo extrapolation
CN109001736B (en) * 2018-06-12 2022-04-05 中国人民解放军国防科技大学 Radar echo extrapolation method based on deep space-time prediction neural network
CN108648457B (en) * 2018-06-28 2021-07-13 苏州大学 Method, device and computer readable storage medium for speed prediction
CN108732550B (en) * 2018-08-01 2021-06-29 北京百度网讯科技有限公司 Method and apparatus for predicting radar echo
CN109283505B (en) * 2018-09-03 2022-06-07 南京信息工程大学 Method for correcting divergence phenomenon of radar echo extrapolated image
CN109444863A (en) * 2018-10-23 2019-03-08 广西民族大学 A kind of estimation method of the narrowband ultrasonic echo number based on convolutional neural networks
WO2020124948A1 (en) * 2018-12-21 2020-06-25 中科寒武纪科技股份有限公司 Network offline model processing method, artificial intelligence processing device, and related product
CN111694617B (en) 2018-12-29 2023-05-02 中科寒武纪科技股份有限公司 Processing method of network offline model, artificial intelligence processing device and related products
CN109884625B (en) * 2019-02-22 2020-01-14 中国人民解放军军事科学院国防科技创新研究院 Radar correlation imaging method based on convolutional neural network
CN109871829B (en) * 2019-03-15 2021-06-04 北京行易道科技有限公司 Detection model training method and device based on deep learning
CN110009150B (en) * 2019-04-02 2019-11-19 中国石油大学(华东) The compact reservoir physical parameter intelligent Forecasting of data-driven
CN112116060B (en) * 2019-06-21 2023-07-25 杭州海康威视数字技术股份有限公司 Network configuration implementation method and device
CN110568442B (en) * 2019-10-15 2021-08-20 中国人民解放军国防科技大学 Radar echo extrapolation method based on confrontation extrapolation neural network
CN111062410B (en) * 2019-11-05 2023-05-30 复旦大学 Star information bridge weather prediction method based on deep learning
CN111142109A (en) * 2019-12-30 2020-05-12 上海眼控科技股份有限公司 Marking method, marking device, computer equipment and storage medium
CN111488887B (en) * 2020-04-09 2023-04-18 腾讯科技(深圳)有限公司 Image processing method and device based on artificial intelligence
CN111507474B (en) * 2020-06-18 2022-07-01 四川大学 Neural network distributed training method for dynamically adjusting Batch-size
CN111736157B (en) * 2020-08-26 2021-01-05 蔻斯科技(上海)有限公司 PPI data-based prediction method and device for nowcasting
CN112232361B (en) * 2020-10-13 2021-09-21 国网电子商务有限公司 Image processing method and device, electronic equipment and computer readable storage medium
CN113239722B (en) * 2021-03-31 2022-08-30 成都信息工程大学 Deep learning based strong convection extrapolation method and system under multi-scale
CN113657477B (en) * 2021-08-10 2022-04-08 南宁五加五科技有限公司 Method, device and system for forecasting short-term rainfall
CN113640769B (en) * 2021-08-27 2023-06-13 南京信息工程大学 Weather radar basic reflectivity prediction method based on deep neural network
CN115390164B (en) * 2022-10-27 2023-01-31 南京信息工程大学 Radar echo extrapolation forecasting method and system
CN115755227B (en) * 2023-01-10 2023-04-14 南京信大气象科学技术研究院有限公司 Three-dimensional radar extrapolation method based on deep neural network model

Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104537387A (en) * 2014-12-16 2015-04-22 广州中国科学院先进技术研究所 Method and system for classifying automobile types based on neural network
CN104811276A (en) * 2015-05-04 2015-07-29 东南大学 DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication
CN106127725A (en) * 2016-05-16 2016-11-16 北京工业大学 A kind of millimetre-wave radar cloud atlas dividing method based on multiresolution CNN
CN106228201A (en) * 2016-06-20 2016-12-14 电子科技大学 A kind of anti-Deceiving interference method of synthetic aperture radar based on shade characteristic

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8026840B2 (en) * 2005-10-28 2011-09-27 Raytheon Company Biometric radar system and method for identifying persons and positional states of persons

Patent Citations (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104537387A (en) * 2014-12-16 2015-04-22 广州中国科学院先进技术研究所 Method and system for classifying automobile types based on neural network
CN104811276A (en) * 2015-05-04 2015-07-29 东南大学 DL-CNN (deep leaning-convolutional neutral network) demodulator for super-Nyquist rate communication
CN106127725A (en) * 2016-05-16 2016-11-16 北京工业大学 A kind of millimetre-wave radar cloud atlas dividing method based on multiresolution CNN
CN106228201A (en) * 2016-06-20 2016-12-14 电子科技大学 A kind of anti-Deceiving interference method of synthetic aperture radar based on shade characteristic

Also Published As

Publication number Publication date
CN106886023A (en) 2017-06-23

Similar Documents

Publication Publication Date Title
CN106886023B (en) A kind of Radar Echo Extrapolation method based on dynamic convolutional neural networks
Zhou et al. Forecasting different types of convective weather: A deep learning approach
CN108846409A (en) Radar echo extrapolation model training method based on cyclic dynamic convolution neural network
CN106203430B (en) A kind of conspicuousness object detecting method based on foreground focused degree and background priori
Franz et al. Ocean eddy identification and tracking using neural networks
CN106355151B (en) A kind of three-dimensional S AR images steganalysis method based on depth confidence network
CN107527352A (en) Remote sensing Ship Target contours segmentation and detection method based on deep learning FCN networks
CN108254741A (en) Targetpath Forecasting Methodology based on Recognition with Recurrent Neural Network
CN107016357A (en) A kind of video pedestrian detection method based on time-domain convolutional neural networks
CN107392130A (en) Classification of Multispectral Images method based on threshold adaptive and convolutional neural networks
CN110163836A (en) Based on deep learning for the excavator detection method under the inspection of high-altitude
CN113496104B (en) Precipitation prediction correction method and system based on deep learning
CN103258324B (en) Based on the method for detecting change of remote sensing image that controlled kernel regression and super-pixel are split
CN105741309B (en) A kind of remote sensing image variation detection method based on the conversion of card side and samples selection
CN104680151B (en) A kind of panchromatic remote sensing image variation detection method of high-resolution for taking snow covering influence into account
CN109657616A (en) A kind of remote sensing image land cover pattern automatic classification method
CN103366371B (en) Based on K distribution and the SAR image segmentation method of textural characteristics
CN107967474A (en) A kind of sea-surface target conspicuousness detection method based on convolutional neural networks
CN112949407B (en) Remote sensing image building vectorization method based on deep learning and point set optimization
CN102938073A (en) Method for classifying remote sensing images
CN107607942A (en) Based on the large scale electromagnetic scattering of deep learning model and the Forecasting Methodology of back scattering
CN109598711A (en) A kind of thermal image defect extracting method based on feature mining and neural network
CN111383273B (en) High-speed rail contact net part positioning method based on improved structure reasoning network
CN106845343A (en) A kind of remote sensing image offshore platform automatic testing method
CN106407975B (en) Multiple dimensioned layering object detection method based on space-optical spectrum structural constraint

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant