CN113256536B - Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis - Google Patents
Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis Download PDFInfo
- Publication number
- CN113256536B CN113256536B CN202110682443.3A CN202110682443A CN113256536B CN 113256536 B CN113256536 B CN 113256536B CN 202110682443 A CN202110682443 A CN 202110682443A CN 113256536 B CN113256536 B CN 113256536B
- Authority
- CN
- China
- Prior art keywords
- neural network
- image data
- network
- wavelet
- data
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
- 238000000034 method Methods 0.000 title claims abstract description 31
- 238000013135 deep learning Methods 0.000 title claims abstract description 21
- 238000004458 analytical method Methods 0.000 title claims abstract description 10
- 238000013528 artificial neural network Methods 0.000 claims abstract description 46
- 238000012549 training Methods 0.000 claims abstract description 27
- 230000009466 transformation Effects 0.000 claims abstract description 15
- 238000005070 sampling Methods 0.000 claims abstract description 14
- 238000010586 diagram Methods 0.000 claims description 16
- 238000000354 decomposition reaction Methods 0.000 claims description 15
- 230000006870 function Effects 0.000 claims description 14
- 238000010606 normalization Methods 0.000 claims description 12
- 238000002790 cross-validation Methods 0.000 claims description 5
- 238000011156 evaluation Methods 0.000 claims description 5
- 238000007781 pre-processing Methods 0.000 claims description 4
- 238000004140 cleaning Methods 0.000 claims description 3
- 239000011159 matrix material Substances 0.000 claims description 3
- 230000035945 sensitivity Effects 0.000 claims description 3
- 230000002159 abnormal effect Effects 0.000 claims description 2
- 238000011478 gradient descent method Methods 0.000 claims description 2
- 238000013519 translation Methods 0.000 claims description 2
- 238000002372 labelling Methods 0.000 claims 2
- 239000000284 extract Substances 0.000 claims 1
- 238000005259 measurement Methods 0.000 claims 1
- 238000013508 migration Methods 0.000 claims 1
- 230000005012 migration Effects 0.000 claims 1
- 230000000644 propagated effect Effects 0.000 claims 1
- 230000011218 segmentation Effects 0.000 abstract description 9
- 230000008569 process Effects 0.000 abstract description 8
- 238000006243 chemical reaction Methods 0.000 abstract 1
- 108020001568 subdomains Proteins 0.000 abstract 1
- 238000011176 pooling Methods 0.000 description 12
- 238000002598 diffusion tensor imaging Methods 0.000 description 11
- 238000004422 calculation algorithm Methods 0.000 description 9
- 238000013527 convolutional neural network Methods 0.000 description 5
- 230000004913 activation Effects 0.000 description 3
- 230000003044 adaptive effect Effects 0.000 description 3
- 238000012545 processing Methods 0.000 description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 description 2
- 238000004364 calculation method Methods 0.000 description 2
- 230000009849 deactivation Effects 0.000 description 2
- 238000009792 diffusion process Methods 0.000 description 2
- 230000004927 fusion Effects 0.000 description 2
- 238000000691 measurement method Methods 0.000 description 2
- 238000003062 neural network model Methods 0.000 description 2
- 230000001902 propagating effect Effects 0.000 description 2
- 230000003042 antagnostic effect Effects 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 210000004556 brain Anatomy 0.000 description 1
- 230000006835 compression Effects 0.000 description 1
- 238000007906 compression Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000010191 image analysis Methods 0.000 description 1
- 238000002595 magnetic resonance imaging Methods 0.000 description 1
- 238000012986 modification Methods 0.000 description 1
- 230000004048 modification Effects 0.000 description 1
- 238000005457 optimization Methods 0.000 description 1
- 238000005381 potential energy Methods 0.000 description 1
- 238000006467 substitution reaction Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
- G06N3/084—Backpropagation, e.g. using gradient descent
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20048—Transform domain processing
- G06T2207/20064—Wavelet transform [DWT]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Data Mining & Analysis (AREA)
- Molecular Biology (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Artificial Intelligence (AREA)
- Evolutionary Computation (AREA)
- General Health & Medical Sciences (AREA)
- Biomedical Technology (AREA)
- Computing Systems (AREA)
- General Engineering & Computer Science (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a wavelet analysis-based ultrahigh-dimensional data reconstruction deep learning method, which utilizes high-dimensional high-order discrete wavelet packet transformation to expand high-dimensional data to different frequency domain channels, and combines a plurality of parallel neural networks to realize the reconstruction task of the high-dimensional data; data are preprocessed firstly, wavelet packet coefficients of sub-domains of different frequency bands are converted through wavelet packets, the wavelet packet coefficients are input into an independent network which is built and trained for the data, and the output of the network is subjected to wavelet packet inverse conversion to reconstruct an original image. The invention utilizes the independent property of each frequency domain of the high-dimensional data after wavelet transformation and the GPU memory in parallel, accelerates the training process of the neural network and makes the deep learning artificial task originally limited by hardware computing resources possible. The invention also extends to the segmentation and generation tasks. For the segmentation task, the output result of the U-net network is subjected to deconvolution up-sampling to form an original image resolution segmentation label. For the generation task, the neural network of each channel is changed to GAN.
Description
Technical Field
The invention relates to the technical field of image processing and neural networks, in particular to a super-high dimensional data reconstruction deep learning method based on wavelet analysis.
Background
In the field of image processing, the fast discrete wavelet transform applies a series of filters to expand image information into different independent frequency domain sub-bands and is represented by wavelet coefficients.
The CNN is the basis of a neural network in the field of processing images, the convolutional layer is the core of the CNN, and the detailed information of the images is extracted through a series of filters to generate a feature vector diagram. The pooling layer introduces invariance for CNN, and simultaneously performs down-sampling to expand the receptive field of the next layer of convolution kernel, so that the network can well learn the characteristic information of the image under different scales. The active layer is generally a nonlinear function, so that the network can better fit an arbitrary function, and the overfitting phenomenon of the ultra-deep network can be relieved.
U-net is a common image semantic segmentation and reconstruction model and is composed of an encoder and a decoder.
The GAN architecture is composed of a generator and a discriminator in cascade connection, and is widely applied to image generation tasks. The core idea is to make two networks compete with each other, which tends to reduce the discrimination accuracy of the discriminator from the generator point of view, i.e. strive to generate images similar to the Truth (Ground Truth) in order to be spurious; from the perspective of the discriminator, it tends to reduce the output image similarity of the generator, i.e., continuously improve the auditing standard. After the training is completed, the output of the discriminator is converged to a certain value, and dynamic balance is achieved.
In the prior art, DTI (Diffusion Tensor Imaging) is a method for describing a brain structure by multi-directional MRI scanning data, generally, scanning data of a tested individual is as large as several Gigabytes (GB), deep learning often requires a very large sample size, and data directly participating in training is about 100GB-200GB, but the memory capacity of the current deep learning chip is limited, for example, england tesla V100 only supports a memory capacity of 32GB, so that a large amount of DTI data cannot be directly stored in the memory of the deep learning chip for deep learning training.
Disclosure of Invention
In order to solve the defects of the prior art, overcome the problem that the deep learning network is difficult to support under the existing hardware conditions to train the ultrahigh-dimensional data and achieve the aim of training the large-volume ultrahigh-dimensional data by using limited computing resources, the invention adopts the following technical scheme:
as shown in fig. 1, a parallel depth network model based on wavelet transform includes the following specific steps;
s1, preprocessing data;
s11, data tag: tagging data collected by an individual;
s12, data enhancement: expanding the database by using translation, rotation, scaling and other modes;
s13, data cleaning: filling missing values, and checking abnormal objects;
s14, data normalization: a unified picture coding scheme is used.
S2, data frequency domain expansion;
s21, selecting proper wavelet basis functions according to different tasks and data;
s22, extracting low-frequency and high-frequency information of each dimension of image data by utilizing multi-order high-dimensional wavelet packet transformation, placing the information in different frequency domain sub-bands, and determining the number of decomposition layers according to the memory of a graphic processor and the data scale;
in the invention, high-dimensional N-order wavelet packet transformation is adopted, and the wavelet transformed numerical value, namely the wavelet basis coefficient which is used for fitting the original image space information, can be regarded as the projection value of the image information on each wavelet basis. After an original image is decomposed by a group of orthogonal wavelet bases, the original image is actually divided into a high-frequency component and a low-frequency component by a group of filters, on the basis, wavelet packet transformation carries out further decomposition on the high-frequency sub-band and the low-frequency sub-band of each level, and an optimal decomposition path is calculated by maximizing the information entropy of a step-by-step decomposition signal, so the wavelet packet decomposition is also called as an optimal sub-band tree structure. Taking three-dimensional image input as an example, the sub-blocks (wavelet packet decomposition coefficients) obtained by first-order wavelet packet decomposition are 1/8 of the original image size at the minimum, the decomposition path is optimized according to the maximum information entropy, sub-blocks which are not completely decomposed are also remained, the size of the sub-blocks is 1/4 of the original image size, and the decomposition mode is as shown in fig. 2. Wavelet packet inverse transformation can reconstruct the original high-dimensional image without distortion.
S3, building a network;
s31, respectively building independent neural networks for each frequency domain channel according to different tasks, building a U-net-based neural network for segmentation and reconstruction tasks, and building a GAN network model for generating tasks, so that the neural networks are trained in parallel in S4;
s32, building a U-net: the U-net is composed of the same number of encoders and decoders, each encoder comprises a convolutional layer and a downsampling layer, wherein each convolutional layer comprises a batch normalization layer, an active layer and a convolution kernel, the downsampling layer comprises a batch normalization layer, an active layer and a pooling layer, the pooling layer is used for downsampling image features, each decoder comprises a convolutional layer and an upsampling layer, each convolutional layer comprises a batch normalization layer, an active layer and a convolution kernel, the upsampling layer comprises a batch normalization layer, an active layer and a sampler, pixels are recovered through an inverse convolution algorithm or an interpolation algorithm, the batch normalization layer is used for normalizing data and calibration, and the active function is used for introducing a nonlinear factor to enable a neural network to be fitted to data distribution. Splicing multi-scale feature graphs by using Skip Connection among codecs at the same level, and ensuring the fusion of the multi-scale feature graphs, thereby improving the reconstruction and segmentation precision;
the U-net uniquely designs skip layer connection, splices the features before down-sampling into a feature map for recovering equivalent pixels, helps a decoder to lose as little detail information as possible in the process of recovering the pixels, an encoder comprises a plurality of convolution kernels and is responsible for extracting detail features under the receptive field to form a feature vector map, and a pooling layer introduces invariance and down-sampling images; the decoder also comprises a plurality of convolution kernels, partial feature vector graphs cut from the decoder at the same level are spliced through layer jump connection after the feature vector graphs are obtained, the situation that when the pixel is recovered by the up-sampler, the detail information is lost as little as possible is guaranteed, meanwhile, the image features learned by the encoder and the decoder are utilized, the pixel is recovered by the up-sampler corresponding to the pooling layer of the encoder, and a reverse convolution algorithm or an interpolation algorithm is usually adopted. Before each convolver and pooling layer, the U-net uses a batch normalization layer to normalize the data and calibrate the data distribution that may produce drift; the activation function typically uses modified linear units (relus) in order to introduce non-linear factors to better fit the network to the data distribution. The U-net network structure is shown in fig. 3.
S33, generating a countermeasure network GAN building: the GAN architecture is formed by cascading a generator and a discriminator, is a Convolutional Neural Network (CNN) architecture, the generator is an up-sampling convolutional network, is similar to a U-net decoder, inputs random signal distribution, outputs image data with the same size as a target image, and mainly comprises a convolutional layer and an up-sampling layer, the convolutional layer also comprises a batch normalization layer, an activation layer and a convolver, the up-sampling layer comprises an up-sampler, and outputs a high-dimensional image; the discriminator is a downsampled convolution network and comprises a convolution layer and a pooling layer, the convolution layer comprises a convolver, a batch normalization layer and an activation layer, the input is image data, and the output is one-dimensional probability E [0,1 ].
Before training the GAN network, the network weights of the generator and the discriminator need to be initialized. Firstly, network parameters of an initial generator are migrated and learned or randomized, then, false image data are generated by random seeds, a real image data training discriminator is mixed, and a network weight is updated. Usually, the number of database true and false data is equal, and the initialization is completed when the output of the discriminator is about 0.5. Then, random noise is used as the input of a generator, the generator and the discriminator are connected in series, the weight of the generator network is updated by using the back propagation error, the generator and the discriminator alternately rise in the repeated iterative training, and the training is continued until the stability is reached. The training process for GAN is shown in fig. 4.
S4, network training;
s41, initializing network parameters;
s42, calculating a loss function by using the labeled data (true value) and the network predicted value in training, updating the network weight by using residual back propagation algorithms such as random gradient descent, potential energy method (Momentum) or adaptive gradient (AdaGrad), adaptive Momentum estimation (Adam) and the like, specifically, according to the true value and the predicted value, calculating partial differential of each network parameter by the summation of a similar loss function and a cross entropy loss function, reversely propagating the residual from the predicted end to the input end among network layers, bringing the residual value of the previous layer into each network parameter, updating the partial differential value (gradient) and the residual value of each weight, updating each network weight by the partial differential value (gradient) and adopting a gradient descent method, and reducing the prediction error. The complete one-time forward propagation and backward propagation are one-time iteration, and the training process is shown in fig. 5 through a plurality of times of iteration optimization models;
s43, hyper-parameter adjustment: the overfitting problem is avoided by adjusting the Batch Size (Batch Size), the learning rate, the Momentum (Momentum), the Weight Decay (Weight Decay), the regularization coefficient, the random deactivation (Dropout) and the like.
S5, reconstructing data;
s51, after the neural network of the same level frequency domain channel is trained, performing wavelet packet inverse transformation on each channel network output (predicted value) and reconstructing upper level frequency domain information;
s6, evaluating the model;
s61, comparing the difference between the predicted value and the true value by Cross Validation (Cross Validation) for the network corresponding to each frequency domain channel, and adjusting the neural network structure and parameters of the frequency domain channel with poor performance;
s62, evaluation and measurement method of the model: similarity Coefficient (Dice coeffient), degree of overlap (IoU), Confusion Matrix (fusion Matrix) and its parameter Precision (Precision), Sensitivity (Sensitivity), Accuracy (Accuracy), receiver operating characteristic curve (ROC) and area under the curve, etc. (AUC).
The invention has the advantages and beneficial effects that:
the invention provides a set of deep learning method aiming at large-volume high-dimensional data by utilizing the independent property of each frequency domain of a high-dimensional image after wavelet transformation, and the invention effectively utilizes the GPU memory, remarkably accelerates the training process of a neural network and makes the deep learning artificial task originally limited by hardware computing resources possible. The method also extends to segmentation and generation tasks. For the segmentation task, the output result of the U-net network is subjected to deconvolution up-sampling to form an original image resolution segmentation label. For the image generation task, the neural network of each channel is changed to the antagonistic neural network GAN.
Drawings
FIG. 1 is a schematic flow chart of a reconstruction model of an ultra-high dimensional neural network according to the present invention.
Fig. 2 is a schematic diagram of wavelet packet transformation in the present invention.
Fig. 3 is a schematic diagram of a U-net network structure in the present invention.
Fig. 4 is a schematic diagram of the GAN network training procedure in the present invention.
FIG. 5 is a schematic diagram of single-channel neural network training in the present invention.
FIG. 6 is a block diagram of an exemplary system for reconstructing DTI data according to the present invention.
Fig. 7 is a schematic diagram of a wavelet packet decomposition process according to the present invention.
FIG. 8 is a schematic diagram of the top layer codec of the first-order tile input U-net of the present invention
FIG. 9 is a schematic diagram of the top layer codec of the two-level tile input U-net according to the present invention.
Detailed Description
The following detailed description of embodiments of the invention refers to the accompanying drawings. It should be understood that the detailed description and specific examples, while indicating the present invention, are given by way of illustration and explanation only, not limitation.
The invention provides a multi-frequency domain parallel neural network, creatively provides that ultra-high dimensional data (such as video data, diffusion weighted magnetic resonance data, functional magnetic resonance data and the like) are decomposed into different frequency sub-bands (sub-bands) by utilizing wavelet packet transformation, and an independent neural network is built for each sub-domain, so that the tasks of segmenting, generating, reconstructing and the like of high dimensional image data are completed. The method can be applied to deep learning tasks of large-volume ultrahigh-dimensional data input, such as medical image analysis, video analysis and the like. Note: each (wavelet transform-neural network-inverse wavelet transform) is referred to as a frequency domain channel or channel.
Because different frequency domain channels after transformation have certain independence, the original image can be perfectly recovered after the image passes through the neural network, and the error only comes from the prediction error of each neural network. In short, the invention provides a method for parallel learning of lossless compression large-volume data in deep learning, in the process, ultrahigh-dimensional large-volume data is decomposed into a plurality of channels, a neural network model in each channel can be trained independently, the calculation and memory resources of a graphic processor required by data training are greatly reduced, and finally, the output of each network reconstructs superior frequency domain information through wavelet packet inverse transformation to restore ultrahigh-dimensional data information.
A deep learning framework which has no information loss and can decompose original data and train independently in a sub-frequency domain solves the problem that deep learning is not suitable for ultra-high-dimensional large-volume data to a certain extent. As shown in fig. 6, a flow of steps for reconstructing DTI under the technical framework provided by the present invention is described.
1-1) data enhancement: circularly translating the DTI data along x, y and z axes by 2-10 voxel points (voxel) to obtain a new data expansion database;
1-2) data normalization: and aligning the direction (b-vector) of a diffusion sensitive field by using a nearest neighbor method, and extracting a corresponding scanning image.
2-1) selecting a multi-bezier (Daubechies) wavelet basis;
2-2) the database is 290 DTI data, one DTI data comprises four dimensions of length, width, height and direction, and the size is 148 multiplied by 176 multiplied by 148 multiplied by 288. The wavelet packet transform selects an optimal decomposition path according to the maximum information entropy principle, and a DTI object is decomposed into 1 first-order sub-tile block (first-order wavelet packet coefficient) with the size of 74 × 88 × 74 × 288 and 56 second-order sub-tile blocks (second-order wavelet packet coefficient) with the size of 37 × 44 × 37 × 288 after second-order wavelet packet transform. The decomposition order is automatically determined by a wavelet packet decomposition algorithm. Particularly, the space occupied by a single data sample is correspondingly reduced from 3.92GB to 0.49GB and 61.2MB, and a deep neural network model with about 64 or 128 data sample objects and about 400 ten thousand parameters can be trained by simultaneously loading the data sample objects under the display memory of a 24GB graphics processor. The wavelet packet decomposition process for a DTI object is shown in fig. 7.
3-1) building 8 independent networks based on U-net, and training in parallel in step 4;
3-2) building U-net: the U-net is composed of 6 encoders and decoders, each encoder convolution layer and Pooling layer, the convolution layer includes two convolution calculations, the convolution kernel is 3 × 3 × 3 × 3, the zero Padding (Padding) value is 1, the step size (Stride) is 1, besides, a batch layer and a modified linear unit are used, the Pooling layer uses a 2 × 2 × 2 × 2 maximum Pooling algorithm (Max Pooling), the input 1 data of 37 × 44 × 37 × 288 is output by the convolution layer to obtain Feature maps (Feature maps) with the size of 64, 37 × 44 × 37 × 288, and the Feature maps with the size of 64, 19 × 22 × 19 × 144 are obtained by the Pooling layer; each decoder comprises a convolution layer and an upper sampling layer of convolution kernel, the convolution parameters of the convolution layer are identical to those of the encoder, the upper sampling layer uses a bilinear interpolation algorithm, the input of the decoder at the uppermost layer is 64 characteristic diagrams of 19 × 22 × 19 × 144, 64 outputs of 37 × 44 × 37 × 288 obtained by the upper sampler, the encoder at the same level and the decoder are connected by using a layer jump, the characteristic diagram output by the convolution layer of the encoder is partially fused to the input part of the convolution layer of the decoder, 32 output characteristic diagrams of the encoder at the uppermost layer are randomly selected and spliced to the input end of the convolution layer of the decoder at the uppermost layer, the input of the convolution layer of the decoder is 96 characteristic diagrams of 37 × 44 × 37 × 288, and a reconstructed DTI second-order sub-tile block (second-order wavelet packet coefficient) with the original size of 37 × 44 × 37 × 288 is obtained after the convolution layer of the final decoder; as shown in fig. 8 and 9, the composition of the uppermost codec is described, and the input and output of each layer are described when the input is a first-order sub-block (first-order wavelet packet coefficients) and a second-order sub-block (second-order wavelet packet coefficients), respectively.
4-1) the step is explained as a training method of a single channel, and 8 independent networks are the same and can be respectively and independently trained;
4-2) randomly initializing network parameters by using standard normal distribution;
4-3) forward propagating the label data through the network, and obtaining two loss functions by using the truth value and the network predicted value: adding a similar Loss function (Dice Loss) and a Cross entropy Loss function (Cross-entropy Loss), obtaining a gradient by using a back propagation algorithm, updating a network weight by using an adaptive momentum estimation (Adam) optimizer, and optimizing the model through multiple iterations until the model converges;
4-3) adjusting the hyper-parameters: the learning rate was set to 0.0005 in the first 50 iterations and 0.0001 in the last 30 iterations, the weight decay was 0.01, and the random deactivation rate was 0.5.
5-1) after the neural network of the same level frequency domain channel is trained, performing wavelet packet inverse transformation on the neural network output (each channel predicted value) of each frequency domain channel, and reconstructing upper level frequency domain information to obtain reconstructed high-dimensional data.
6-1) comparing the difference between a predicted value and a true value by Cross Validation (Cross Validation) for networks corresponding to all frequency domain channels (namely the same frequency domain channel), and pertinently adjusting the neural network structure and parameters of the frequency domain channels with poor performance;
6-2) evaluation and measurement method of model: similarity factor, degree of overlap.
The above examples are only intended to illustrate the technical solution of the present invention, but not to limit it; although the present invention has been described in detail with reference to the foregoing embodiments, it will be understood by those of ordinary skill in the art that: the technical solutions described in the foregoing embodiments may still be modified, or some or all of the technical features may be equivalently replaced; and the modifications or the substitutions do not make the essence of the corresponding technical solutions depart from the scope of the technical solutions of the embodiments of the present invention.
Claims (10)
1. A super-high dimensional data reconstruction deep learning method based on wavelet analysis is characterized by comprising the following steps:
s1, preprocessing the ultrahigh-dimensional image data;
s2, the frequency domain expansion of the ultra-high dimensional image data comprises the following steps:
s21, selecting a wavelet basis;
s22, extracting frequency information of different dimensions of the ultra-high dimensional image data by using N-order high-dimensional wavelet packet transformation, decomposing the ultra-high dimensional image data into N values which are transformed by the wavelet packet, namely, decomposing wavelet basis coefficients fitting image space information, and placing the N values in different frequency domain sub-bands, wherein the number of decomposing layers is determined according to the memory of the image processor and the data scale;
s3, building a neural network, and building independent neural networks for each frequency domain channel;
s4, training a neural network, and performing parallel training on the neural network corresponding to each frequency domain channel;
and S5, reconstructing data, namely inputting the decomposed ultrahigh-dimensional image data into trained neural networks in the same-level frequency domain channels, performing wavelet packet inverse transformation on the output of each channel neural network, and reconstructing upper-level frequency domain information.
2. The method as claimed in claim 1, wherein in S22, the image data is decomposed into high frequency components and low frequency components by a set of orthogonal wavelet bases, and then divided into high frequency components and low frequency components by a set of filters, and based on the high frequency components and the low frequency components, the wavelet packet transform further decomposes the high frequency sub-bands and the low frequency sub-bands of each level, and the optimal decomposition path is calculated by maximizing the entropy of the signals decomposed in stages.
3. The method for deep learning based on wavelet analysis for ultrahigh-dimensional data reconstruction as claimed in claim 1, wherein S3 is implemented to build a U-net neural network, the U-net neural network comprises a set of pairs of encoder and decoder corresponding to each other, and between peer encoders and decoders, using layer-hopping connection to concatenate multi-scale feature maps, and concatenating image features before downsampling by the encoders into feature maps of equivalent pixels restored by the decoders.
4. The method as claimed in claim 3, wherein the encoder includes an encoder convolutional layer and a down-sampling layer, the convolutional core of the encoder convolutional layer extracts detail features under its receptive field to form a feature vector diagram, and the down-sampling layer introduces invariance and down-samples image features; the decoder comprises a decoder convolution layer and an upper sampling layer, after the feature vector diagram is obtained by the convolution kernel of the decoder convolution layer, the feature vector diagram is spliced by layer jump connection, and the pixels are recovered by the upper sampling layer.
5. The method for deep learning based on wavelet analysis and reconstruction of ultrahigh-dimensional data according to claim 1, wherein S3 is used for constructing a GAN neural network, the GAN neural network comprises a cascade generator and a discriminator, the generator is an upsampled convolutional network, the discriminator is a downsampled convolutional network, one end of the discriminator is used for respectively acquiring a target image and image data which is output by the generator and has the same size as the target image, and is used for training the discriminator, and the other end of the discriminator is used for training the generator with a one-dimensional probability e [0,1] output by the discriminator.
6. The method for ultrahigh-dimensional data reconstruction deep learning based on wavelet analysis as recited in claim 1, wherein said S4 comprises the steps of:
s41, initializing neural network parameters;
s42, the marked image data is propagated forward through a neural network to obtain a neural network prediction value, a loss function is calculated according to the prediction value and the true value, the loss function is utilized to carry out backward propagation to obtain a gradient, a gradient descent method is adopted to update a network weight, and the neural network is optimized through multiple iterations until the neural network converges;
and S43, adjusting the neural network hyper-parameters to prevent the neural network from being over-fitted.
7. The method of claim 6, wherein in step S4, a U-net neural network is trained, the U-net neural network includes a set of encoder and decoder pairs corresponding to each other, the original image data is input into the encoder, a prediction value is obtained from an output of the corresponding decoder, and the loss function is calculated according to the prediction value and a true value corresponding to the original image.
8. The method according to claim 6, wherein in step S4, a GAN neural network is trained, the GAN neural network comprises cascaded generators and classifiers, before training, network weights of the generators and classifiers are initialized, false image data is generated by using random seeds after network parameters of the generators are initialized through migration learning and/or randomization, the classifiers are trained in a true value and updated, then random noise is used as an input of the generators, the network weights of the generators are updated through back propagation errors according to a true value and a predicted value output by the classifiers, and the performances of the generators and the classifiers alternately rise in repeated iterative training and training is continued until errors between the true value and the predicted value tend to be stable.
9. The method according to claim 1, wherein after S5, model evaluation is performed to compare the difference between the predicted value and the true value for the neural network corresponding to each frequency domain channel by cross validation, and the neural network structure and parameters of the frequency domain channel are adjusted according to the difference; the evaluation and measurement modes comprise: similarity coefficients, degree of overlap, confusion matrix and its parameter accuracy, sensitivity, accuracy, receiver operating characteristics curve and area under the curve.
10. The method for ultrahigh-dimensional data reconstruction deep learning based on wavelet analysis as recited in claim 1, wherein in said S1, the image data preprocessing includes: image data labeling, image data enhancement, image data cleaning and image data normalization;
the image data label is used for labeling the collected image data;
the image data enhancement: expanding the image database, wherein the expansion mode comprises translation, rotation and scaling of image data;
and image data cleaning: filling missing values in the image data, and checking an abnormal object;
the image data normalization: a unified picture coding scheme is used.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110682443.3A CN113256536B (en) | 2021-06-18 | 2021-06-18 | Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110682443.3A CN113256536B (en) | 2021-06-18 | 2021-06-18 | Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113256536A CN113256536A (en) | 2021-08-13 |
CN113256536B true CN113256536B (en) | 2021-11-23 |
Family
ID=77188734
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110682443.3A Active CN113256536B (en) | 2021-06-18 | 2021-06-18 | Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113256536B (en) |
Families Citing this family (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114488217B (en) * | 2021-12-24 | 2022-11-18 | 中国人民解放军战略支援部队航天工程大学 | High-orbit satellite CEI signal frequency estimation method based on deep learning |
WO2023197219A1 (en) * | 2022-04-13 | 2023-10-19 | Guangdong Oppo Mobile Telecommunications Corp., Ltd. | Cnn-based post-processing filter for video compression with multi-scale feature representation |
CN114511581B (en) * | 2022-04-20 | 2022-07-08 | 四川大学华西医院 | Multi-task multi-resolution collaborative esophageal cancer lesion segmentation method and device |
CN114757930B (en) * | 2022-04-26 | 2022-12-06 | 西安电子科技大学 | Chip hardware Trojan detection method based on heat transfer |
CN115456918B (en) * | 2022-11-11 | 2023-03-24 | 之江实验室 | Image denoising method and device based on wavelet high-frequency channel synthesis |
CN116359836B (en) * | 2023-05-31 | 2023-08-15 | 成都金支点科技有限公司 | Unmanned aerial vehicle target tracking method and system based on super-resolution direction finding |
Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108416819A (en) * | 2018-02-24 | 2018-08-17 | 南京医科大学 | A kind of compression sampling MR image reconstruction method based on curvelet-fista |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN109472743B (en) * | 2018-10-25 | 2021-01-15 | 中国科学院电子学研究所 | Super-resolution reconstruction method of remote sensing image |
CN110310227B (en) * | 2019-06-27 | 2020-09-08 | 电子科技大学 | Image super-resolution reconstruction method based on high-low frequency information decomposition |
CN111798396A (en) * | 2020-07-01 | 2020-10-20 | 中通服咨询设计研究院有限公司 | Multifunctional image processing method based on wavelet transformation |
-
2021
- 2021-06-18 CN CN202110682443.3A patent/CN113256536B/en active Active
Patent Citations (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN108416819A (en) * | 2018-02-24 | 2018-08-17 | 南京医科大学 | A kind of compression sampling MR image reconstruction method based on curvelet-fista |
Also Published As
Publication number | Publication date |
---|---|
CN113256536A (en) | 2021-08-13 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN113256536B (en) | Ultrahigh-dimensional data reconstruction deep learning method based on wavelet analysis | |
CN107610194B (en) | Magnetic resonance image super-resolution reconstruction method based on multi-scale fusion CNN | |
US10489943B2 (en) | System and method for sparse image reconstruction | |
Chen et al. | Learning memory augmented cascading network for compressed sensing of images | |
CN109214989A (en) | Single image super resolution ratio reconstruction method based on Orientation Features prediction priori | |
CN111476719A (en) | Image processing method, image processing device, computer equipment and storage medium | |
Larraondo et al. | A data-driven approach to precipitation parameterizations using convolutional encoder-decoder neural networks | |
Nguyen-Duc et al. | Frequency-splitting dynamic MRI reconstruction using multi-scale 3D convolutional sparse coding and automatic parameter selection | |
CN111667445A (en) | Image compressed sensing reconstruction method based on Attention multi-feature fusion | |
US20240161251A1 (en) | Image denoising method and apparatus based on wavelet high-frequency channel synthesis | |
Thapa et al. | A performance comparison among different super-resolution techniques | |
CN113744136A (en) | Image super-resolution reconstruction method and system based on channel constraint multi-feature fusion | |
Goodwill et al. | Adaptively lossy image compression for onboard processing | |
CN116757930A (en) | Remote sensing image super-resolution method, system and medium based on residual separation attention mechanism | |
Chavez et al. | A comparison of deep-learning-based inpainting techniques for experimental X-ray scattering | |
WO2024021796A1 (en) | Image processing method and apparatus, electronic device, storage medium, and program product | |
CN117333750A (en) | Spatial registration and local global multi-scale multi-modal medical image fusion method | |
CN111882512A (en) | Image fusion method, device and equipment based on deep learning and storage medium | |
CN111932473B (en) | Multi-resolution sparse coding phase information noise reduction algorithm and storage medium | |
CN116626570A (en) | Multi-contrast MRI sampling and image reconstruction | |
CN116778016A (en) | MRI image reconstruction method, system and medium | |
US20240185570A1 (en) | Undecimated image processing method and device | |
US20230409374A1 (en) | Information processing apparatus, information processing method, and information processing system | |
US20230014745A1 (en) | Deep learning based image reconstruction | |
CN114219738A (en) | Single-image multi-scale super-resolution reconstruction network structure and method |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant | ||
CP02 | Change in the address of a patent holder | ||
CP02 | Change in the address of a patent holder |
Address after: 310023 building 10, artificial intelligence Town, 1818 Wenyi West Road, Yuhang District, Hangzhou City, Zhejiang Province Patentee after: ZHEJIANG LAB Address before: 310023 building 10, artificial intelligence Town, 1818 Wenyi West Road, Hangzhou District, Hangzhou City, Zhejiang Province Patentee before: ZHEJIANG LAB |