CN117236201B - Diffusion and ViT-based downscaling method - Google Patents

Diffusion and ViT-based downscaling method Download PDF

Info

Publication number
CN117236201B
CN117236201B CN202311525721.XA CN202311525721A CN117236201B CN 117236201 B CN117236201 B CN 117236201B CN 202311525721 A CN202311525721 A CN 202311525721A CN 117236201 B CN117236201 B CN 117236201B
Authority
CN
China
Prior art keywords
model
diffusion
steps
precipitation
resolution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202311525721.XA
Other languages
Chinese (zh)
Other versions
CN117236201A (en
Inventor
季焱
智协飞
张永宏
卢楚翰
彭婷
张玲
王靖宇
陈超辉
吉璐莹
吕阳
朱寿鹏
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing Institute Of Meteorological Science And Technology Innovation
National University of Defense Technology
Nanjing University of Information Science and Technology
Wuxi University
Original Assignee
Nanjing Institute Of Meteorological Science And Technology Innovation
National University of Defense Technology
Nanjing University of Information Science and Technology
Wuxi University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing Institute Of Meteorological Science And Technology Innovation, National University of Defense Technology, Nanjing University of Information Science and Technology, Wuxi University filed Critical Nanjing Institute Of Meteorological Science And Technology Innovation
Priority to CN202311525721.XA priority Critical patent/CN117236201B/en
Publication of CN117236201A publication Critical patent/CN117236201A/en
Application granted granted Critical
Publication of CN117236201B publication Critical patent/CN117236201B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y02TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
    • Y02ATECHNOLOGIES FOR ADAPTATION TO CLIMATE CHANGE
    • Y02A90/00Technologies having an indirect contribution to adaptation to climate change
    • Y02A90/10Information and communication technologies [ICT] supporting adaptation to climate change, e.g. for weather forecasting or climate simulation

Abstract

The invention discloses a downscaling method based on Diffusion and ViT, which comprises the following steps: s1, establishing a low-resolution numerical mode precipitation forecast and a high-resolution precipitation observation sample, and preprocessing; s2, constructing a Diffusion-Vision-transformation precipitation prediction model; s3, training a model until errors of the Diffusion-Vision-transformation converge, and storing the model and predicting; according to the invention, the Vision Transformer model is used for replacing the U-Net structure in the original Diffusion model, so that the training efficiency of the model is greatly improved, and the time of the model for prediction is reduced.

Description

Diffusion and ViT-based downscaling method
Technical Field
The invention relates to the technical field of weather forecast, in particular to a scale-down method based on Diffusion and ViT.
Background
Most of the traditional statistical downscaling methods are models based on linear frames, and are difficult to process complex and high-dimensional meteorological field data and characterize an atmospheric nonlinear dynamic process. The rise in deep learning provides new directions for characterizing complex data that are highly dimensional and strongly nonlinear, such as meteorological element fields. By utilizing the efficient spatial feature extraction module to extract key information of high-dimensional spatial data, a statistical model of low-resolution input to high-resolution output is established, the deep learning model can be effectively applied to scenes such as picture denoising and picture resolution improvement, and the like, and the method is generally called as a super-resolution model. However, how to efficiently migrate the model to the down-scale problem of meteorology, and further improve the calculation efficiency and the prediction accuracy of the model, still needs further research and exploration.
Disclosure of Invention
The invention aims to: the invention aims to provide a downscaling method based on Diffusion and ViT to solve the problems of insufficient spatial resolution and large prediction error of numerical mode precipitation prediction.
The technical scheme is as follows: the invention discloses a downscaling method based on Diffusion and ViT, which comprises the following steps:
s1: establishing a low-resolution numerical mode precipitation forecast and a high-resolution precipitation observation sample, and preprocessing;
s2: constructing a Diffusion-Vision-transformation precipitation prediction model; the method comprises the following steps:
s21: forward noise adding is carried out on the high-resolution precipitation observation sample in the Diffusion model;
s22: extracting high-order spatial features of low-resolution numerical mode precipitation prediction by using a Vision-transducer model;
s23: denoising the result obtained in the step S21 in a Diffusion model, and introducing the high-order spatial features obtained in the step S22 as condition information to obtain a reduced-scale high-resolution precipitation forecast;
s3: training the model until the error of the dispersion-Vision-transducer converges, and storing the model and predicting.
Further, in the step S1, the preprocessing includes: the data set is subjected to operations of logarithmization and normalization.
Further, the specific process of step S21 is as follows:
setting a high-resolution precipitation observation sample pretreated at a certain momentGaussian noise +.A. The original observation was added stepwise in T times>Obtain->Data distribution at time t +.>Before +.>The formula is as follows:
wherein,is a preset constant superparameter, and ranges between 0 and 1;
data distribution at last time tCan be made of data +.0 time instant>The distribution is obtained by the following formula:
wherein,and for->Then->
Further, the step S22 is specifically as follows: input paired high resolution precipitation observation sampleAnd low resolution numberValue mode precipitation forecast ++>And determining the step number T of forward noise and the variance super-parameter of the added random Gaussian noise
Further, the step S23 includes the following steps:
s231: dividing the low-resolution numerical mode precipitation forecast into a plurality of image blocks, and then carrying out linear mapping on the divided image blocks;
s232: the position information of different image blocks is represented by position codes, and the processed coding information is used as the input of N groups of self-attention modules;
s233: the convolution operation is replaced with a spatial self-attention module.
Further, the formula of step S231 is as follows:
;
wherein,for a group of segmented tiles, +.>For the weight coefficient to be trained, +.>For the truncation coefficient to be trained, +.>Is a set of vectors that have undergone linear mapping.
Further, the step S232 position encoding is a two-dimensional position embedding method.
Further, the step S233 specifically includes the following steps:
set a group of divided blocks asThree sets of weights are utilized, namely query weight +.>Key weight->Numerical weight->Raw data is divided into three features: query matrix->Key value matrixMatrix of values->The method comprises the steps of carrying out a first treatment on the surface of the Then->Corresponding self-attention->The formula is as follows:
wherein,is->Square root of dimension.
Further, the step S3 specifically includes the following steps:
the results obtained through steps S21-S22 are:wherein->For the model obtained for steps S21-S22, and (2)>For low resolution numerical model precipitation forecast +.>For paired high-resolution precipitation observation samples,the super parameter preset in the step S21 is T, and the number of steps of forward noise adding in the step S21 is T; the prediction error of the Diffusion-Vision-transformation model in step S3>The formula is as follows:
wherein,is a random Gaussian distribution, then +.>
Forecast errors when the diffion-Vision-transform modelUpon convergence, deducing step T in reverse until model prediction +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein, the former step +.>By the next step->The formula is as follows:
wherein,for the model obtained for steps S21-S22, and (2)>For low resolution numerical model precipitation forecast +.>For the super parameter preset in step S21, < ->Is a random Gaussian distribution, then +.>
An apparatus of the present invention includes a memory, a processor, and a program stored on the memory and executable on the processor, the processor implementing steps in any of the methods of downscaling based on diffion and ViT when the program is executed.
The beneficial effects are that: compared with the prior art, the invention has the following remarkable advantages: (1) The refinement degree of the downscaling prediction is improved by utilizing the Diffusion model, and the method has more advantages particularly in the task aiming at downscaling multiple exceeding 4; (2) By using the Vision Transformer model to replace the U-Net structure in the original Diffusion model, the training efficiency of the model is greatly improved, and the time of the model for prediction is reduced.
Drawings
FIG. 1 is a general flow chart of the present invention;
FIG. 2 is a schematic diagram of a training flow of the diffration-ViT model;
FIG. 3 is a schematic diagram of a diffration model;
FIG. 4 is a schematic diagram of a Vision-transducer model.
Description of the embodiments
The technical scheme of the invention is further described below with reference to the accompanying drawings.
As shown in fig. 1, an embodiment of the present invention provides a downscaling method based on diffration and ViT, which includes the following steps:
s1: establishing a low-resolution numerical mode precipitation forecast and a high-resolution precipitation observation sample, and preprocessing; the pretreatment comprises the following steps: the data set is subjected to operations of logarithmization and normalization.
As shown in fig. 2, S2: constructing a Diffusion-Vision-transformation precipitation prediction model; the method comprises the following steps:
s21: forward noise adding is carried out on the high-resolution precipitation observation sample in the Diffusion model; the method comprises the following steps: as shown in FIG. 3, a high-resolution precipitation observation sample pretreated at a certain moment is setGaussian noise +.A. The original observation was added stepwise in T times>Obtain->Data distribution at time t +.>Before +.>The formula is as follows: data distribution at time t +.>Before +.>The formula is as follows:
wherein,is a preset constant excessParameters ranging between 0 and 1;
data distribution at last time tCan be made of data +.0 time instant>The distribution is obtained by the following formula:
wherein,and for->Then->
S22: extracting high-order spatial features of low-resolution numerical mode precipitation prediction by using a Vision-transducer model; the method comprises the following steps: input paired high resolution precipitation observation sampleAnd low resolution numerical mode precipitation forecast +.>And determining the step number T of forward noise addition and the variance super-parameter of the added random Gaussian noise +.>
Denoising the result obtained in the step S21 in a Diffusion model, and introducing the high-order spatial features obtained in the step S22 as condition information to obtain a reduced-scale high-resolution precipitation forecast;
the method comprises the following steps:
s231: as shown in fig. 4, the low-resolution numerical mode precipitation prediction is divided into a plurality of blocks, and then the divided blocks are subjected to linear mapping; the formula is as follows:
;
wherein,for a group of segmented tiles, +.>For the weight coefficient to be trained, +.>For the truncation coefficient to be trained, +.>Is a set of vectors that have undergone linear mapping.
S232: the position information of different image blocks is represented by position codes, and the processed coding information is used as the input of N groups of self-attention modules; the position coding is a two-dimensional position embedding method, and specifically comprises the following steps: by encoding the position of each tile relative to the X-axis and the Y-axis, different tiles are represented with different position encodings.
S233: the convolution operation is replaced with a spatial self-attention module. The method comprises the following steps:
set a group of divided blocks asThree sets of weights are utilized, namely query weight +.>Key weight->Numerical weight->Raw data is divided into three features: query matrix->Key value matrixMatrix of values->The method comprises the steps of carrying out a first treatment on the surface of the Then->Corresponding self-attention->The formula is as follows:
wherein,is->Square root of dimension. The spatial self-attention module consists of a regularization layer, a multi-head self-attention, a residual structure and a feedforward neural network.
S3: training the model until the error of the dispersion-Vision-transducer converges, and storing the model and predicting. The method comprises the following steps:
the results obtained through steps S21-S22 are:wherein->For the model obtained for steps S21-S22, and (2)>For low resolution numerical model precipitation forecast +.>For paired high-resolution precipitation observation samples,the super parameter preset in the step S21 is T, and the number of steps of forward noise adding in the step S21 is T; the prediction error of the Diffusion-Vision-transformation model in step S3>The formula is as follows:
wherein,is a random Gaussian distribution, then +.>
Forecast errors when the diffion-Vision-transform modelUpon convergence, deducing step T in reverse until model prediction +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein, the former step +.>By the next step->The formula is as follows:
wherein,for the model obtained for steps S21-S22, and (2)>Precipitation pre-treatment for low resolution numerical modeNewspaper (I) of->For the super parameter preset in step S21, < ->Is a random Gaussian distribution, then +.>
The embodiment of the invention also provides equipment, which comprises a memory, a processor and a program stored on the memory and capable of running on the processor, wherein the processor realizes the steps in any one of the downscaling methods based on the Diffusion and ViT when executing the program.

Claims (10)

1. A downscaling method based on Diffusion and ViT, comprising the steps of:
s1: establishing a low-resolution numerical mode precipitation forecast and a high-resolution precipitation observation sample, and preprocessing;
s2: constructing a Diffusion-Vision-transformation precipitation prediction model; the method comprises the following steps:
s21: forward noise adding is carried out on the high-resolution precipitation observation sample in the Diffusion model;
s22: extracting high-order spatial features of low-resolution numerical mode precipitation prediction by using a Vision-transducer model;
s23: denoising the result obtained in the step S21 in a Diffusion model, and introducing the high-order spatial features obtained in the step S22 as condition information to obtain a reduced-scale high-resolution precipitation forecast;
s3: training the model until the error of the dispersion-Vision-transducer converges, and storing the model and predicting.
2. The downscaling method based on Diffusion and ViT of claim 1, wherein the preprocessing in step S1 comprises: the data set is subjected to operations of logarithmization and normalization.
3. The downscaling method based on the Diffusion and the ViT according to claim 1, wherein the specific procedure of the step S21 is as follows:
setting a high-resolution precipitation observation sample pretreated at a certain momentAdding Gaussian noise to the original observation step by step in T times to obtain + ->Data distribution at time t +.>Before +.>The formula is as follows:
wherein,is a preset constant superparameter, and ranges between 0 and 1;
data distribution at last time tCan be made of data +.0 time instant>The distribution is obtained by the following formula:
wherein,and for->Then->
4. The downscaling method based on Diffusion and ViT of claim 1, wherein S22 is specifically as follows: input paired high resolution precipitation observation sampleAnd low resolution numerical mode precipitation forecast +.>And determining the step number T of forward noise addition and the variance super-parameter of the added random Gaussian noise +.>
5. The downscaling method based on Diffusion and ViT of claim 1, wherein the step S23 comprises the steps of:
s231: dividing the low-resolution numerical mode precipitation forecast into a plurality of image blocks, and then carrying out linear mapping on the divided image blocks;
s232: the position information of different image blocks is represented by position codes, and the processed coding information is used as the input of N groups of self-attention modules;
s233: the convolution operation is replaced with a spatial self-attention module.
6. The downscaling method based on Diffusion and ViT of claim 4, wherein the formula of step S231 is as follows:
;
wherein,for a group of segmented tiles, +.>For the weight coefficient to be trained, +.>As the truncated coefficients to be trained,is a set of vectors that have undergone linear mapping.
7. The downscaling method of claim 4 wherein the step S232 position encoding is a two-dimensional position embedding method.
8. The downscaling method based on Diffusion and ViT of claim 4, wherein the step S233 is specifically as follows:
set a group of divided blocks asThree sets of weights are utilized, namely query weight +.>Key weight->Numerical weight->Raw data is divided into three features: query matrix->Key value matrix->Matrix of values->The method comprises the steps of carrying out a first treatment on the surface of the Then->Corresponding self-attention->The formula is as follows:
wherein,is->Square root of dimension.
9. The downscaling method based on Diffusion and ViT according to claim 1, wherein the step S3 is specifically as follows:
the results obtained through steps S21-S22 are:wherein->For the model obtained in steps S21-S22, and (2)>For low resolution numerical model precipitation forecast +.>For paired high resolution precipitation observation samples, +.>The super parameter preset in the step S21 is T, and the number of steps of forward noise adding in the step S21 is T; the prediction error of the Diffusion-Vision-transducer model in step S3>The formula is as follows:
wherein,is a random Gaussian distribution, then +.>
Forecast error when using a Diffuse-Vision-transducer modelUpon convergence, deducing step T in reverse until model prediction +.>The method comprises the steps of carrying out a first treatment on the surface of the Wherein, the former step +.>By the next step->The formula is as follows:
wherein,for the model obtained in steps S21-S22, and (2)>For low resolution numerical model precipitation forecast +.>For the super parameter preset in step S21, < ->Is a random Gaussian distribution, then +.>
10. An apparatus comprising a memory, a processor and a program stored on the memory and executable on the processor, wherein the processor performs the steps in a method of downscaling based on diffion and ViT as claimed in any one of claims 1-9 when the program is executed.
CN202311525721.XA 2023-11-16 2023-11-16 Diffusion and ViT-based downscaling method Active CN117236201B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202311525721.XA CN117236201B (en) 2023-11-16 2023-11-16 Diffusion and ViT-based downscaling method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202311525721.XA CN117236201B (en) 2023-11-16 2023-11-16 Diffusion and ViT-based downscaling method

Publications (2)

Publication Number Publication Date
CN117236201A CN117236201A (en) 2023-12-15
CN117236201B true CN117236201B (en) 2024-02-23

Family

ID=89098904

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202311525721.XA Active CN117236201B (en) 2023-11-16 2023-11-16 Diffusion and ViT-based downscaling method

Country Status (1)

Country Link
CN (1) CN117236201B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6688180B1 (en) * 1999-07-05 2004-02-10 Sinvent As Multi-test assembly for evaluating, detecting and mountoring processes at elevated pressure
CN109524061A (en) * 2018-10-23 2019-03-26 中国人民解放军陆军防化学院 A kind of radionuclide diffusion calculation method based on transmission coefficient matrix
CN115964869A (en) * 2022-12-14 2023-04-14 西北核技术研究所 High-space-time resolution atmospheric pollution diffusion migration simulation method
CN116740223A (en) * 2023-04-26 2023-09-12 先进操作系统创新中心(天津)有限公司 Method for generating image based on text
CN116953642A (en) * 2023-06-29 2023-10-27 安徽大学 Millimeter wave radar gesture recognition method based on adaptive coding Vision Transformer network

Family Cites Families (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP1882693B1 (en) * 2006-07-24 2012-11-14 Imec Method and solution for growing a charge-transfer complex salt onto a metal surface
JP6697394B2 (en) * 2014-04-10 2020-05-20 イェール ユニバーシティーYale University Methods and compositions for detecting a misfolded protein
JP2024517412A (en) * 2021-04-16 2024-04-22 ストロング フォース ヴィーシーエヌ ポートフォリオ 2019,エルエルシー Systems, methods, kits, and apparatus for digital product network systems and biology-based value chain networks
CA3177620A1 (en) * 2021-05-06 2022-11-06 Strong Force Iot Portfolio 2016, Llc Quantum, biological, computer vision, and neural network systems for industrial internet of things
KR20240019771A (en) * 2021-05-11 2024-02-14 스트롱 포스 브이씨엔 포트폴리오 2019, 엘엘씨 Systems, methods, kits and devices for edge distributed storage and querying of value chain networks
US20220301097A1 (en) * 2022-06-03 2022-09-22 Intel Corporation Methods and apparatus to implement dual-attention vision transformers for interactive image segmentation

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US6688180B1 (en) * 1999-07-05 2004-02-10 Sinvent As Multi-test assembly for evaluating, detecting and mountoring processes at elevated pressure
CN109524061A (en) * 2018-10-23 2019-03-26 中国人民解放军陆军防化学院 A kind of radionuclide diffusion calculation method based on transmission coefficient matrix
CN115964869A (en) * 2022-12-14 2023-04-14 西北核技术研究所 High-space-time resolution atmospheric pollution diffusion migration simulation method
CN116740223A (en) * 2023-04-26 2023-09-12 先进操作系统创新中心(天津)有限公司 Method for generating image based on text
CN116953642A (en) * 2023-06-29 2023-10-27 安徽大学 Millimeter wave radar gesture recognition method based on adaptive coding Vision Transformer network

Non-Patent Citations (3)

* Cited by examiner, † Cited by third party
Title
基于点扩散函数的条件植被温度指数降尺度转换方法;王鹏新;《农业机械学报》;第48卷(第12期);165-173 *
江淮梅雨锋暴雨的中尺度可预报性研究;杨舒楠;《中国博士学位论文全文数据库 基础科学辑》(第6期);A009-6 *
面向多天气退化图像恢复的自注意力扩散模型;秦菁;《上海交通大学学报》;1-22 *

Also Published As

Publication number Publication date
CN117236201A (en) 2023-12-15

Similar Documents

Publication Publication Date Title
CN111598786B (en) Hyperspectral image unmixing method based on depth denoising self-coding network
CN112733965B (en) Label-free image classification method based on small sample learning
CN111259397B (en) Malware classification method based on Markov graph and deep learning
CN112560966B (en) Polarized SAR image classification method, medium and equipment based on scattering map convolution network
Huang et al. Compressing multidimensional weather and climate data into neural networks
CN111931813A (en) CNN-based width learning classification method
CN115311555A (en) Remote sensing image building extraction model generalization method based on batch style mixing
CN112417752A (en) Cloud layer track prediction method and system based on convolution LSTM neural network
CN110807497A (en) Handwritten data classification method and system based on deep dynamic network
CN108388918B (en) Data feature selection method with structure retention characteristics
CN117236201B (en) Diffusion and ViT-based downscaling method
CN111080516A (en) Super-resolution image reconstruction method based on self-sampling enhancement
CN113835964B (en) Cloud data center server energy consumption prediction method based on small sample learning
CN109840888B (en) Image super-resolution reconstruction method based on joint constraint
CN114581789A (en) Hyperspectral image classification method and system
CN109919200B (en) Image classification method based on tensor decomposition and domain adaptation
CN113537573A (en) Wind power operation trend prediction method based on dual space-time feature extraction
CN112926670A (en) Garbage classification system and method based on transfer learning
CN111967580B (en) Low-bit neural network training method and system based on feature migration
CN116698410B (en) Rolling bearing multi-sensor data monitoring method based on convolutional neural network
CN113627073B (en) Underwater vehicle flow field result prediction method based on improved Unet++ network
CN117878928B (en) Wind power prediction method and device based on deep learning
CN113627556B (en) Method and device for realizing image classification, electronic equipment and storage medium
CN114022360B (en) Rendered image super-resolution system based on deep learning
CN117495153A (en) Method and system for quickly correcting regional power grid line loss influence factors

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant