CN107818555B - Multi-dictionary remote sensing image space-time fusion method based on maximum posterior - Google Patents

Multi-dictionary remote sensing image space-time fusion method based on maximum posterior Download PDF

Info

Publication number
CN107818555B
CN107818555B CN201711022705.3A CN201711022705A CN107818555B CN 107818555 B CN107818555 B CN 107818555B CN 201711022705 A CN201711022705 A CN 201711022705A CN 107818555 B CN107818555 B CN 107818555B
Authority
CN
China
Prior art keywords
dictionary
resolution
image
low
group
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201711022705.3A
Other languages
Chinese (zh)
Other versions
CN107818555A (en
Inventor
何楚
张芷
郭闯创
熊德辉
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Wuhan University WHU
Original Assignee
Wuhan University WHU
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Wuhan University WHU filed Critical Wuhan University WHU
Priority to CN201711022705.3A priority Critical patent/CN107818555B/en
Publication of CN107818555A publication Critical patent/CN107818555A/en
Application granted granted Critical
Publication of CN107818555B publication Critical patent/CN107818555B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformation in the plane of the image
    • G06T3/40Scaling the whole image or part thereof
    • G06T3/4053Super resolution, i.e. output image resolution higher than sensor resolution
    • G06T3/4061Super resolution, i.e. output image resolution higher than sensor resolution by injecting details from a different spectral band
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10036Multispectral image; Hyperspectral image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Abstract

The invention relates to a multi-dictionary remote sensing image space-time fusion method based on maximum posterior, which comprises the steps of firstly, roughly classifying high-low resolution difference images, taking out image blocks for each class, and forming a training sample matrix of each class, thereby training a multi-class high-low resolution dictionary. The multi-dictionary learning considers that different landforms in the image have different shapes and textures, so that the trained dictionary is more targeted, and the difference between the landforms can be better captured. Before solving the sparse coefficient, a maximum posterior probability model is adopted to select the dictionary group, the maximum posterior probability is calculated through a likelihood function from the region pixel to the dictionary and a prior function between the region pixel dictionaries, and therefore the low-resolution differential input image pixel points are distributed to the corresponding dictionary groups. And carrying out sparse coding on each group of pixel points under the low-resolution dictionaries of the corresponding group to obtain sparse representation coefficients. And multiplying the sparse representation coefficient by the corresponding high-resolution dictionary to obtain a high-resolution difference image.

Description

Multi-dictionary remote sensing image space-time fusion method based on maximum posterior
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a multi-dictionary remote sensing image space-time fusion method based on maximum posterior.
Background
Due to hardware technology and budget constraints, there are obstacles to obtaining remote sensing images with both high spatial resolution and high temporal coverage. For example, a mode-resolution imaging spectrometer (MODIS) can observe the same region every day, with a higher time resolution. However, at the same time, their spatial resolution ranges from 250 to 1000 meters, and since the spatial resolution is too low to perform well on the detailed features of the target area, the surface coverage and ecosystem changes of a certain area or a designated place cannot be monitored well. On the other hand, remote sensing images with higher spatial resolution can be obtained from SPOT and earth surface satellite instruments (such as Landsat, the terrestrial satellite in the United states), and the spatial resolution range of the remote sensing images is 10-30 meters. Such remote sensing images are generally suitable for land use, surface mapping and the prediction of coverage changes, but the time between the shots of these high spatial resolution remote sensing images is typically half a month to a month, and the long shot intervals and the effects of severe weather and climate conditions make the images taken by this type of satellite undetectable for rapid surface changes due to seasonal changes or disturbances caused by human activities. In order to better observe the change condition of ground objects and analyze the change of the ground surface, a space-time fusion technology is developed, and the technology fuses the existing images with rich time information and space information to obtain high-spatial resolution images with short time intervals.
A space-time Adaptive reflection fusion Model (STARFM) is a classical space-time fusion Model, and many improved models based on the Model are proposed later. The STARFM model predicts a central pixel value by using the similar pixel values of the adjacent spectrums in the three aspects of distance, spectrum similarity and time difference, so that the precision of a fusion result is greatly improved. Then, an image super-resolution method based on sparse representation is proposed: and by applying a sparse representation theory, the mapping relation of the high-resolution image and the low-resolution image is obtained by establishing a high-resolution dictionary and a low-resolution dictionary, so that the super-resolution of the image is realized. The sparse representation is applied to the field of remote sensing image space-time fusion, and a space-time fusion model (SPSTFM) based on the sparse representation is provided: the change in reflectivity is predicted by uniformly training a dictionary between the low spatial resolution high temporal coverage difference images and the high spatial resolution low temporal coverage difference images, which makes it very robust to both geophysical changes and earth surface coverage type changes.
However, the above model has the following problems: the STARFM model does not fully utilize current data resources, and lacks an effective means for short-time varying areas, since it assumes that the earth coverage type and the proportion of each earth coverage type are constant during the observation period; the model based on sparse representation is learned and reconstructed based on the unified dictionary, and the reconstruction effect is restrained by the effectiveness of the learning dictionary.
Disclosure of Invention
The invention aims to train a more targeted dictionary aiming at different shapes and textures of different landforms in an image through multi-dictionary learning and can better capture the difference between the landforms. And a Bayesian framework is introduced, so that the dictionary can be selected in the process of generating the image by fully utilizing the prior information. The invention provides a multi-dictionary space-time fusion model based on the maximum posterior on the basis of a sparse representation space-time fusion model: learning multiple dictionaries aiming at different types of regions in the image; the method comprises the steps of regarding a dictionary selection problem of an input image as a solving problem of maximum posterior probability, constructing a likelihood function between a region pixel and a dictionary and a prior function between region pixel dictionaries under a Bayes framework, and obtaining dictionary selection of the region through an optimization method; and generating a high-resolution image through front and back high-resolution remote sensing images and a simultaneous low-resolution image under a classical sparse expression space-time fusion framework.
The technical scheme of the invention is a multi-dictionary remote sensing image space-time fusion method based on maximum posterior, which comprises the following steps: step 1, classifying training images, comprising the following substeps,
step 1.1, for any two interval times t1+2nAnd t1Is differenced for any two interval times t1+2nAnd t1The low-resolution images are differentiated to obtain a high-resolution differential image and a low-resolution differential image respectively, wherein n is 1,2 and 3;
step 1.2, carrying out simple rough classification on the high-resolution differential image and the low-resolution differential image respectively to obtain a plurality of groups of high-resolution and low-resolution training sample images of different classes;
step 2, learning a multi-class dictionary, comprising the following substeps,
step 2.1, according to the different categories obtained in the step 1, partitioning each category of training sample images, and stacking each image block into columns to form a training sample matrix;
2.2, putting the high-resolution and low-resolution training sample matrixes of the same category together for joint training;
2.3, respectively training high-low resolution dictionary pairs belonging to each class of training sample matrix to obtain dictionaries of different groups;
step 3, inputting t to be reconstructed1+nLow resolution image of time, and t1+2nDifferentiating the low-resolution images of the time instants to obtain an input image, and selecting a pixel-by-pixel dictionary set for the input image, comprising the sub-steps of,
step 3.1, sampling the low-resolution training sample image of each category obtained in the step 1 to obtain the likelihood probability of any pixel point under each category;
step 3.2, obtaining the prior probability of the pixel point according to the classification condition of the surrounding points of the input image pixel point;
step 3.3, traversing each pixel point in the input image, and selecting a pixel point dictionary group for each pixel point by adopting a maximum posterior probability method;
step 4, for t1+nThe high resolution image at a time is reconstructed, comprising the sub-steps of,
step 4.1, partitioning the input image according to the selection result of the dictionary group, putting the image blocks of the same group together and stacking the image blocks into columns to form an input matrix, and finally generating the input matrix with the same number as the dictionary group;
step 4.2, aiming at each input matrix, the low-resolution dictionary obtained in the step 2.3 is utilized, and an OMP method is adopted to obtain a corresponding sparse representation coefficient;
4.3, multiplying the high-resolution dictionary of each group with the sparse representation coefficient of the corresponding group to obtain each group image block of the reconstructed image; then, overlapping each group of image blocks, and obtaining a reconstructed difference image by adopting an average value in an overlapping area;
step 4.4, the reconstructed difference image is added or subtracted with the known t1+2nThe high-resolution image at the moment is a reconstructed high-resolution image L21;
step 5, mixing t1+nTime of dayLow resolution image of, and t1And (3) taking the difference of the low-resolution images at the moment as an input image, repeating the step 3-4 to obtain a reconstructed high-resolution image L22, and adding the two images L21 and L22 for averaging to obtain a final reconstruction result.
Moreover, in the step 2.3, the high-low resolution dictionary pairs belonging to each class are respectively trained for each class of training sample matrix by adopting a K-SVD method, which is realized as follows,
let the training sample matrix be x ═ x1,x2,…,xN},xiIs of the formula RNSparse representation assumes that these signals can be linearly represented by several atoms in an overcomplete dictionary matrix, namely:
x=Dα (1)
overcomplete dictionary
D={d1,d2,…,dN}∈RM×N(M<N) (2)
Coefficient of sparseness
α={α12,…,αN}T∈RN(3)
Wherein M and N are rows and columns of a dictionary matrix respectively, each column in the dictionary matrix is called a dictionary atom d, N dictionary atoms are shared, the dimension of each atom is M, α is a sparse representation coefficient, most values of α are 0, only a few values are nonzero, if the number of nonzero values is K, and K < M, α is called K sparse;
as can be seen from the principle of sparse representation and sparse coding, each class of high-resolution dictionary pair and low-resolution dictionary pair is obtained by optimizing the formula (4),
Figure BDA0001447777550000031
where λ is a regularization parameter used to balance sparsity and reconstruction errors of the sparsity-represented signal, | α | | survival1Is 11Norm, representing the sum of absolute values, | | Z-Da | | non-conducting phosphor2Is represented by2Norm, which is modulo in the usual sense, Z ═ Y; x]Y is a high resolution training sample matrix and X is lowA resolution training sample matrix, wherein X and Y are subjected to normalization processing; d ═ DI;Dh],DI,DhRespectively a low resolution dictionary and a high resolution dictionary, and D, α represent the dictionary and the sparse representation coefficient obtained after optimizing the right side of the equation;
obtaining high-low resolution dictionary pair D by adopting K-SVD method optimization formulaI,DhThe method comprises the following specific steps:
a. inputting a training sample matrix Z, wherein the atom number of the dictionary is N, and the iteration number is J;
b. initializing a dictionary, and randomly selecting K columns from a training sample matrix Z as initial values of the dictionary;
c. solving a sparse representation coefficient α according to the initialized dictionary by using an Orthogonal Matching Pursuit (OMP) algorithm;
d. updating the kth column of the dictionary:
① let the k row in the sparse matrix multiplied by the k column of the dictionary be noted as
Figure BDA0001447777550000041
② calculating a representation error matrix for the population after removing the kth column of the dictionary
Figure BDA0001447777550000042
djColumn j in the dictionary;
③Ekof which only the k column of the dictionary is reserved
Figure BDA0001447777550000043
The term after the product of the medium non-zero positions forms
Figure BDA0001447777550000044
④ pairs
Figure BDA0001447777550000045
Singular value decomposition is carried out, so that the kth column of the dictionary and the corresponding sparse coefficient are updated;
e. and d, repeating the step d until the iteration times are met, and obtaining the final high-low resolution dictionary pair.
Moreover, each pixel in the step 3.3 adopts the maximum posterior probability method to select the category of the pixel dictionary group, the realization method is as follows,
the pixel value of a certain pixel point in the known input image is xmnUnder the condition, m and n are coordinates, the posterior probability under the assumption of each dictionary group is calculated through a Bayes formula (5), the maximum posterior probability is taken as the final dictionary group of the pixel point, wherein the Bayes formula is as follows:
Figure BDA0001447777550000046
in the formula xmnPixel value representing a pixel point, ciRepresenting an ith class dictionary; p (x)mn|ci) Represented in class i dictionary ciMiddle pixel value is xmnThe probability of (2) is called likelihood probability; p (c)i) Is the prior probability of the class i dictionary; p (x)mn) Representing a pixel value of xmnA priori of P (c)i|xmn) Is a posterior probability, i.e. the pixel value at this point is known to be xmnUnder the condition (2), the point belongs to the dictionary class ciThe probability of (d);
due to the denominator P (x)mn) Is a constant independent of the dictionary group, i.e. P (x) is a constant independent of the dictionary group, i.e. no matter the pixel belongs to any dictionary groupmn) Is always kept constant, so that there is no influence in calculating the maximum a posteriori probability, P (x)mn) Without taking part in the calculation, the following formula can be converted,
Figure BDA0001447777550000051
wherein, I represents the category of the dictionary, for an input image I, the image size is MxN, and pixel values x of all pixel points in the image are taken outmn(M is 1,2, 1.. multidot.M; N is 1, 2.. multidot.N), obtaining the dictionary type I through formula (6), and combining the same pixel points of I together to form the image block I of the ith dictionaryi
Furthermore, the simple coarse classification described in step 1.2 is a binary classification.
Compared with the prior art, the invention has the advantages and beneficial effects that: the multi-dictionary learning method provided by the invention considers that different landforms in the image have different shapes and textures, so that the trained dictionary has pertinence, can better capture the difference between landforms, and can better select dictionaries in different areas of the input image based on the maximum posterior probability method of the Bayes framework. Thereby improving the quality of the reconstructed high resolution image.
Drawings
FIG. 1 is a flow chart of an embodiment of the present invention.
FIG. 2 is a diagram illustrating a Bayesian maximum a posteriori classification scheme in accordance with an embodiment of the present invention.
Detailed Description
The technical scheme of the invention is explained in detail in the following by combining the drawings and the embodiment.
The algorithm of the invention introduces multiple dictionaries and a maximum posterior. After the high-low resolution difference images are obtained, rough classification is carried out, image blocks are taken out of each class, a training sample matrix of each class is formed, and therefore a multi-class high-low resolution dictionary can be trained. The multi-dictionary learning considers that different landforms in the image have different shapes and textures, so that the trained dictionary is more targeted, and the difference between the landforms can be better captured. Before solving the sparse coefficient, a maximum posterior probability model is adopted to select the dictionary group, the maximum posterior probability is calculated through a likelihood function from the region pixel to the dictionary and a prior function between the region pixel dictionaries, and therefore the low-resolution differential input image pixel points are distributed to the corresponding dictionary groups. And carrying out sparse coding on each group of pixel points under the low-resolution dictionaries of the corresponding group to obtain sparse representation coefficients. And multiplying the sparse representation coefficient by the corresponding high-resolution dictionary to obtain a high-resolution difference image. A Bayesian framework is adopted for calculating the maximum posterior probability, so that prior information is utilized more fully, and the reconstruction effect is closer to a real image.
Referring to fig. 1, the flow chart of the embodiment of the present invention includes the following 3 steps:
step one training multi-class dictionary
(1)t3And t1The Landsat images Y3, Y1 at the time are differentiated to obtain L31, t3And t1M31 is obtained by differentiating MODIS images X3 and X1 at the time points, wherein the time intervals of the time points are uncertain, the time intervals can be multiple time points in the middle, and the images at the middle time are mainly reconstructed by using the images obtained from the time points at the two ends.
(2) And carrying out simple rough classification on L31 and M31, such as first binaryzation, screening out areas with areas larger than a specific value, and obtaining a binaryzation image of a lake, a lake degradation track and a forest land. And separating three groups of training sample images of lakes, lake degradation tracks and woodland from the difference image by using the binarization image.
(3) And partitioning each type of training sample image, wherein the blocks are overlapped, each image block is stacked into columns to form a training sample matrix, and the sample matrices with high and low resolutions are put together to be jointly trained to obtain a plurality of pairs of dictionary pairs with high and low resolutions. The K-SVD method is specifically as follows,
for an image, it is divided into image blocks, each of which is represented by { x after being stacked in columns1,x2,…,xN},xiIs of the formula RN. Sparse representation it is assumed that these signals can be linearly represented by several atoms in an overcomplete dictionary matrix, namely:
x=Dα (1)
overcomplete dictionary
D={d1,d2,…,dN}∈RM×N(M<N) (2)
Coefficient of sparseness
α={α12,…,αN}T∈RN(3)
Where M and N are the rows and columns, respectively, of a dictionary matrix, each column in the dictionary matrix is called a dictionary atom d, there are N dictionary atoms in total, the dimension of each atom is M, α is a sparse representation coefficient, where α has most values of 0 and only a few values are non-zero, if the number of non-zero values is K, and K < M, then α is K sparse, and the dictionary pairs are obtained by optimizing formula (4),
Figure BDA0001447777550000061
where λ is a regularization parameter used to balance sparsity and reconstruction errors of the sparsity-represented signal, | α | | survival1Is 11Norm, representing the sum of absolute values, | | Z-D α | | purple2Is represented by2Norm, which is modulo in the usual sense, Z ═ Y; x],D=[DI;Dh],DI,DhRespectively, low-resolution dictionary and high-resolution dictionary, D**And representing the dictionary and sparse representation coefficients obtained after optimizing the right side of the equation. And Y is a high-resolution training matrix formed by dividing the high-resolution differential image into image blocks and then stacking each image block into columns. Similarly, X is a low-resolution differential image block, and is stacked into columns to form a low-resolution training matrix. The high-resolution and low-resolution training matrixes are combined to ensure that the high-resolution sparse representation coefficient and the low-resolution sparse representation coefficient of the image block at the same position are the same during training. The training matrix is normalized to account for differences in reflectivity between different frequency bands and between high and low resolution images. And X and Y are training matrixes after normalization processing.
Optimizing formula (4) to obtain high-low resolution dictionary DI,DhIn the method, a K-SVD method is adopted, and the specific steps are as follows: a) inputting a training sample matrix Z, wherein the atom number of the dictionary is N, and the iteration number is J;
b) initializing a dictionary, and randomly selecting K columns from a training sample matrix Z as initial values of the dictionary;
c) solving a sparse representation coefficient α according to the initialized dictionary by using an Orthogonal Matching Pursuit (OMP) algorithm;
for a certain overcomplete dictionary, the solution for sparse representation coefficients α is not unique, and in order to make α the most sparse, it is therefore necessary to find the solution with the fewest non-zero values, and the problem translates into:
min||α||0s.t.x=Dα (7)
||α||0is represented by0Norm, representing the number of non-zero values; the number N of atoms in D is greater than the dimension M of the signal x, i.e. M < N, so that the overcomplete of the dictionary is guaranteed.
Solving for sparse representation coefficients can be converted into solving for l1And (4) norm. The solution to the sparse coefficients uses an Orthogonal Matching Pursuit (OMP) algorithm. The main idea of the algorithm is as follows: from the dictionary matrix D, an atom (i.e., a column) that best matches the sample matrix Z is selected, a sparse approximation is constructed, the signal residuals are solved, then the atom that best matches the signal residuals is continuously selected, and iteration is repeated, so that Z can be represented by the linear sum of the atoms plus the final residual value. When the residual values are within a negligible range, then Z is the linear combination of these atoms. The selected atoms need to be orthogonalized before being selected, so that each iteration is optimal, and the calculation time is gradually reduced along with the iteration.
d) Updating the kth column of the dictionary:
① let the k row in the sparse matrix multiplied by the k column of the dictionary be noted as
Figure BDA0001447777550000071
② calculating a representation error matrix for the population after removing the kth column of the dictionary
Figure BDA0001447777550000072
djColumn j in the dictionary;
③Ekonly the kth column of the dictionary and
Figure BDA0001447777550000073
those terms after the product of the medium non-zero positions form
Figure BDA0001447777550000074
④ pairs
Figure BDA0001447777550000075
Singular value decomposition is carried out, so that the kth column of the dictionary and the corresponding sparse coefficient are updated;
e) and d), repeating the step d) until the iteration times are met, and obtaining the final dictionary.
(4) Inputting t to be reconstructed2The low resolution image X2 at the time is differentiated from the low resolution image X3 at the time t3 to obtain an input image X32, and at an intermediate time, a low resolution image is known.
Step two, the method of the maximum posterior probability selects the dictionary group of the pixel points
The main idea of obtaining the likelihood function is to perform a rough classification on a series of low-resolution difference images, then perform pixel sampling on each category respectively, count the gray values of the pixels, draw a probability density curve of each dictionary category, and obtain the probability of each gray value under dictionaries of different groups according to the probability density curve. For example, the rough classification is roughly classified into three categories according to lakes, lake degradation tracks and woodlands, and the corresponding dictionaries are also classified into the three categories.
The prior probability is determined mainly by considering the dictionary class condition of the surrounding points. For example, when the classification results of the surrounding points are the same, the probability that the classification of the intermediate point is the same as that of the surrounding points is higher, and may be set to 0.8, and the probability that the intermediate point is classified into other classes is lower, and is assumed to be 0.2; when the grouping results of the surrounding points are not consistent, the probability that the intermediate points are classified into any class is considered to be the same. The prior function is simpler, the surrounding points can be divided into more different situations, not only four points up, down, left and right can be considered when the surrounding points are considered, but also more points can participate in determining the prior probability, so that the obtained prior probability is more accurate, and the classification result is better.
In the embodiment, the method of maximum posterior probability is adopted for X32 to select the category of the dictionary group of the pixel points, the probabilities of the whole image under all different grouping conditions are traversed, and the maximum probability is taken as the final result of selecting the category of the dictionary group; the realization is as follows,
and the dictionary group of the pixel points is selected by adopting a Bayes framework, and a Bayes formula is as follows:
Figure BDA0001447777550000081
in the formula xmnPixel value, C, representing a pixel pointiRepresenting the ith class dictionary. P (x)mn|ci) Represented in class i dictionary ciMiddle pixel value is xmnIs called likelihood probability, P (c)i) Is the prior probability of the class i dictionary. The prior probability reflects the probability that the intermediate point is assigned to the dictionary set based on the grouping results of the surrounding points. P (x)mn) Representing a pixel value of xmnA priori of P (c)i|xmn) Is a posterior probability, i.e. the pixel value at this point is known to be xmnUnder the condition (2), the point belongs to the dictionary class ciThe probability of (c).
The selection of the dictionary category of the pixel points adopts the Maximum A Posteriori (MAP) principle, which means that the known pixel value is xmnUnder the condition (3), the posterior probability under each dictionary group hypothesis is calculated by the formula (5), and the hypothesis with the highest probability is taken as the final group. Denominator P (x)mn) Is a constant independent of the dictionary class, i.e. P (x) regardless of the dictionary class to which the point belongsmn) The value of (2) is always kept unchanged, so that the maximum posterior probability is calculated without any influence and can not participate in the calculation.
Figure BDA0001447777550000082
I represents the category of the dictionary, which can be determined by the maximum posterior probability, for an input image I, the image size is M multiplied by N, and the pixel value x of all pixel points in the image is takenmn(M is 1,2, 1.. multidot.M; N is 1, 2.. multidot.N), obtaining the dictionary type I through formula (6), and combining the same pixel points of I together to form the image block I of the ith dictionaryi
Referring to fig. 2, a schematic diagram of determining a pixel group according to a maximum posterior probability in the embodiment of the present invention; each pixel point in the image is corresponding to likelihood probability, pixel value, dictionary group and prior probability. For an image, it can be seen as having four layers. Assuming that the group of intermediate points is unknown and the group of surrounding points is known. The first step is to get the likelihood probability of the intermediate point according to the pixel value. And thirdly, multiplying the likelihood probability and the prior probability to obtain the posterior probability of the intermediate point, and taking the group with the maximum posterior probability as the grouping result of the intermediate point.
Step three high resolution image reconstruction
Partitioning the input image X32 according to the selection result of the dictionary group, and combining the same pixel points of I to form an image block I of the ith dictionaryiThe image blocks of the same group are put together and stacked in columns to form an input matrix, and finally the same number of input matrices as the dictionary group are generated. For each input matrix, a corresponding low resolution dictionary is used, and the corresponding sparse representation coefficients are obtained by OMP.
High resolution dictionary D of each grouphMultiplying with each group of sparse representation coefficients respectively to obtain each group of image blocks of the reconstructed image, overlapping each group of image blocks, and obtaining the reconstructed difference image Y by adopting the mean value in the overlapping area32
Suppose that the OMP algorithm is used to solve the input matrix X32Is α, the high resolution difference image Y32Can be expressed as:
Y32=Dh*α (8)
differential image Y32The high-resolution image L3 was added to the image to obtain-L2, and the negative image was subtracted to obtain the reconstructed high-resolution image L21.
In the same manner, a difference image X12 between X1 and X2 is obtained as an input image, a high-resolution image L22 is reconstructed, and L21 and L22 are added and averaged to obtain a final reconstruction result.
The specific embodiments described herein are merely illustrative of the spirit of the invention. Various modifications or additions may be made to the described embodiments or alternatives may be employed by those skilled in the art without departing from the spirit or ambit of the invention as defined in the appended claims.

Claims (4)

1. A multi-dictionary remote sensing image space-time fusion method based on maximum posterior is characterized by comprising the following steps:
step 1, classifying training images, comprising the following substeps,
step 1.1, for any two interval times t1+2nAnd t1Is differenced for any two interval times t1+2nAnd t1The low-resolution images are differentiated to obtain a high-resolution differential image and a low-resolution differential image respectively, wherein n is 1,2 and 3;
step 1.2, carrying out simple rough classification on the high-resolution differential image and the low-resolution differential image respectively to obtain a plurality of groups of high-resolution and low-resolution training sample images of different classes;
step 2, learning a multi-class dictionary, comprising the following substeps,
step 2.1, according to the different categories obtained in the step 1, partitioning each category of training sample images, and stacking each image block into columns to form a training sample matrix;
2.2, putting the high-resolution and low-resolution training sample matrixes of the same category together for joint training;
2.3, respectively training high-low resolution dictionary pairs belonging to each class of training sample matrix to obtain dictionaries of different groups;
step 3, inputting t to be reconstructed1+nLow resolution image of time, and t1+2nDifferentiating the low-resolution images of the time instants to obtain an input image, and selecting a pixel-by-pixel dictionary set for the input image, comprising the sub-steps of,
step 3.1, sampling the low-resolution training sample image of each category obtained in the step 1 to obtain the likelihood probability of any pixel point under each category;
step 3.2, obtaining the prior probability of the pixel point according to the classification condition of the surrounding points of the input image pixel point;
step 3.3, traversing each pixel point in the input image, and selecting a pixel point dictionary group for each pixel point by adopting a maximum posterior probability method;
step 4, for t1+nThe high resolution image at a time is reconstructed, comprising the sub-steps of,
step 4.1, partitioning the input image according to the selection result of the dictionary group, putting the image blocks of the same group together and stacking the image blocks into columns to form an input matrix, and finally generating the input matrix with the same number as the dictionary group;
step 4.2, aiming at each input matrix, the low-resolution dictionary obtained in the step 2.3 is utilized, and an OMP method is adopted to obtain a corresponding sparse representation coefficient;
4.3, multiplying the high-resolution dictionary of each group with the sparse representation coefficient of the corresponding group to obtain each group image block of the reconstructed image; then, overlapping each group of image blocks, and obtaining a reconstructed difference image by adopting an average value in an overlapping area;
step 4.4, the reconstructed difference image is added or subtracted with the known t1+2nThe high-resolution image at the moment is a reconstructed high-resolution image L21;
step 5, mixing t1+nLow resolution image of time, and t1And (3) taking the difference of the low-resolution images at the moment as an input image, repeating the step 3-4 to obtain a reconstructed high-resolution image L22, and adding the two images L21 and L22 for averaging to obtain a final reconstruction result.
2. The maximum a posteriori based spatio-temporal fusion method for multi-dictionary remote sensing images according to claim 1, characterized in that: in the step 2.3, a high-low resolution dictionary pair belonging to each class is trained for each class of training sample matrix by adopting a K-SVD method in the following way,
let the training sample matrix be x ═ x1,x2,…,xN},xiIs of the formula RNSparse representation assumes that these signals can be linearly represented by several atoms in an overcomplete dictionary matrix, namely:
x=Dα (1)
overcomplete dictionary
D={d1,d2,…,dN}∈RM×N(2)
Coefficient of sparseness
α={α12,…,αN}T∈RN(3)
Wherein M and N are rows and columns of the dictionary matrix respectively, wherein M is less than N, each column in the dictionary matrix is called a dictionary atom d, N dictionary atoms are shared, and the dimension of each atom is M;
as can be seen from the principle of sparse representation and sparse coding, each class of high-resolution dictionary pair and low-resolution dictionary pair is obtained by optimizing the formula (4),
Figure FDA0002238719140000021
where λ is a regularization parameter used to balance sparsity and reconstruction errors of the sparsity-represented signal, | α | | survival1Is 11Norm, representing the sum of absolute values, | | Z-Da | | non-conducting phosphor2Is represented by2Norm, Z ═ Y; x]Y is a high-resolution training sample matrix, X is a low-resolution training sample matrix, and both X and Y are subjected to normalization processing; d ═ DI;Dh],DI,DhRespectively, low-resolution dictionary and high-resolution dictionary, D**Expressing a dictionary and a sparse representation coefficient obtained after optimizing the right side of the equation;
obtaining high-low resolution dictionary pair D by adopting K-SVD method optimization formulaI,DhThe method comprises the following specific steps:
a. inputting a training sample matrix Z, wherein the atom number of the dictionary is N, and the iteration number is J;
b. initializing a dictionary, and randomly selecting K columns from a training sample matrix Z as initial values of the dictionary;
c. solving a sparse representation coefficient α according to the initialized dictionary by using an Orthogonal Matching Pursuit (OMP) algorithm;
d. updating the kth column of the dictionary:
① let the k row in the sparse matrix multiplied by the k column of the dictionary be noted as
Figure FDA0002238719140000022
② calculating a representation error matrix for the population after removing the kth column of the dictionary
Figure FDA0002238719140000031
djColumn j in the dictionary;
③Ekof which only the k column of the dictionary is reserved
Figure FDA0002238719140000032
The term after the product of the medium non-zero positions forms
Figure FDA0002238719140000033
④ pairs
Figure FDA0002238719140000034
Singular value decomposition is carried out, so that the kth column of the dictionary and the corresponding sparse coefficient are updated;
e. and d, repeating the step d until the iteration times are met, and obtaining the final high-low resolution dictionary pair.
3. The maximum a posteriori based spatio-temporal fusion method for multi-dictionary remote sensing images according to claim 1 or 2, characterized in that: in the step 3.3, each pixel point adopts the maximum posterior probability method to select the category of the pixel point dictionary group, the realization method is as follows,
the pixel value of a certain pixel point in the known input image is xmnUnder the condition (1), m and n are coordinates, and the posterior is calculated under the assumption of each dictionary group through a Bayes formula (5)And (3) probability, taking the maximum posterior probability as the final dictionary group of the pixel point, wherein the Bayesian formula is as follows:
Figure FDA0002238719140000035
in the formula xmnPixel value representing a pixel point, ciRepresenting an ith class dictionary; p (x)mn|ci) Represented in class i dictionary ciMiddle pixel value is xmnThe probability of (2) is called likelihood probability; p (c)i) Is the prior probability of the class i dictionary; p (x)mn) Representing a pixel value of xmnA priori of P (c)i|xmn) Is a posterior probability, i.e. the pixel value at this point is known to be xmnUnder the condition (2), the point belongs to the dictionary class ciThe probability of (d);
p (x) when calculating the maximum a posteriori probabilitymn) Without involving in the calculation, the formula (5) can be converted into the following formula,
Figure FDA0002238719140000036
wherein, I represents the category of the dictionary, for an input image I, the image size is MxN, and pixel values x of all pixel points in the image are taken outmnWherein M is 1, 2.. times.m; n is 1,2, the dictionary type I is obtained through the formula (6), and the same pixel points of I are combined together to form the image block I of the I-th dictionaryi
4. The maximum a posteriori based spatio-temporal fusion method for multi-dictionary remote sensing images according to claim 1, characterized in that: the simple coarse classification described in step 1.2 is a binary classification.
CN201711022705.3A 2017-10-27 2017-10-27 Multi-dictionary remote sensing image space-time fusion method based on maximum posterior Active CN107818555B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201711022705.3A CN107818555B (en) 2017-10-27 2017-10-27 Multi-dictionary remote sensing image space-time fusion method based on maximum posterior

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201711022705.3A CN107818555B (en) 2017-10-27 2017-10-27 Multi-dictionary remote sensing image space-time fusion method based on maximum posterior

Publications (2)

Publication Number Publication Date
CN107818555A CN107818555A (en) 2018-03-20
CN107818555B true CN107818555B (en) 2020-03-10

Family

ID=61604143

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201711022705.3A Active CN107818555B (en) 2017-10-27 2017-10-27 Multi-dictionary remote sensing image space-time fusion method based on maximum posterior

Country Status (1)

Country Link
CN (1) CN107818555B (en)

Families Citing this family (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN108932710B (en) * 2018-07-10 2021-11-12 武汉商学院 Remote sensing space-time information fusion method
CN108921116B (en) * 2018-07-10 2021-09-28 武汉商学院 Time-varying information extraction method for remote sensing image
CN109410165B (en) * 2018-11-14 2022-02-11 太原理工大学 Multispectral remote sensing image fusion method based on classification learning
CN111242848B (en) * 2020-01-14 2022-03-04 武汉大学 Binocular camera image suture line splicing method and system based on regional feature registration
CN112183595B (en) * 2020-09-18 2023-08-04 中国科学院空天信息创新研究院 Space-time remote sensing image fusion method based on grouping compressed sensing
CN113160100A (en) * 2021-04-02 2021-07-23 深圳市规划国土房产信息中心(深圳市空间地理信息中心) Fusion method, fusion device and medium based on spectral information image
CN113449683B (en) * 2021-07-15 2022-12-27 江南大学 High-frequency ultrasonic sparse denoising method and system based on K-SVD training local dictionary
CN116958717B (en) * 2023-09-20 2023-12-12 山东省地质测绘院 Intelligent geological big data cleaning method based on machine learning

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102360498A (en) * 2011-10-27 2012-02-22 江苏省邮电规划设计院有限责任公司 Reconstruction method for image super-resolution
CN104103052A (en) * 2013-04-11 2014-10-15 北京大学 Sparse representation-based image super-resolution reconstruction method
CN104778671A (en) * 2015-04-21 2015-07-15 重庆大学 Image super-resolution method based on SAE and sparse representation
CN105374060A (en) * 2015-10-15 2016-03-02 浙江大学 PET image reconstruction method based on structural dictionary constraint
CN106227015A (en) * 2016-07-11 2016-12-14 中国科学院深圳先进技术研究院 A kind of hologram image super-resolution reconstruction method and system based on compressive sensing theory
CN106919952A (en) * 2017-02-23 2017-07-04 西北工业大学 EO-1 hyperion Anomaly target detection method based on structure rarefaction representation and internal cluster filter

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN102360498A (en) * 2011-10-27 2012-02-22 江苏省邮电规划设计院有限责任公司 Reconstruction method for image super-resolution
CN104103052A (en) * 2013-04-11 2014-10-15 北京大学 Sparse representation-based image super-resolution reconstruction method
CN104778671A (en) * 2015-04-21 2015-07-15 重庆大学 Image super-resolution method based on SAE and sparse representation
CN105374060A (en) * 2015-10-15 2016-03-02 浙江大学 PET image reconstruction method based on structural dictionary constraint
CN106227015A (en) * 2016-07-11 2016-12-14 中国科学院深圳先进技术研究院 A kind of hologram image super-resolution reconstruction method and system based on compressive sensing theory
CN106919952A (en) * 2017-02-23 2017-07-04 西北工业大学 EO-1 hyperion Anomaly target detection method based on structure rarefaction representation and internal cluster filter

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
"基于组稀疏表示的自然图像超分辨率算法研究";李祥灿;《中国优秀硕士学位论文全文数据库 信息科技辑》;20140715(第7期);I138-785页 *

Also Published As

Publication number Publication date
CN107818555A (en) 2018-03-20

Similar Documents

Publication Publication Date Title
CN107818555B (en) Multi-dictionary remote sensing image space-time fusion method based on maximum posterior
Postels et al. Sampling-free epistemic uncertainty estimation using approximated variance propagation
CN109035142B (en) Satellite image super-resolution method combining countermeasure network with aerial image prior
CN110109060B (en) Radar radiation source signal sorting method and system based on deep learning network
CN113128134B (en) Mining area ecological environment evolution driving factor weight quantitative analysis method
CN111369487B (en) Hyperspectral and multispectral image fusion method, system and medium
CN103488968B (en) The mixed pixel material of remote sensing images constitutes decomposer and the method for becoming more meticulous
CN107730482B (en) Sparse fusion method based on regional energy and variance
CN111461146B (en) Change detection method based on sparse cross reconstruction
CN107491793B (en) Polarized SAR image classification method based on sparse scattering complete convolution
CN113327231B (en) Hyperspectral abnormal target detection method and system based on space-spectrum combination
CN106845343B (en) Automatic detection method for optical remote sensing image offshore platform
CN110490894B (en) Video foreground and background separation method based on improved low-rank sparse decomposition
CN109671019B (en) Remote sensing image sub-pixel mapping method based on multi-objective optimization algorithm and sparse expression
CN114723631A (en) Image denoising method, system and device based on depth context prior and multi-scale reconstruction sub-network
CN106097290A (en) SAR image change detection based on NMF image co-registration
CN110956601A (en) Infrared image fusion method and device based on multi-sensor mode coefficients and computer readable storage medium
CN113421198B (en) Hyperspectral image denoising method based on subspace non-local low-rank tensor decomposition
Wang et al. SAR images super-resolution via cartoon-texture image decomposition and jointly optimized regressors
CN112200752B (en) Multi-frame image deblurring system and method based on ER network
Yufeng et al. Research on SAR image change detection algorithm based on hybrid genetic FCM and image registration
CN113222206A (en) Traffic state prediction method based on ResLS-C deep learning combination
CN113989612A (en) Remote sensing image target detection method based on attention and generation countermeasure network
Vyshnevyi Two–Stage Segmentation of SAR Images Distorted by Additive Noise with Uncorrelated Samples
CN111899226A (en) Hyperspectral image target prior optimization method based on multitask sparse learning

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant