CN106023122A - Image fusion method based on multi-channel decomposition - Google Patents
Image fusion method based on multi-channel decomposition Download PDFInfo
- Publication number
- CN106023122A CN106023122A CN201610291857.2A CN201610291857A CN106023122A CN 106023122 A CN106023122 A CN 106023122A CN 201610291857 A CN201610291857 A CN 201610291857A CN 106023122 A CN106023122 A CN 106023122A
- Authority
- CN
- China
- Prior art keywords
- image
- coefficient
- texture
- sparse
- dictionary
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10004—Still image; Photographic image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20212—Image combination
- G06T2207/20221—Image fusion; Image merging
Abstract
The invention provides an image fusion method based on multi-channel decomposition. The method comprises the steps that (1) multi-channel decomposition feature extraction is carried out on a source image, namely multi-channel decomposition is carried out on the source image to acquire the structural component and the texture component of the image; (2) a fusion rule based on sparse representation is set; (3) coefficient fusion is carried out, namely the sparse representation coefficient of an image block corresponding to each image is fused according to a certain fusion rule to acquire a sparse coefficient after structural image and texture fusion; and (4) image reconstruction is carried out. The step (2) comprises the steps that block vectorization is carried out on the multi-channel-decomposed image; sparse representation is carried out on a number of vectorized blocks; an over-complete dictionary is learned; and the sparse representation coefficient of each column vector in the dictionary is calculated. The step (4) comprises the steps that since reconstruction is the inverse process of sparse decomposition, the over-complete dictionary and the fusion sparse coefficient are combined for reconstruction; and structural and texture reconstruction is carried out to acquire a final fusion image.
Description
Technical field
The invention belongs to image fusion technology field, particularly a kind of based on multichannel image interfusion method.
Background technology
Nowadays, pattern recognition becomes more and more important at daily life with the application in industrial circle.The little image of arriving
Feature extraction, the big information identification to scene of a crime.Thus, it is found that a kind of efficient image interfusion method becomes increasingly to weigh
Want.The process of image co-registration is the process to source images feature extraction.Texture image and structural images are the important of source images
Feature and the embodiment of different aspect, therefore take into full account that the texture of image and structural information can improve the quality of fusion image.
We mainly make use of TV-L1 algorithm and the K-SVD algorithm of rarefaction representation of exploded view picture.
The structure that TV-L1 model obtains under different scale is different, thus can be well by different at metric space
Structural area separates.Along with parameter lambda value tapers into, little structure is smoothed successively, and smooth effect is very clean,
The key of this effect multi-resolution decomposition just.Meanwhile, details in the structure remained is it is clear that this point is for image
Decomposition is critically important.So using TV-L1 model to carry out resolving into texture and structure components by image.
K-SVD algorithm is to be proposed by Michal Aharon, Michael Elad of the Institute of Technology of Israel et al. for 2006
Come, be the most classical a kind of dictionary training algorithm, and reached good training effect.Its objective is to solve lower column matrix
The solution of equation:
Y=DX (1)
Wherein D is intended to the dictionary of training, and X is intended to training, the sparse coefficient matrix of Y correspondence dictionary.Dimension when matrix
Time the highest, even if using computer software also to be difficult to solution matrix equation, and this algorithm solves what higher dimensional matrix solved just
Problem.
Summary of the invention
It is an object of the invention to provide a kind of based on multichannel image fusion method, it is possible to realize multi-source image is carried out
Efficient image co-registration.
The technical solution realizing the object of the invention is: a kind of based on multichannel image interfusion method, including following
Step:
Step 1), the feature extraction of source images Multichannel Decomposition: source images is carried out Multichannel Decomposition, obtain the knot of image
Structure component (structure) and texture component (texture);
Step 2), fusion rule based on rarefaction representation is arranged: image is after Multichannel Decomposition, during image co-registration
Introduce sparse representation theory.According to the super complete dictionary of structure, the image to be fused of input is carried out piecemeal vectorization arranged side by side, so
Rear calculating each column vector rarefaction representation coefficient under dictionary;
Step 3), coefficient merges: the rarefaction representation coefficient of image block corresponding for each image according to certain fusion
Rule merges, it is thus achieved that the sparse coefficient of image to be reconstructed.
Step 4), image reconstruction: reconstruct is the inverse process of Its Sparse Decomposition, enters with merging sparse coefficient in conjunction with super complete dictionary
Line reconstruction, obtains structural images and the texture image of synthesis, the structure obtained and texture image is summed into fusion image.
Described step 1) particularly as follows: utilize TV-L1 model to carry out manifold source images A and B of two width M × N to be fused
Road decomposes.It is decomposed into 2 major parts: comprise the structure components (structure image) of large scale change and to comprise image little
The details coefficients (texture image) of yardstick texture.First, in catabolic process, original image f is newly defined as newly by we
Image u, v ∈ RM×N, calculate its graded.
Wherein
Therefore, TV-L1 model can be drawn by following formula
According to above-mentioned computational methods, we will be as follows based on TV-L1 model:
min∫D|▽u|dx+λ∫D|u-f|dx (5)
Wherein, u is the structure components of image, and v is the texture component of image.Formula (5) is also a convex optimization problem.First half
TV (the u)=∫ of partD| u | dx can regard image sparse signal | | u | | as2L1Norm, the fit term of latter half,
Also it is l1Norm.Then, (5) formula can be by minimum l1Norm method solves.l1Norm has focused energy and substantially locates to feature,
The feature of the most floating tiny characteristics, TV-L1 model is based on l1Norm, there is l1The advantage of norm, thus TV-L1 model
Preferably performance is had than TV-L2 model at metric space.At metric space, l1The smooth effect of norm be " as far as possible " do
Only, l2Norm is not very clean to the smooth effect of small-scale structure, remains to find out it under the yardstick of only one of which structure
The blurred contour of his structure.The structure that TV-L2 model obtains under different scale is different, thus can be well at metric space
Different structural areas is separated.Along with parameter lambda value tapers into, little structure is smoothed successively, and smooth effect
Fruit is very clean, the key of this effect multi-resolution decomposition just.Meanwhile, details in the structure remained it is clear that this
Point is critically important for picture breakdown.
Through step 1) obtain 4 width images.Structural images S of image AAWith texture image TAAnd the structural images of image B
SBWith texture image TB。
Described step 2) particularly as follows: image is after Multichannel Decomposition, during image co-registration, introduce sparse representation theory.
According to the super complete dictionary of structure, the image to be fused of input is carried out piecemeal vectorization arranged side by side, then calculates each column vector and exist
Rarefaction representation coefficient under dictionary.The process of the method mainly has following two parts:
(1) image column vectorization.Assume that two width sizes are the image of M × N, the sliding window utilizing a size to be 4 × 4
Carry out these images respectively overlapping being sampled as the image block that size is identical, and press row expansion 16 × 1 dimensional vector.
(2) rarefaction representation.The rarefaction representation coefficient of the column vector that each image block is corresponding obtains by solving following formula.
Here consider noise jamming, utilize the image sparse improved to represent model, meanwhile, use SOMP algorithm super complete
Solve sparse coefficient on standby dictionary D, thus obtain the sparse coefficient of each image block.
We are by TA,TB,SA,SBVectorization obtains T respectivelyA1,TB1,SA1,SB1, make T=[TA1,TB1], S=[SA1,SB1],
K-SVD algorithm carries out study respectively obtain D1 T, S are brought into respectively, [XTA1,XTB1] and D2, [XSA1,XSB1]
Described step 3) particularly as follows: the rarefaction representation coefficient of image block corresponding for each image according to certain fusion
Rule merges, it is thus achieved that the sparse coefficient of image to be reconstructed.The normal form the greater i.e. taking two corresponding sparse coefficient is made
For last sparse coefficient.
Assuming that the sparse coefficient that two signal training are obtained is respectively x1 and x2, size is a*b, and fusion rule is:
I.e. take the greater sparse coefficient (A) as the new image merged of two sparse coefficient column vector one normal forms.
If by sparse coefficient [XTA1, XTB1] and [XSA1,XSB1] merge the sparse coefficient that obtains and be respectively A1, A2.
Described step 4) particularly as follows: be the inverse process of Its Sparse Decomposition according to reconstruct, sparse with fusion in conjunction with super complete dictionary
Coefficient is reconstructed, and according to the principle of dictionary learning, the dictionary D1, D2 study obtained is respectively at sparse A1 and the A2 phase merged
Take advantage of.That is:
Y1=D1 × A1, (8)
Y2=D2 × A2; (9)
In (8) formula, Y1 is not the structural images merged, but the vectorization of structural images represents, so using step
Rapid 2) inverse process of vectorization procedure obtains structural images structure merged, and in like manner has (9) formula to obtain the texture merged
Image texture.
So, the final image Image merged i.e. is represented by:
Image=structure+texture. (10)
Accompanying drawing explanation
Fig. 1 is present invention flow chart based on multichannel Image Fusion.
Fig. 2 (a) is the left focusedimage of the inventive method.
Fig. 2 (b) is the right focusedimage of the inventive method.
Fig. 2 (c) is the structural images of left focusedimage.
Fig. 2 (d) is the texture image of left focusedimage.
Fig. 2 (e) is the structural images of right focusedimage.
Fig. 2 (f) is the texture image of right focusedimage.
Fig. 2 (g) is the fusion structure image that sliding window step-length is 8 of 8 × 8.
Fig. 2 (h) is the fusion texture image that sliding window step-length is 8 of 8 × 8.
Fig. 2 (i) be the sliding window step-length of 8 × 8 be the final fusion image of 8.
Fig. 2 (j) is the fusion structure image that sliding window step-length is 4 of 8 × 8.
Fig. 2 (k) is the fusion texture image that sliding window step-length is 4 of 8 × 8.
Fig. 2 (l) be the sliding window step-length of 8 × 8 be the final fusion image of 4.
Fig. 2 (m) is the fusion structure image that sliding window step-length is 2 of 8 × 8.
Fig. 2 (n) is the fusion texture image that sliding window step-length is 2 of 8 × 8.
Fig. 2 (o) be the sliding window step-length of 8 × 8 be the final fusion image of 2.
Detailed description of the invention
Below in conjunction with the accompanying drawings the detailed description of the invention of the present invention is described in further detail.As shown in figure (1), including
Following steps:
Step 1 prepares image M × N (M=480, N=640 in the present embodiment) source images A and B to be fused, utilizes TV-
Picture breakdown is become structure components and texture component by L1 model.TV-L1 model is as follows:
min∫D|▽u|dx+λ∫D|u-f|dx
Wherein, u is the structure components of image, and v is the texture component of image.Formula (5) is also a convex optimization problem.First half
TV (the u)=∫ of partD| u | dx can regard image sparse signal | | u | | as2L1Norm, the fit term of latter half,
Also it is l1Norm.
Step 2 fusion rule based on rarefaction representation is arranged: image, after Multichannel Decomposition, draws during image co-registration
Enter sparse representation theory.According to the super complete dictionary of structure, the image to be fused of input is carried out piecemeal vectorization arranged side by side, then
Calculate each column vector rarefaction representation coefficient under dictionary;The process of the method mainly has following two parts:
(1) image column vectorization.Two width sizes are the image of 480 × 640, the sliding window utilizing a size to be 8 × 8
Carry out these images respectively overlapping being sampled as the image block that size is identical, and press row expansion 64 × 1 dimensional vector.Concrete grammar
For:
Xn (::, n)=X ((i*step)+1:(i*step+8), (j*step)+1:(j*step+8));
Wherein, Xn (::, n) represent n-component column vector, X represents original image, and i, j represent line number and columns, step respectively
Representing step-length, step-length is set as 8 in this example, the different specifications such as 4,2.Constantly fine along with step-length, the fusion knot obtained
Fruit is also become better and better, and the time used is more and more longer.
(2) rarefaction representation.The rarefaction representation coefficient of the column vector that each image block is corresponding obtains by solving following formula.
Here consider noise jamming, utilize the image sparse improved to represent model, meanwhile, use SOMP algorithm super complete
Solve sparse coefficient on standby dictionary D, thus obtain the sparse coefficient of each image block.
Step 3) coefficient fusion:
Assuming that the sparse coefficient that two signal training are obtained is respectively x1 and x2, size is a*b, and fusion rule is:
I.e. take the greater sparse coefficient (A) as the new image merged of two sparse coefficient column vector one normal forms.
Step 4 image reconstruction: according to the principle of dictionary learning, the dictionary D1, D2 study obtained is dilute respectively at merge
Dredge A1 with A2 to be multiplied.That is:
Y1=D1 × A1, (8)
Y2=D2 × A2; (9)
In (8) formula, Y1 is not the structural images merged, but the vectorization of structural images represents, so using step
Rapid 2) inverse process of vectorization procedure obtains structural images structure merged, and in like manner has (9) formula to obtain the texture merged
Image texture.
So, the final image Image merged i.e. is represented by:
Image=structure+texture.
Below in conjunction with example, the present invention is further detailed explanation.
Focusedimage as left in Fig. 2 (a) the inventive method, shown in the right focusedimage of Fig. 2 (b) the inventive method, first distinguishes
Read right focusedimage A and left focusedimage B, it is carried out TV-L1 decomposition, obtain structural images S1 corresponding to A with B
(structure image), texture image T1 (texture image) and S2, T2, i.e. Fig. 2 (c), Fig. 2 (d), Fig. 2 (e) and figure
2(f).S1, T1, S2 and T2 are carried out respectively vectorization and obtains S11, T11, S22 and T22;Again by [S11, S22] and [T11,
T22] carry out K-SVD dictionary training and obtain D1, [X11, X12] and D2, [X21, X22].Then carry out step 3) coefficient merge,
[X11, X12] is fused to X1, [X21, X22] is fused to X2.Finally D1 × the X1 obtained and D2 × X2 is carried out contravariant vector
Change to such as Fig. 2 (g), Fig. 2 (h) (step-length is 8), Fig. 2 (j), Fig. 2 (k) (step-length is 4) and Fig. 2 (m), Fig. 2 (n), (step-length is
2).Finally its structural images and texture image are separately summed design sketch Fig. 2 (i) that set obtains merging, Fig. 2 (l), Fig. 2
(o).It can be seen that from Fig. 2 (i) to Fig. 2 (l) again to Fig. 2 (o), the effect of fusion is become better and better, merge increasingly finer, merge
Required time is more and more longer.
Claims (5)
1. an image interfusion method based on Multichannel Decomposition, it is characterised in that comprise the following steps:
Step 1), the feature extraction of source images Multichannel Decomposition: source images is carried out Multichannel Decomposition, obtain the structure of original image
Component (structure image) and texture component (texture image);
Step 2), fusion rule based on rarefaction representation is arranged: image carries out piecemeal vectorization after Multichannel Decomposition, to vector
Multiple piecemeals after change carry out rarefaction representation, learn super complete dictionary, then calculate each column vector rarefaction representation under dictionary
Coefficient;
Step 3), coefficient merges: the rarefaction representation coefficient of image block corresponding for each image according to certain fusion rule
Merge, it is thus achieved that the sparse coefficient of image to be reconstructed;
Step 4), image reconstruction: reconstruct is the inverse process of Its Sparse Decomposition, carry out weight in conjunction with super complete dictionary with merging sparse coefficient
Structure, obtains structural images and the texture image of synthesis, the structure obtained and texture image is summed into fusion image.
Image interfusion method based on Multichannel Decomposition the most according to claim 1, it is characterised in that described step 1) bag
Include:
Two width source images A and B to be fused are carried out Multichannel Decomposition respectively, is decomposed into 2 major parts: comprise large scale change
Structure components (structure image) and comprise the details coefficients (texture image) of image little yardstick texture.
Based on TV-L1 model, source images being carried out Multichannel Decomposition, TV-L1 model is as follows:
Wherein, u is the structure components of image, and v is the texture component of image.
Image interfusion method based on Multichannel Decomposition the most according to claim 1, it is characterised in that described step 2) bag
Include: image, after Multichannel Decomposition, carries out rarefaction representation respectively.According to the super complete dictionary of structure, by the figure to be fused of input
As carrying out piecemeal vectorization arranged side by side, then calculate each column vector rarefaction representation coefficient under dictionary.The process of the method is main
There are following two parts:
(1) image column vectorization.Assuming that two width sizes are the image of m × n, the sliding window utilizing a size to be 4 × 4 is respectively
Carry out these images overlapping and be sampled as the image block that size is identical, and press row expansion 16 × 1 dimensional vector.
(2) dictionary training.For the matrix signal of two vectorizations of input, carried out K-SVD dictionary learning training.
Y=Dx
Wherein, Y represents the matrix signal through vectorization, and D represents the dictionary that study obtains, and x represents the sparse system that study obtains
Number.
(3) rarefaction representation.The rarefaction representation coefficient of the column vector that each image block is corresponding obtains by solving following formula.
Here consider noise jamming, utilize the image sparse improved to represent model, meanwhile, use SOMP algorithm at super complete word
Solve sparse coefficient in allusion quotation D, thus obtain the sparse coefficient of each image block.
Image interfusion method based on Multichannel Decomposition the most according to claim 1, it is characterised in that described step 3) bag
Include following steps:
Assuming that the sparse coefficient that two signal training are obtained is respectively x1 and x2, size is a*b, and fusion rule is:
If matrix x=[x1, x2],
I.e. take the greater sparse coefficient (A) as the new image merged of two sparse coefficient column vector one normal forms.
Image interfusion method based on Multichannel Decomposition the most according to claim 1, it is characterised in that described step 4) bag
Include following steps:
It is the inverse process of Its Sparse Decomposition according to reconstruct, the dictionary D that will obtain1With the coefficient x merged1It is multiplied and obtains Y1, then by this
Matrix signal is reduced to image, i.e. obtains structural images structure merged, obtains the texture maps merged by same method
As texture, if original image is I, then have:
I=structure+texture.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610291857.2A CN106023122A (en) | 2016-05-05 | 2016-05-05 | Image fusion method based on multi-channel decomposition |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN201610291857.2A CN106023122A (en) | 2016-05-05 | 2016-05-05 | Image fusion method based on multi-channel decomposition |
Publications (1)
Publication Number | Publication Date |
---|---|
CN106023122A true CN106023122A (en) | 2016-10-12 |
Family
ID=57081684
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN201610291857.2A Pending CN106023122A (en) | 2016-05-05 | 2016-05-05 | Image fusion method based on multi-channel decomposition |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN106023122A (en) |
Cited By (6)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN107680070A (en) * | 2017-09-15 | 2018-02-09 | 电子科技大学 | A kind of layering weight image interfusion method based on original image content |
CN108305245A (en) * | 2017-12-29 | 2018-07-20 | 上海交通大学医学院附属瑞金医院 | A kind of analysis of image data method |
CN109671044A (en) * | 2018-12-04 | 2019-04-23 | 重庆邮电大学 | A kind of more exposure image fusion methods decomposed based on variable image |
CN110019865A (en) * | 2017-09-08 | 2019-07-16 | 北京京东尚科信息技术有限公司 | Mass picture processing method, device, electronic equipment and storage medium |
CN110232699A (en) * | 2019-05-05 | 2019-09-13 | 淮阴工学院 | A kind of image multilayer feature decomposition method based on the sparse statistical property of L0 |
CN110689513A (en) * | 2019-09-26 | 2020-01-14 | 石家庄铁道大学 | Color image fusion method and device and terminal equipment |
-
2016
- 2016-05-05 CN CN201610291857.2A patent/CN106023122A/en active Pending
Non-Patent Citations (2)
Title |
---|
汪秀莉 等: "基于TV-L1模型的多尺度遥感影像分解方法", 《海军航空工程学院学报》 * |
罗晖 等: "基于稀疏表示的WMSN红外和可见光图像融合", 《计算机工程与科学》 * |
Cited By (10)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110019865A (en) * | 2017-09-08 | 2019-07-16 | 北京京东尚科信息技术有限公司 | Mass picture processing method, device, electronic equipment and storage medium |
CN110019865B (en) * | 2017-09-08 | 2021-01-26 | 北京京东尚科信息技术有限公司 | Mass image processing method and device, electronic equipment and storage medium |
US11395010B2 (en) | 2017-09-08 | 2022-07-19 | Beijing Jingdong Shangke Information Technology Co., Ltd. | Massive picture processing method converting decimal element in matrices into binary element |
CN107680070A (en) * | 2017-09-15 | 2018-02-09 | 电子科技大学 | A kind of layering weight image interfusion method based on original image content |
CN108305245A (en) * | 2017-12-29 | 2018-07-20 | 上海交通大学医学院附属瑞金医院 | A kind of analysis of image data method |
CN108305245B (en) * | 2017-12-29 | 2021-09-10 | 上海交通大学医学院附属瑞金医院 | Image data analysis method |
CN109671044A (en) * | 2018-12-04 | 2019-04-23 | 重庆邮电大学 | A kind of more exposure image fusion methods decomposed based on variable image |
CN110232699A (en) * | 2019-05-05 | 2019-09-13 | 淮阴工学院 | A kind of image multilayer feature decomposition method based on the sparse statistical property of L0 |
CN110689513A (en) * | 2019-09-26 | 2020-01-14 | 石家庄铁道大学 | Color image fusion method and device and terminal equipment |
CN110689513B (en) * | 2019-09-26 | 2022-09-02 | 石家庄铁道大学 | Color image fusion method and device and terminal equipment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN106023122A (en) | Image fusion method based on multi-channel decomposition | |
CN111179167B (en) | Image super-resolution method based on multi-stage attention enhancement network | |
CN101980284B (en) | Two-scale sparse representation-based color image noise reduction method | |
CN111325165B (en) | Urban remote sensing image scene classification method considering spatial relationship information | |
CN113888744A (en) | Image semantic segmentation method based on Transformer visual upsampling module | |
CN105741252A (en) | Sparse representation and dictionary learning-based video image layered reconstruction method | |
CN113345082B (en) | Characteristic pyramid multi-view three-dimensional reconstruction method and system | |
CN112967178B (en) | Image conversion method, device, equipment and storage medium | |
CN105844653A (en) | Multilayer convolution neural network optimization system and method | |
CN112651316B (en) | Two-dimensional and three-dimensional multi-person attitude estimation system and method | |
CN111241963B (en) | First person view video interactive behavior identification method based on interactive modeling | |
CN113506224A (en) | Image restoration method based on multi-scale generation countermeasure network | |
CN115984494A (en) | Deep learning-based three-dimensional terrain reconstruction method for lunar navigation image | |
CN105303545A (en) | Fusion method for multispectral image and panchromatic satellite image | |
CN115482382A (en) | Image semantic segmentation method based on Transformer architecture | |
CN104008536A (en) | Multi-focus noise image fusion method based on CS-CHMT and IDPCNN | |
CN115345866B (en) | Building extraction method in remote sensing image, electronic equipment and storage medium | |
Mishra et al. | Self-FuseNet: data free unsupervised remote sensing image super-resolution | |
CN109558880B (en) | Contour detection method based on visual integral and local feature fusion | |
CN116797787B (en) | Remote sensing image semantic segmentation method based on cross-modal fusion and graph neural network | |
CN107133921A (en) | The image super-resolution rebuilding method and system being embedded in based on multi-level neighborhood | |
CN104123707A (en) | Local rank priori based single-image super-resolution reconstruction method | |
CN108305219A (en) | A kind of image de-noising method based on uncorrelated sparse dictionary | |
CN115705493A (en) | Image defogging modeling method based on multi-feature attention neural network | |
CN117252892B (en) | Automatic double-branch portrait matting device based on light visual self-attention network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
C06 | Publication | ||
PB01 | Publication | ||
C10 | Entry into substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20161012 |