CN115170398A - Image super-resolution reconstruction method and device for chrysanthemum storage warehouse - Google Patents

Image super-resolution reconstruction method and device for chrysanthemum storage warehouse Download PDF

Info

Publication number
CN115170398A
CN115170398A CN202210853635.0A CN202210853635A CN115170398A CN 115170398 A CN115170398 A CN 115170398A CN 202210853635 A CN202210853635 A CN 202210853635A CN 115170398 A CN115170398 A CN 115170398A
Authority
CN
China
Prior art keywords
module
feature
chrysanthemum
feature map
output
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202210853635.0A
Other languages
Chinese (zh)
Other versions
CN115170398B (en
Inventor
杨雪梅
李黎
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Yunyang Yunshan Agricultural Development Co ltd
Original Assignee
Chongqing Yunshan Industrial Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing Yunshan Industrial Co ltd filed Critical Chongqing Yunshan Industrial Co ltd
Priority to CN202210853635.0A priority Critical patent/CN115170398B/en
Publication of CN115170398A publication Critical patent/CN115170398A/en
Application granted granted Critical
Publication of CN115170398B publication Critical patent/CN115170398B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4038Image mosaicing, e.g. composing plane images from plane sub-images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2200/00Indexing scheme for image data processing or generation, in general
    • G06T2200/32Indexing scheme for image data processing or generation, in general involving image mosaicing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • Biophysics (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Biomedical Technology (AREA)
  • Health & Medical Sciences (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • General Engineering & Computer Science (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses an image super-resolution reconstruction method and equipment for a chrysanthemum storage warehouse. According to the method, in the advanced information refining module, information of different spatial positions of the characteristic diagram is calibrated by using two spatial attention mechanisms, then information on different channels of the characteristic diagram is secondarily calibrated by using a channel attention mechanism, the characteristic diagram input into the attention module contains a large amount of differentiated characteristic information, and a test result shows that the CNN network provided by the invention has a good reconstruction effect on-site chrysanthemum images.

Description

Image super-resolution reconstruction method and device for chrysanthemum storage warehouse
Technical Field
The invention belongs to the technical field of chrysanthemum and image processing, and particularly relates to an image super-resolution reconstruction method and device for a chrysanthemum storage warehouse.
Background
After being processed, the chrysanthemum is stored in a warehouse to wait for classified packaging. In the storage process of the chrysanthemum, the chrysanthemum is influenced by various environmental factors and may deteriorate. Moreover, once deteriorated, the chrysanthemum flowers will be infected around, resulting in extended loss. The camera can be installed at the top of the warehouse, and the state of the chrysanthemum in the warehouse is monitored by combining the computer vision technology. However, the distance between the camera and the chrysanthemum is far, and the image is limited by resolution, so that the deteriorated chrysanthemum only occupies a small area, and the target detection network has low identification precision.
Disclosure of Invention
The invention provides an image super-resolution reconstruction method and equipment for a chrysanthemum storage warehouse.
In order to solve the technical problems, the invention adopts the following technical scheme: an image super-resolution reconstruction method for a chrysanthemum storage warehouse comprises the following steps:
the method comprises the following steps that P100, a CNN network is built on a computer according to a designed structure, the CNN network is sequentially provided with a low-level information refining module, a high-level information refining module and a super-resolution mapping module along the depth direction, and the high-level information refining modules are sequentially connected;
the mathematical model of the advanced information extraction module is as follows:
N1=γ1(j1c3(T m ))
N2=γ2(j1d3(T m ))
N3=fu(N1,N2)
N4=jb(N3,S m )*N1
N5=jd(N3,L m )*N2
C m+1 =jc(||N4,N5||,C m )×||N4,N5||
S m+1 =γ3(j2c3(C m+1 ))
L m+1 =γ4(j2d3(C m+1 ))
T m+1 =γ5(j3c1(||S m+1 ,L m+1 ||))
wherein, T m Feature graph representing input of advanced information refining modules from the core input, S m Feature graph, L, representing input of advanced information refinement modules from a first bypass input m A characteristic diagram representing the input of a high-level information refinement module from a second bypass input, C m A characteristic diagram representing the high-level information extraction module input from the middle input end; j1C3 () and j2C3 () each represent a normal convolution operation with a convolution kernel size of 3 x 3, j1d3 () and j2d3 () each represent a dilated convolution operation with a convolution kernel size of 3 x 3 and a dilation rate of 2, j3C1 () represents a normal convolution operation with a convolution kernel size of 1 x 1, γ 1 (), γ 2 (), γ 3 (), γ 4 (), and γ 5 () each represent an activation function ReLU, jb () represents a first bypass spatial attention module, jd () represents a second bypass spatial attention module, and represents a feature map that is de-aligned using a spatial alignment map (i.e., the spatial alignment map is multiplied by the element correspondence of each layer of the feature map), jc represents a middle channel attention module, and x represents a feature map that is de-aligned using a channel alignment map (i.e., the feature map is multiplied by the corresponding value in the channel alignment map in each layer of the feature map), and | represents a concatenation operation with N () of the feature map, and N () each layer of the feature map and N2N m+1 Feature graph representing the output of advanced information refining modules from the intermediate output, S m+1 Feature graph, L, representing the output of advanced information refinement modules from a first bypass output m+1 Feature graph, T, representing the output of a high level information refinement module from a second bypass output m+1 A feature graph representing the high-level information extraction module output from the core output;
p200, acquiring a training data set, and training the CNN network by using the training data set, wherein the training data set comprises high-resolution warehouse chrysanthemum images and low-resolution warehouse chrysanthemum images corresponding to the high-resolution warehouse chrysanthemum images;
p300, acquiring a field chrysanthemum image needing to be reconstructed, inputting the field chrysanthemum image into the CNN network trained in the step P200, wherein the field chrysanthemum image sequentially passes through the low-level information refining module and each high-level information refining module, and then the core output end of the last high-level information refining module outputs a comprehensive characteristic diagram;
and P400, inputting the comprehensive characteristic map into the super-resolution mapping module, performing super-resolution reconstruction on the comprehensive characteristic map by the super-resolution mapping module, and outputting a target chrysanthemum image with the resolution being greater than that of the field chrysanthemum image.
Furthermore, an intermediate information fusion module is arranged in the CNN network, and the intermediate information fusion module fuses the feature information output by the middle output end of each advanced information refining module with the comprehensive feature map and then inputs the feature information into the super-resolution mapping module.
Further, the mathematical model of the intermediate information fusion module is as follows:
CCS=jr 1 c1(C 1 )+jr 2 c1(C 2 )+...+jr m+1 c1(C m+1 )
CSTL=||CCS,CVT||
wherein, C 1 、C 2 、…、C m+1 A characteristic diagram, jr, representing the output of the central output of each of said advanced information refining modules 1 c1()、jr 2 c1()、…、jr m+1 c1 The values (c) all represent common convolution operation with convolution kernel size of 1 x 1, the value of | | · | | represents splicing operation of the feature diagram therein, the CVT represents a comprehensive feature diagram, and the CSTL represents the feature diagram output after fusion operation of the intermediate information fusion module.
Further, the fusion process of the feature map N1 and the feature map N2 is expressed as the following mathematical formula:
Nt=Ta1(j1t(N1))+Ta2(j2t(N2))+||N1,N2||
N3=Lc(j3t(Nt))
wherein, | | · | represents that the feature maps therein are subjected to splicing operation, j1t (), j2t () and j3t () all represent common convolution operation with a convolution kernel size of 1 × 1, ta1 () and Ta2 () both represent Tanh activation functions, lc () represents logics activation functions, and N3 represents the feature map obtained by fusing the feature map N1 with the feature map N2.
Further, the mathematical model of the first bypass space attention module is:
Figure BDA0003739746290000031
PV2=SMe(S m )
PV3=SMe(N3)
PV4=SVr(N3)
S1T=δ1(ja1c(||PV1,PV2,PV3,PV4||))
wherein, the characteristic diagram S m And feature map N3 is the input to the first bypass space attention module, SMa () represents a full channel max pooling operation on the feature map (i.e., a global max pooling operation in the channel direction), SMe () represents a full channel mean pooling operation on the feature map (i.e., a global mean pooling operation in the channel direction), SVr () represents a full channel variance pooling operation on the feature map (i.e., a global variance pooling operation in the channel direction),
Figure BDA0003739746290000041
the method comprises the steps of representing a product corresponding to elements, | · | | represents splicing operation of feature graphs in the product, ja1c () represents common convolution operation with a convolution kernel of 1 × 1, delta 1 () represents a sigmoid function, and S1T represents a first bypass space calibration graph output by a first bypass space attention module.
Further, the mathematical model of the second bypass spatial attention module is:
Figure BDA0003739746290000042
ZO2=SMe(L m )
PV3=SMe(N3)
PV4=SVr(N3)
S2T=δ2(ja2c(||ZO1,ZO2,PV3,PV4||))
wherein, the characteristic diagram L m And feature map N3 is the input to the second bypass spatial attention module, SMa () representing a pairThe feature map is subjected to full-channel maximum pooling, SMe () represents the full-channel average pooling of the feature map, SVr () represents the full-channel variance pooling of the feature map,
Figure BDA0003739746290000043
the corresponding product of elements is represented, | · | | represents that the feature graph is subjected to splicing operation, ja2c () represents common convolution operation with the convolution kernel size of 1 × 1, δ 2 () represents a sigmoid function, and S2T represents a second bypass space calibration graph output by the second bypass space attention module.
Further, the mathematical model of the middle channel attention module is:
CW1=CMaK(||N4,N5||)+CMaK(jw1c(C m ))
CTW=δ3(FC2(γ6(FC1(CW1))))
wherein, the characteristic diagram is N4, N5 and the characteristic diagram C m For the input of the middle channel attention module, | | · | | represents that the feature maps therein are subjected to splicing operation, CMaK () represents that full-space MaK pooling operation is performed on each layer of the feature maps, jw1c () represents general convolution operation with a convolution kernel size of 1 × 1, FC1 () and FC2 () both represent full join operation, γ 6 () represents an activation function ReLU, δ 3 () represents a sigmoid function, and CTW represents a channel calibration map output by the middle channel attention module.
Since the conventional ultrasound image contains considerable noise, the noise will cause a large interference to the center channel attention module, which is likely to cause a wrong calibration. The present invention employs a full-space MaK pooling operation in the middle aisle attention module. Further, the process of the full-space MaK pooling operation includes the following steps:
p1, arranging elements in the matrix according to a sequence from large to small to obtain a sequence U;
p2, obtaining numerical values of the first three elements in the sequence U, and calculating the result according to the following formula:
GK=0.6Lag1+0.3Lag2+0.1Lag3
wherein, lag1 represents the numerical value of the first element in the sequence U, lag2 represents the numerical value of the second element in the sequence U, lag3 represents the numerical value of the third element in the sequence U, and GK represents the output value after the MaK pooling operation of the full space. Comparative experimental tests show that the middle channel attention module performs best when the three parameters are set to 0.6, 0.3 and 0.1 respectively.
The invention also provides an image super-resolution reconstruction device for the chrysanthemum storage warehouse, which comprises a processor and a memory, wherein the memory stores a computer program, and the processor is used for executing the image super-resolution reconstruction method for the chrysanthemum storage warehouse by loading the computer program.
The invention has the beneficial effects that:
(1) In the advanced information extraction module, firstly, two space attention mechanisms are adopted to calibrate information of different space positions of the characteristic diagram, and then a channel attention mechanism is adopted to calibrate information on different channels of the characteristic diagram for the second time, so that the calibration pertinence is strong, and the quality of characteristic information in a convolution operation output characteristic diagram is greatly improved;
(2) j1c3 () and j1d3 () are arranged in parallel, and feature information of images is extracted on different visual fields, so that key information under a plurality of visual fields is reserved in the fused N3 feature map, partial invalid information is eliminated, the N3 feature map is used as one input of a space attention module, visual information with different visual fields is provided for the space attention module, the visual information is richer, and the generated first bypass space calibration map and the second bypass space calibration map have a more accurate space information calibration effect;
(3) The spatial attention module not only takes the N3 feature map as input, but also takes the feature map output by the upstream advanced information refining module (S) m /L m ) As input, the spatial attention module has different levels of characteristic information, and the differentiated level information improves the quality of information acquired by the spatial attention module, so that the accuracy of the spatial information calibration by the spatial calibration graph is improved; similarly, the middle channel attention module takes the feature map in the current-level advanced information refinement module and the feature map in the previous-level module as input at the same time, compared with the single information inputThe output channel calibration chart can better calibrate the information of different channels;
(4) In the spatial attention module, after the full-channel maximum pooling operation is respectively carried out on the two characteristic diagrams, the two matrixes are subjected to element corresponding product, and compared with the splicing and dimension reduction of the two matrixes, the spatial attention module can strengthen important information to a greater extent.
Drawings
Fig. 1 is a schematic diagram of a CNN network structure according to the present invention;
fig. 2 is a schematic structural diagram of an advanced information refining module in the CNN network shown in fig. 1;
FIG. 3 is a schematic diagram of the fusion process of the feature map N1 and the feature map N2;
FIG. 4 is a block diagram of a first bypass space attention module of the advanced information mining module of FIG. 2;
FIG. 5 is a block diagram of a middle channel attention module of the advanced information mining module of FIG. 2;
fig. 6 is a schematic structural diagram of a super-resolution mapping module in the CNN network shown in fig. 1;
FIG. 7 is a schematic diagram of a modified first bypass spatial attention module of comparative example 1;
FIG. 8 is a schematic structural view of a modified central passage attention module of comparative example 2;
in the drawings:
the method comprises the steps of 1-field chrysanthemum image, 2-low-level information refining module, 3-high-level information refining module, 31-first bypass space attention module, 32-second bypass space attention module, 33-middle channel attention module, 4-super-resolution mapping module, 41-preset convolution layer, 42-Pixelshuffle layer, 43-rear convolution layer, 5-middle information fusion module and 6-target chrysanthemum image.
Detailed Description
The present invention will be described in further detail with reference to the accompanying drawings.
Example 1:
based on a TensorFlow framework, a CNN network for reconstructing chrysanthemum images in a storage warehouse is realized by programming on a computer, and the specific structure is shown in FIG. 1. The CNN network includes a low-level information extraction module 2, a high-level information extraction module 3, an intermediate information fusion module 5, and a super-resolution mapping module 4.
The low-level information refining module 2 is realized by adopting common convolution operation with the convolution kernel size of 3 x 3, and the number of output characteristic diagram channels is 32 after the on-site chrysanthemum image 1 is input into the low-level information refining module 2 and is subjected to convolution operation.
The number of the high-level information extraction modules 3 is set to 6, and the specific structure of the high-level information extraction modules 3 is shown in fig. 2. Each advanced information refining module 3 receives as input a plurality of profiles, where T m The profile (number of channels 32) is input from its core input, S m The profile (number of channels 32) is input from its first bypass input, L m The characteristic diagram (number of channels 32) is input from its second bypass input, C m The characteristic diagram is input from the middle input end thereof. For the first high-level information extraction module 3, the feature maps input from the four input ends thereof are all the outputs of the low-level information extraction module 2.
Each advanced information refining module 3 will also output a plurality of characteristic diagrams, C m+1 The characteristic diagram (the number of channels is 64) is output from the middle output end thereof, C m+1 The feature map is not only input into the intermediate information fusion module 5, C m+1 The profile is also input to the mid-lane attention module 33 in the downstream high-level information refinement module 3. S m+1 The signature (number of lanes 32) is output from its first bypass output, S m+1 The characteristic diagram is used as the input of the first bypass input end of the downstream advanced information refining module 3, L m+1 The signature (number of channels 32) is output from its second bypass output, L m+1 The feature map is used as an input of a second bypass input end of the downstream high-level information refining module 3, T m+1 The feature map (number of channels 32) is output from its core output, T m+1 The feature map is used as the input of the core input end of the downstream high-level information refining module 3.
T m The dimensions of the feature map, the N1 feature map, the N2 feature map and the N3 feature map are identical. Please refer to fig. 3When the N1 feature map is fused with the N2 feature map, three operations are firstly performed, that is, the N1 feature map is spliced with the N2 feature map (the number of channels of the output feature map is 64), the N1 feature map is subjected to 1 × 1 convolution and Tanh function activation (the number of channels of the output feature map is 64), and the N2 feature map is subjected to 1 × 1 convolution and Tanh function activation (the number of channels of the output feature map is 64). And then adding the three feature maps, activating by a j3t () convolution and a Logitics function, and outputting to obtain an N3 feature map. In the process of fusing the N1 characteristic diagram and the N2 characteristic diagram, the invention simultaneously adopts two modes of splicing and adding, and in the process, the dimension of the N1 characteristic diagram and the dimension of the N2 characteristic diagram are increased and then reduced, thereby well overcoming the respective defects of the two fusion modes of splicing and adding and improving the characteristic fusion effect.
The operation process inside the first bypass space attention module 31 and the second bypass space attention module 32 is almost the same, please refer to fig. 4, which takes the first bypass space attention module 31 as an example. Characteristic diagram S m After the first bypass spatial attention module 31 is inputted, two matrices are obtained through the full pass maximum pooling operation and the full pass average pooling operation. Similarly, after the feature map N3 is inputted into the first bypass space attention module 31, three matrices are obtained through the full-channel maximum pooling operation, the full-channel average pooling operation, and the full-channel variance pooling operation. And then performing element corresponding product operation on the matrix output by the maximum pooling operation of the two full channels to obtain PV1. And finally, splicing PV1, PV2, PV3 and PV4, and sequentially performing ja1c () convolution and delta 1 () activation to obtain a first bypass space calibration graph (a two-dimensional matrix).
Referring to fig. 5, after the N4 feature map and the N5 feature map are spliced and input to the central channel attention module 33, each layer is output to obtain a vector with a length of 64 through a full-space MaK pooling operation. Characteristic diagram C m When the input data is input to the middle channel attention module 33, the jw1c () convolution operation is performed, then the full-space MaK pooling operation is performed on each layer, and a vector with the length of 64 is obtained through the same output. For the first high-level information refining module 3, the number of channels of the characteristic diagram before the convolution operation is 32, and the number of channels after the convolution is 64; for other high-level information refinement modules 3, jw1c () pre-and post-convolution profilesThe number of channels is 64. After the two vectors are added, the channel calibration graph is obtained after sequentially passing through FC1 (), gamma 6 (), FC2 () and delta 3 (). For FC1 (), its input node number is 64, and its output node number is 12; for FC2 (), its input node number is 12 and its output node number is 64.
In order to avoid information loss and gradient disappearance, the CNN network of this embodiment is provided with an intermediate information fusion module 5, and a feature map C output by the middle output end of the advanced information refining module 3 m+1 After 1 × 1 convolution (the number of channels is 64), the sum is added, and then the sum is spliced with the comprehensive characteristic diagram to be used as the input of the super-resolution mapping module 4. It is emphasized that the characteristic diagram of the input intermediate information fusion module 5 in this embodiment comes from the middle output (C) m+1 ) Rather than the core output (T) m+1 ) Therefore, the feature map input into the super-resolution mapping module 4 can contain more differentiated information, and the quality of the reconstructed target chrysanthemum image 6 can be improved.
As shown in fig. 6, the super-resolution mapping module 4 includes a pre-modulation convolution layer 41, a pixelsuffle layer 42, and a post-modulation convolution layer 43 connected in sequence. When the image resolution is increased by a multiple of A, the number of channels of the output feature map of the pre-modulation convolution layer 41 is 32A 2 The Pixelshuffle layer 42 outputs a signature having 32 channels, but has a length-wise and width-wise pixel value a times that of the output signature of the preconditioning convolutional layer 41. Finally, the post-tone convolutional layer 43 outputs the target chrysanthemum image 6 with the number of channels being 3.
568 high-resolution warehouse chrysanthemum images are collected and obtained, 150 images are randomly extracted to serve as a test set, and the rest images serve as a training set. And manufacturing a low-resolution warehouse chrysanthemum image corresponding to the high-resolution warehouse chrysanthemum image by adopting a bicubic downsampling mode. The CNN network, the CS-NL model (from the article Image Super-Resolution with Cross-Scale Non-Local extension and extreme Self-extensions Mining) and the SRFeat (from the article SRFeat: single Image Super-Resolution with Feature discovery) provided by the present invention are respectively trained by using the same training set, and then the results are shown as follows:
the test results on the same test set are as follows:
Figure BDA0003739746290000091
Figure BDA0003739746290000101
as can be seen from the measurement results of the two parameters of the peak signal-to-noise ratio and the structural similarity, compared with the existing model, the CNN network provided in embodiment 1 can better reconstruct a high-resolution chrysanthemum image of a warehouse site, which is beneficial to more accurately detect and discover a quality-changed chrysanthemum.
Example 2:
in order to verify the beneficial effects of the spatial attention module and the full-space MaK pooling operation provided by the present invention, the first bypass spatial attention module 31 and the second bypass spatial attention module 32 are both configured as a conventional structure, and the modified spatial attention module structure is shown in fig. 7. The resulting model was named the example 2A network without changing the rest of the CNN network in example 1.
Similarly, also on the basis of embodiment 1, the full-space MaK pooling operation in the middle channel attention module 33 is modified to a conventional full-space max pooling operation, and the modified channel attention module structure is shown in fig. 8. The resulting model was named example 2B network without changing the rest of the CNN network in example 1.
The example 2A network and the example 2B network were trained on the same training set, respectively, and then tested on the test set, respectively, with the following results:
Figure BDA0003739746290000102
Figure BDA0003739746290000111
according to the measurement results of two parameters of the peak signal-to-noise ratio and the structural similarity, after the spatial attention module and the full-space MaK pooling operation provided by the invention are replaced by the prior art, the performance of the network reconstruction of the field chrysanthemum image 1 is reduced to different degrees, and the spatial attention module and the full-space MaK pooling operation provided by the invention are well proved to have positive effects on improving the chrysanthemum image reconstruction quality.
The above is only a preferred embodiment of the present invention, and it should be noted that several modifications and improvements made by those skilled in the art without departing from the technical solution should also be considered as falling within the scope of the claims.

Claims (9)

1. An image super-resolution reconstruction method for a chrysanthemum storage warehouse is characterized by comprising the following steps: the method comprises the following steps:
p100, building a CNN network on a computer according to a designed structure, wherein the CNN network is sequentially provided with a low-level information refining module, a high-level information refining module and a super-resolution mapping module along the depth direction, and the high-level information refining modules are sequentially connected;
the mathematical model of the advanced information extraction module is as follows:
N1=γ1(j1c3(T m ))
N2=γ2(j1d3(T m ))
N3=fu(N1,N2)
N4=jb(N3,S m )*N1
N5=jd(N3,L m )*N2
C m+1 =jc(||N4,N5||,C m )×||N4,N5||
S m+1 =γ3(j2c3(C m+1 ))
L m+1 =γ4(j2d3(C m+1 ))
T m+1 =γ5(j3c1(||S m+1 ,L m+1 ||))
wherein, T m Feature graph representing input of advanced information refinement modules from core inputs,S m Feature graph, L, representing input of advanced information refinement modules from a first bypass input m A characteristic diagram representing the input of a high-level information refinement module from a second bypass input, C m A characteristic diagram representing the high-level information extraction module input from the middle input end; j1C3 () and j2C3 () each represent a normal convolution operation with a convolution kernel size of 3 x 3, j1d3 () and j2d3 () each represent a dilated convolution operation with a convolution kernel size of 3 x 3 and a dilation rate of 2, j3C1 () represents a normal convolution operation with a convolution kernel size of 1 x 1, γ 1 (), γ 2 (), γ 3 (), γ 4 () and γ 5 () each represent an activation function ReLU, jb () represents a first bypass space attention module, jd () represents a second bypass space attention module, which represents a feature map that is de-aligned using a space alignment map, jc () represents a middle channel attention module, x represents a feature map that is de-aligned using a channel alignment map, | | | | represents a feature map in which a splicing operation is performed, fu () represents a feature map N1 fused with a feature map N2, C m+1 Feature graph representing the output of advanced information refining modules from the intermediate output, S m+1 Feature graph, L, representing the output of advanced information refinement modules from a first bypass output m+1 Feature graph, T, representing the output of a high level information refinement module from a second bypass output m+1 A feature graph representing the high-level information refinement module output from the core output;
p200, acquiring a training data set, and training the CNN network by using the training data set, wherein the training data set comprises high-resolution warehouse chrysanthemum images and low-resolution warehouse chrysanthemum images corresponding to the high-resolution warehouse chrysanthemum images;
p300, acquiring a field chrysanthemum image needing to be reconstructed, inputting the field chrysanthemum image into the CNN network trained in the step P200, wherein the field chrysanthemum image sequentially passes through the low-level information refining module and each high-level information refining module, and then the core output end of the last high-level information refining module outputs a comprehensive characteristic diagram;
and P400, inputting the comprehensive characteristic map into the super-resolution mapping module, performing super-resolution reconstruction on the comprehensive characteristic map by the super-resolution mapping module, and outputting a target chrysanthemum image with the resolution being greater than that of the field chrysanthemum image.
2. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 1, wherein: and the CNN network is provided with an intermediate information fusion module which fuses the feature information output by the middle output end of each advanced information refining module with the comprehensive feature map and inputs the feature information and the comprehensive feature map into the super-resolution mapping module.
3. The super-resolution image reconstruction method for the chrysanthemum storage warehouse according to claim 2, wherein the super-resolution image reconstruction method comprises the following steps: the mathematical model of the intermediate information fusion module is as follows:
CCS=jr 1 c1(C 1 )+jr 2 c1(C 2 )+...+jr m+1 c1(C m+1 )
CSTL=||CCS,CVT||
wherein, C 1 、C 2 、…、C m+1 A characteristic diagram, jr, representing the output of the central output of each of said advanced information mining modules 1 c1()、jr 2 c1()、…、jr m+1 c1 The values (c) all represent common convolution operation with convolution kernel size of 1 x 1, the value of | | · | | represents splicing operation of the feature diagram therein, the CVT represents a comprehensive feature diagram, and the CSTL represents the feature diagram output after fusion operation of the intermediate information fusion module.
4. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 1, wherein: the fusion process of the feature map N1 and the feature map N2 is expressed by the following mathematical formula:
Nt=Ta1(j1t(N1))+Ta2(j2t(N2))+||N1,N2||
N3=Lc(j3t(Nt))
wherein, | | · | represents that the feature maps therein are subjected to splicing operation, j1t (), j2t () and j3t () all represent common convolution operation with a convolution kernel size of 1 × 1, ta1 () and Ta2 () both represent Tanh activation functions, lc () represents logics activation functions, and N3 represents the feature map obtained by fusing the feature map N1 with the feature map N2.
5. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 1, wherein: the mathematical model of the first bypass space attention module is:
Figure FDA0003739746280000031
PV2=SMe(S m )
PV3=SMe(N3)
PV4=SVr(N3)
S1T=δ1(ja1c(||PV1,PV2,PV3,PV4||))
wherein, the characteristic diagram S m And a feature map N3 is an input of the first bypass space attention module, SMa () represents a full channel maximum pooling operation on the feature map, SMe () represents a full channel average pooling operation on the feature map, SVr () represents a full channel variance pooling operation on the feature map,
Figure FDA0003739746280000032
the method comprises the following steps of representing a product corresponding to elements, | · | | | represents splicing operation of feature graphs in the feature graphs, ja1c () represents common convolution operation with a convolution kernel of 1 × 1, delta 1 () represents a sigmoid function, and S1T represents a first bypass space calibration graph output by a first bypass space attention module.
6. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 5, wherein: the mathematical model of the second bypass spatial attention module is:
Figure FDA0003739746280000033
ZO2=SMe(L m )
PV3=SMe(N3)
PV4=SVr(N3)
S2T=δ2(ja2c(||ZO1,ZO2,PV3,PV4||))
whereinCharacteristic diagram L m And a feature map N3 is input to the second bypass spatial attention module, SMa () represents a full channel maximal pooling operation on the feature map, SMe () represents a full channel average pooling operation on the feature map, SVr () represents a full channel variance pooling operation on the feature map,
Figure FDA0003739746280000041
the corresponding product of elements is represented, | · | | represents that the feature graph is subjected to splicing operation, ja2c () represents common convolution operation with the convolution kernel size of 1 × 1, δ 2 () represents a sigmoid function, and S2T represents a second bypass space calibration graph output by the second bypass space attention module.
7. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 1, wherein: the mathematical model of the middle channel attention module is as follows:
CW1=CMaK(||N4,N5||)+CMaK(jw1c(C m ))
CTW=δ3(FC2(γ6(FC1(CW1))))
wherein, the characteristic diagram is N4, N5 and the characteristic diagram C m For the input of the middle channel attention module, | | · | | represents that the feature maps therein are subjected to splicing operation, CMaK () represents that full-space MaK pooling operation is performed on each layer of the feature maps, jw1c () represents general convolution operation with a convolution kernel size of 1 × 1, FC1 () and FC2 () both represent full join operation, γ 6 () represents an activation function ReLU, δ 3 () represents a sigmoid function, and CTW represents a channel calibration map output by the middle channel attention module.
8. The image super-resolution reconstruction method for the chrysanthemum storage warehouse according to claim 7, wherein: the process of the full-space MaK pooling operation comprises the following steps:
p1, arranging elements in the matrix according to a descending order to obtain a sequence U;
p2, obtaining numerical values of the first three elements in the sequence U, and calculating the result according to the following formula:
GK=0.6Lag1+0.3Lag2+0.1Lag3
wherein, lag1 represents the numerical value of the first element in the sequence U, lag2 represents the numerical value of the second element in the sequence U, lag3 represents the numerical value of the third element in the sequence U, and GK represents the output value after the MaK pooling operation of the full space.
9. An image super-resolution reconstruction device for a chrysanthemum storage warehouse is characterized in that: comprises a processor and a memory, wherein the memory stores a computer program, and the processor is used for executing the image super-resolution reconstruction method for the chrysanthemum storage warehouse according to any one of claims 1 to 8 by loading the computer program.
CN202210853635.0A 2022-07-11 2022-07-11 Image super-resolution reconstruction method and device for chrysanthemum storage warehouse Active CN115170398B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202210853635.0A CN115170398B (en) 2022-07-11 2022-07-11 Image super-resolution reconstruction method and device for chrysanthemum storage warehouse

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202210853635.0A CN115170398B (en) 2022-07-11 2022-07-11 Image super-resolution reconstruction method and device for chrysanthemum storage warehouse

Publications (2)

Publication Number Publication Date
CN115170398A true CN115170398A (en) 2022-10-11
CN115170398B CN115170398B (en) 2023-10-13

Family

ID=83494606

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202210853635.0A Active CN115170398B (en) 2022-07-11 2022-07-11 Image super-resolution reconstruction method and device for chrysanthemum storage warehouse

Country Status (1)

Country Link
CN (1) CN115170398B (en)

Cited By (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115358931A (en) * 2022-10-20 2022-11-18 运易通科技有限公司 Image reconstruction method and device for warehouse logistics system

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113362223A (en) * 2021-05-25 2021-09-07 重庆邮电大学 Image super-resolution reconstruction method based on attention mechanism and two-channel network
CN114022750A (en) * 2021-10-28 2022-02-08 重庆大学 Welding spot appearance image identification method and system based on aggregation-calibration CNN
CN114626984A (en) * 2022-03-16 2022-06-14 西安理工大学 Super-resolution reconstruction method for Chinese text image
US20220208355A1 (en) * 2020-12-30 2022-06-30 London Health Sciences Centre Research Inc. Contrast-agent-free medical diagnostic imaging
US20220207648A1 (en) * 2020-12-04 2022-06-30 Commonwealth Scientific And Industrial Research Organisation Creating super-resolution images

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20220207648A1 (en) * 2020-12-04 2022-06-30 Commonwealth Scientific And Industrial Research Organisation Creating super-resolution images
US20220208355A1 (en) * 2020-12-30 2022-06-30 London Health Sciences Centre Research Inc. Contrast-agent-free medical diagnostic imaging
CN113362223A (en) * 2021-05-25 2021-09-07 重庆邮电大学 Image super-resolution reconstruction method based on attention mechanism and two-channel network
CN114022750A (en) * 2021-10-28 2022-02-08 重庆大学 Welding spot appearance image identification method and system based on aggregation-calibration CNN
CN114626984A (en) * 2022-03-16 2022-06-14 西安理工大学 Super-resolution reconstruction method for Chinese text image

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
ABDUL MUQEET等: "HRAN: Hybrid residual attention network for single image super-resolution", IEEE ACCESS, vol. 7, pages 137020 - 137029, XP011748410, DOI: 10.1109/ACCESS.2019.2942346 *

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN115358931A (en) * 2022-10-20 2022-11-18 运易通科技有限公司 Image reconstruction method and device for warehouse logistics system
CN115358931B (en) * 2022-10-20 2023-01-03 运易通科技有限公司 Image reconstruction method and device for warehouse logistics system

Also Published As

Publication number Publication date
CN115170398B (en) 2023-10-13

Similar Documents

Publication Publication Date Title
CN109255755B (en) Image super-resolution reconstruction method based on multi-column convolutional neural network
Ahn et al. Image super-resolution via progressive cascading residual network
Hui et al. Progressive perception-oriented network for single image super-resolution
CN112861729B (en) Real-time depth completion method based on pseudo-depth map guidance
CN110378398B (en) Deep learning network improvement method based on multi-scale feature map jump fusion
CN111870245B (en) Cross-contrast-guided ultra-fast nuclear magnetic resonance imaging deep learning method
CN111862251B (en) Method, device, storage medium and electronic equipment for medical image reconstruction technology
US11822900B2 (en) Filter processing device and method of performing convolution operation at filter processing device
CN115170398A (en) Image super-resolution reconstruction method and device for chrysanthemum storage warehouse
CN111667444A (en) Image compressed sensing reconstruction method based on multi-channel residual error network
Gendy et al. Balanced spatial feature distillation and pyramid attention network for lightweight image super-resolution
Wang et al. STCD: efficient Siamese transformers-based change detection method for remote sensing images
Varghese et al. Unpaired image-to-image translation of structural damage
Ma et al. Towards no-reference image quality assessment based on multi-scale convolutional neural network
CN114519718A (en) Abdomen multi-organ CT image segmentation method and system
He et al. Mambaad: Exploring state space models for multi-class unsupervised anomaly detection
Su et al. High-Similarity-Pass Attention for Single Image Super-Resolution
CN114332625A (en) Remote sensing image colorizing and super-resolution method and system based on neural network
Cai et al. Refconv: Re-parameterized refocusing convolution for powerful convnets
Chen et al. Split-guidance network for salient object detection
CN117635444A (en) Depth completion method, device and equipment based on radiation difference and space distance
CN115760670B (en) Unsupervised hyperspectral fusion method and device based on network implicit priori
CN115984949A (en) Low-quality face image recognition method and device with attention mechanism
CN113450364B (en) Tree-shaped structure center line extraction method based on three-dimensional flux model
Zhou et al. Hierarchical Unified Spectral-Spatial Aggregated Transformer for Hyperspectral Image Classification

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant
TR01 Transfer of patent right
TR01 Transfer of patent right

Effective date of registration: 20240416

Address after: Building 1-1-5, No. 599 Binjiang Road, Qinglong Street, Yunyang County, Chongqing, 404599

Patentee after: Yunyang Yunshan Agricultural Development Co.,Ltd.

Country or region after: China

Address before: Group 1, Yongfu Village, Hongshi Town, Yunyang County, Chongqing 404511 (self-promised)

Patentee before: Chongqing Yunshan Industrial Co.,Ltd.

Country or region before: China