CN114757830A - Image super-resolution reconstruction method based on channel-diffusion double-branch network - Google Patents
Image super-resolution reconstruction method based on channel-diffusion double-branch network Download PDFInfo
- Publication number
- CN114757830A CN114757830A CN202210488529.7A CN202210488529A CN114757830A CN 114757830 A CN114757830 A CN 114757830A CN 202210488529 A CN202210488529 A CN 202210488529A CN 114757830 A CN114757830 A CN 114757830A
- Authority
- CN
- China
- Prior art keywords
- convolution
- layer
- diffusion
- channel
- convolution layer
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000009792 diffusion process Methods 0.000 title claims abstract description 94
- 238000000034 method Methods 0.000 title claims abstract description 28
- 230000004913 activation Effects 0.000 claims abstract description 29
- 238000012549 training Methods 0.000 claims abstract description 24
- 230000003044 adaptive effect Effects 0.000 claims abstract description 21
- 238000012360 testing method Methods 0.000 claims abstract description 14
- 238000005070 sampling Methods 0.000 claims abstract description 10
- 238000011176 pooling Methods 0.000 claims abstract description 8
- 208000037170 Delayed Emergence from Anesthesia Diseases 0.000 claims abstract description 6
- 230000006870 function Effects 0.000 claims description 30
- 238000010586 diagram Methods 0.000 claims description 10
- 239000004576 sand Substances 0.000 claims description 7
- 238000000605 extraction Methods 0.000 claims description 5
- 238000013507 mapping Methods 0.000 claims description 5
- 230000009466 transformation Effects 0.000 claims description 3
- 238000004364 calculation method Methods 0.000 claims description 2
- 230000003321 amplification Effects 0.000 claims 1
- 238000003199 nucleic acid amplification method Methods 0.000 claims 1
- 238000004088 simulation Methods 0.000 description 6
- 230000000694 effects Effects 0.000 description 2
- 238000003384 imaging method Methods 0.000 description 2
- 230000008569 process Effects 0.000 description 2
- 238000012545 processing Methods 0.000 description 2
- 238000013459 approach Methods 0.000 description 1
- 238000004422 calculation algorithm Methods 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 230000007246 mechanism Effects 0.000 description 1
- 238000011084 recovery Methods 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/04—Architecture, e.g. interconnection topology
- G06N3/045—Combinations of networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06N—COMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
- G06N3/00—Computing arrangements based on biological models
- G06N3/02—Neural networks
- G06N3/08—Learning methods
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T7/00—Image analysis
- G06T7/10—Segmentation; Edge detection
- G06T7/11—Region-based segmentation
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/10—Image acquisition modality
- G06T2207/10024—Color image
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20084—Artificial neural networks [ANN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20132—Image cropping
-
- Y—GENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
- Y02—TECHNOLOGIES OR APPLICATIONS FOR MITIGATION OR ADAPTATION AGAINST CLIMATE CHANGE
- Y02T—CLIMATE CHANGE MITIGATION TECHNOLOGIES RELATED TO TRANSPORTATION
- Y02T10/00—Road transport of goods or passengers
- Y02T10/10—Internal combustion engine [ICE] based vehicles
- Y02T10/40—Engine management systems
Landscapes
- Engineering & Computer Science (AREA)
- Physics & Mathematics (AREA)
- Theoretical Computer Science (AREA)
- General Physics & Mathematics (AREA)
- General Health & Medical Sciences (AREA)
- General Engineering & Computer Science (AREA)
- Biophysics (AREA)
- Computational Linguistics (AREA)
- Data Mining & Analysis (AREA)
- Evolutionary Computation (AREA)
- Artificial Intelligence (AREA)
- Molecular Biology (AREA)
- Computing Systems (AREA)
- Biomedical Technology (AREA)
- Life Sciences & Earth Sciences (AREA)
- Mathematical Physics (AREA)
- Software Systems (AREA)
- Health & Medical Sciences (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention discloses an image super-resolution reconstruction method of a channel-diffusion double-branch network, which mainly solves the problems of insufficient texture details and structural distortion of the existing method and comprises the following implementation steps: constructing a training sample set and a testing sample set; constructing a channel-diffusion double branch network: constructing a channel-diffusion double-branch network of a first convolution layer, D channel-diffusion residual modules, a second convolution layer and an up-sampling module which are connected in sequence; the channel-diffusion residual block comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and are arranged in parallel; the adaptive convolution module comprises a plurality of convolution layers and a plurality of nonlinear activation layers; the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer; the channel attention branch comprises a pooling layer, a plurality of nonlinear activation layers and a plurality of convolution layers; and performing iterative training on the channel-diffusion double-branch network. The method can obtain clear and accurate super-resolution reconstructed images.
Description
Technical Field
The invention belongs to the technical field of image processing, relates to an image reconstruction method, and particularly relates to an image super-resolution reconstruction method based on a channel-diffusion double-branch network, which can be used in the technical fields of pedestrian re-identification and the like.
Background
In the image acquisition process, due to the limitation of factors such as imaging equipment, shooting distance, light and the like, the resolution of the shot picture is often too low, and the quality is poor. In order to obtain higher resolution images, super-resolution reconstruction techniques are generally employed. Image super-resolution reconstruction is a technique for generating a high-resolution image from a low-resolution image. In the field with strict requirements on imaging quality, such as pedestrian re-identification, the image is required to have higher resolution, and structural distortion and edge texture loss do not occur in the image so as to prevent identification errors. At present, three methods, namely interpolation-based method, reconstruction-based method and learning-based method, are mainly used as common super-resolution methods. The image restored by the interpolation method has the phenomena of blurring, sawtooth and the like. Compared with the method based on interpolation, the method based on reconstruction makes great progress, but the effect is still poor. The main idea of the learning-based super-resolution algorithm is to learn the correspondence between the low-resolution image and the high-resolution image, and to guide the super-resolution reconstruction of the images according to the correspondence.
In recent years, deep learning is rapidly developed, and many researchers combine the super-resolution reconstruction with the super-resolution reconstruction of images to achieve effective results. For example, in the patent document "super-resolution reconstruction system and method for single image" (patent application No. 202010218624.6, application publication No. CN 111402140 a) applied by the university of metrology in china, a super-resolution reconstruction method for single image is proposed, which includes: extracting features from the original low-resolution image through the two convolution layers by using an embedded network; reconstructing high-resolution residual error characteristics from the low-resolution characteristics by using two cascaded fine extraction blocks through a coarse-to-fine method; the reconstructed high-resolution residual error characteristics are sent to a reconstruction network to obtain a residual error image through deconvolution operation; and adding the up-sampled low-resolution image and the high-resolution residual image to obtain a finally reconstructed high-resolution image. Although the resolution of the reconstructed image is improved, all information in the image is processed identically, important high-frequency information such as texture details, structures and spatial positions is not recovered, and further improvement of the image reconstruction performance is limited.
Disclosure of Invention
The invention aims to provide an image super-resolution reconstruction method based on channel-diffusion double branches aiming at overcoming the defects of the prior art, and aims to strengthen the extraction capability of high-frequency information through a channel-diffusion double branch network so as to obtain a reconstructed image with more accurate structure and more complete edge details.
In order to achieve the purpose, the technical scheme of the invention comprises the following steps:
(1) constructing a training sample set and a testing sample set:
(1a) acquiring N RGB images, and performing 1/4 downsampling on each RGB image to obtain N downsampled RGB images, wherein N is larger than or equal to 100;
(1b) respectively cutting the N RGB images into image blocks with the size of L multiplied by L to obtain H image blocks in total, and simultaneously cutting the down-sampled RGB image corresponding to each RGB image into the image blocks with the size of L multiplied by LObtaining H down-sampled image blocks, and using each clipped image block as the corresponding down-sampled clipped image blockSelecting M downsampled image blocks and labels corresponding to the image blocks to form a training sample set R, and forming other downsampled image blocks and labels corresponding to the image blocks to form a test sample set E, wherein L is larger than or equal to 192, and M is larger than or equal to 1/2H;
(2) building a channel-diffusion double-branch network model O:
constructing a channel-diffusion double-branch network model O of a first convolution layer, D channel-diffusion residual modules, a second convolution layer and an up-sampling module which are connected in sequence; the channel-diffusion residual block comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and are arranged in parallel; the adaptive convolution module comprises a plurality of convolution layers and a plurality of nonlinear activation layers; the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer; the channel attention branch comprises a pooling layer, a plurality of nonlinear active layers and a plurality of convolution layers;
(3) performing iterative training on a channel-diffusion double-branch network model O:
(3a) the number of initialization iterations is S, the maximum number of iterations is S, S is more than or equal to 10000, and the channel-diffusion double branch network model of the S-th iteration is Os,OsThe weight and the bias parameter are w respectivelys、bsAnd let s be 1, Os=O;
(3b) Taking a training sample set R as an input of a channel-diffusion double-branch network O, and performing feature extraction on each training sample by using a first convolution layer; n extracted by D channel-diffusion residual module pairs1Performing feature mapping on the amplitude feature map to obtain n1A non-linear characteristic diagram, a second convolution layer pair n1Extracting the features of the nonlinear feature map, and extracting n from the first convolution layer1Adding the amplitude characteristic diagrams element by element; the up-sampling module performs up-sampling and dimension transformation on the added images to obtain M super-resolution reconstruction images, wherein n is1The number of convolution kernels of the first convolution layer;
(3c) computing a loss function using the L1 norm, and computing O from each reconstructed image and its corresponding training sample labelsLoss value L ofsSeparately calculating L by the chain rulesTo networkWeight parameter omega insAnd bias parameter bsPartial derivatives ofAndand according toFor omegas、bsUpdating to obtain a channel-diffusion double-branch network model of the iteration;
(3d) judging whether S is more than or equal to S, if so, obtaining a trained channel-diffusion double-branch network model O, otherwise, making S be S +1, and executing the step (3 b);
(4) acquiring an image super-resolution reconstruction result:
and (3) performing forward propagation by taking the test sample set E as the input of the trained channel-diffusion double-branch network model O to obtain reconstructed images corresponding to all the test samples.
Compared with the prior art, the invention has the following advantages:
the channel-diffusion double-branch network model O constructed by the method comprises D channel-diffusion residual modules, each channel-diffusion residual module comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and are arranged in parallel, in the process of training the model and obtaining an image super-resolution reconstruction result, the adaptive convolution has a larger receptive field so as to learn the characteristics of an image more comprehensively, the channel attention branch can distribute different weights to different channel characteristics so as to strengthen the expression of important semantics, and the diffusion branch can distribute different weights to different space information so as to strengthen the recovery of important areas.
Drawings
FIG. 1 is a flow chart of an implementation of the present invention;
FIG. 2 is a schematic diagram of a channel-diffusion double branch network according to the present invention;
FIG. 3 is a schematic diagram of a channel-diffusion residual module according to an embodiment of the present invention;
FIG. 4 is a schematic structural diagram of an adaptive convolution module according to an embodiment of the present invention;
FIG. 5 is a schematic structural diagram of a P-M diffusion layer according to an embodiment of the present invention.
Detailed Description
The invention is described in further detail below with reference to the figures and specific examples.
Referring to fig. 1, the present invention includes the steps of:
step 1: constructing a training sample set and a testing sample set:
(1a) acquiring N RGB images, and performing 1/4 downsampling on each RGB image to obtain N downsampled RGB images, wherein N is larger than or equal to 100;
(1b) respectively cutting the N RGB images into image blocks with the size of L multiplied by L to obtain H image blocks in total, and simultaneously cutting the down-sampled RGB image corresponding to each RGB image into the image blocks with the size of L multiplied by LObtaining H downsampled image blocks, taking each cropped image block as a label of the corresponding downsampled cropped image block, selecting M downsampled image blocks and corresponding labels thereof to form a training sample set R, and forming the rest downsampled image blocks and corresponding labels thereof to form a test sample set E, wherein L is more than or equal to 192, and M is more than or equal to 1/2H;
step 2: a channel-diffusion double-branch network O is built, and the structure of the channel-diffusion double-branch network O is shown in figure 2;
constructing a channel-diffusion double-branch network of a first convolution layer, D channel-diffusion residual modules, a second convolution layer and an up-sampling module which are connected in sequence; the channel-diffusion residual error module comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and arranged in parallel, wherein D is more than or equal to 10, and in the embodiment, D is more than or equal to 10; the adaptive convolution module comprises a plurality of convolution layers and a plurality of nonlinear activation layers; the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer; the channel attention branch comprises a pooling layer, a plurality of nonlinear activation layers and a plurality of convolution layers;
the structure of the channel-diffusion residual error module of the present embodiment is shown in fig. 3;
let the input of the channel-diffusion residual block be MD-1Firstly, a group of finer features are generated by self-adaptive convolution and are marked as W; and simultaneously inputting W into the channel attention branch and the diffusion branch to respectively generate two groups of attention weighted feature maps, carrying out weighting operation on the two groups of attention weighted feature maps and the W to recalibrate the features, and adding the calibrated features and the W to obtain the output of the channel-diffusion residual error module.
The structure of the adaptive convolution module of the present embodiment is shown in fig. 4;
the adaptive convolution structure comprises two branches which are arranged in parallel, the first branch comprises a third convolution layer, a first nonlinear activation function, a weighted image generation module and a fourth convolution layer which are arranged in parallel, a fifth convolution layer and a second nonlinear activation function which are connected in sequence, the weighted image generation module comprises a sixth convolution layer and a seventh convolution layer which are arranged in parallel and a third nonlinear activation function connected with the sixth convolution layer and the seventh convolution layer, and the second branch comprises an eighth convolution layer, a fourth nonlinear activation function, a ninth convolution layer and a fifth nonlinear activation function which are connected in sequence; in the weighted image generation module, the feature map is downsampled before the sixth convolutional layer to obtain an image with a smaller size, so that the convolutional layer focuses on larger spatial information to obtain more complete related features, the feature map is upsampled to obtain an image with the same size as the input feature map, the feature map is upsampled before the seventh convolutional layer to obtain an image with a smaller size, so that the convolutional layer focuses on detailed information to obtain finer features, the feature map is downsampled to obtain an image with the same size as the input feature map, the two images are added, the added images are subjected to weight calculation through a Sigmoid function, and the feature map output by the fourth convolutional layer is recalibrated to obtain more comprehensive features.
The structure of the P-M diffusion layer of this embodiment is shown in FIG. 5;
inspired by P-M diffusion, a P-M diffusion layer is designed, the layer is an improved residual block, the input tensor is W, and the P-M diffusion equation is as follows:
wherein For the diffusion coefficient, t is the diffusion step, k is the shape definition constant,can be expressed as:
where the constant k is a threshold. It is easy to find that in the flat or smooth areas where the gradient is smallCoefficient of diffusionWill approach 1 and in areas of rich texture or structural detailThe coefficients will be close to zero. Thus, with such spatially varying diffusion coefficients, the P-M diffusion mechanism can preserve image detail while removing noise.
There are the following identities:
Wαα+Wββ=Wxx+Wyy (4)
wherein ,Wαα and WββSecond order partial derivatives of W in the direction of image gradient and in the direction perpendicular to the image features (edges), respectivelyNumber, Wxx and WyyThe second partial derivatives of W along the coordinate axes x and y, respectively.
In addition, WββCan be expressed as:
wherein ,Wx and WyFirst partial derivatives of W in the x and y directions, WxyIs WxPartial derivatives in the y-direction.
Derived using equation (4):
can be obtained by substituting formula (2):
the vicinity of the edge is where the gradient mode is large, and the gradient direction is not smoothed as much as possible in order to better retain the edge. Therefore, let WααThe coefficient being zero, i.e.If the diffusion step Δ t is set to 1, it can be obtained by substituting equation (6):
the above equation can be expressed as a residual learning block because the right side of the equation can be considered as aw. In each block, Δ W is added to the input tensor Wi。Wx、Wy、Wxy、Andthe residual error results are calculated from the twelfth convolution layer, the thirteenth convolution layer, the fourteenth convolution layer, the fifteenth convolution layer, and the sixteenth convolution layer, respectively, according to equation (7).
Passing W through P-M diffusion layer to obtain W + Δ W, and passing W through a weight of ΩDA1 x 1 convolution layer, and then rescaling the attention map to [0,1 ] using sigmoid function σ (·)]As shown in the following formula:
MDA=σ(ΩDA(W+ΔW)) (8)
wherein fDA(. o) means the multiplication of an input W by an element of the obtained DA map, WDAIs the result of the recalibration of the DA module. Each MDA(i, j) corresponds to the DA weight of the spatial location (i, j) of W.
And step 3: performing iterative training on the channel-diffusion double-branch network O:
(3a) the number of initialization iterations is S, the maximum number of iterations is S, S is more than or equal to 10000, and the channel-diffusion double-branch network model of the S-th iteration is Os,OsThe weight and bias parameters of the middle learnable parameter are w respectivelys、bsAnd let s be 1, Os=O;
(3b) Taking a training sample set R as an input of a channel-diffusion double-branch network O, and performing feature extraction on any training sample by using a first convolution layer; n extracted from D channel-diffusion residual module pairs1Performing feature mapping on the amplitude feature map; n obtained by second convolution layer to characteristic mapping1Extracting the features of the nonlinear feature map, and extracting n from the first convolution layer1Adding the amplitude characteristic diagrams element by element; the up-sampling module comprises PixelShuffle and a convolutional layer, PixelShuffle to n1The amplitude characteristic diagram is obtained after 4 times of upsamplingReconstructing the image, and setting the number of the channels of the convolution kernel to be 1, namely reconstructing the imageConverting the reconstructed image into dimension transformation to obtain a super-resolution reconstructed image, wherein n is1Performing the above processing on the M images for the number of convolution kernels of the first convolution layer and the second convolution layer;
(3c) computing a loss function using the L1 norm, and computing O from each reconstructed image and its corresponding training sample labelsLoss value L ofsSeparately calculating L by the chain rulesFor weight parameter omega in networksAnd bias parameter bsPartial derivatives ofAndand according toFor omegas、bsIs updated according toFor omegas、bsThe updating formulas for updating are respectively as follows:
wherein ,representing the reconstructed image, I representing the labels of the samples in the training sample set, ws、bsRepresents OsWeight, bias parameter, w of all learnable parameterss'、bs' indicates updated learnable parameters,/rIndicates the learning rate, LsIs a function of the loss as a function of,representing a derivative operation.
(3d) Judging whether S is more than or equal to S, if so, obtaining a trained channel-diffusion double-branch network model O, otherwise, making S equal to S +1, and executing the step (3 b);
and 4, step 4: acquiring an image reconstruction result:
and performing forward propagation by taking the test sample set E as the input of the trained channel-diffusion double-branch network model O to obtain reconstructed images corresponding to all the test samples.
The technical effects of the present invention can be further illustrated by the following simulation experiments
1. Simulation conditions and contents:
the hardware platform of the simulation experiment is as follows: the processor is an Intel (R) Core i9-9900K CPU, the main frequency is 3.6GHz, the memory is 32GB, and the display card is NVIDIA GeForce RTX 2080 Ti. The software platform of the simulation experiment is as follows: ubuntu 16.04 operating system, python version 3.7, pitorch version 1.7.1.
The RGB image dataset used in the simulation experiment was the DIV2K dataset. The training data Set used in the simulation experiment is the DIV2K data Set, and the test Set is the Set5 data Set.
The peak snr of the prior art data Set at Set5 was 37.78dB, and the peak snr of the invention data Set at Set5 was 38.14dB, respectively, with the results shown in table 1. Compared with the prior art, the peak signal-to-noise ratio is remarkably improved.
TABLE 1
Method | Prior Art | The invention |
PSNR | 37.78dB | 38.14dB |
SSIM | 0.9042 | 0.9612 |
Claims (3)
1. An image super-resolution reconstruction method based on a channel-diffusion double-branch network is characterized by comprising the following steps:
(1) constructing a training sample set and a testing sample set:
(1a) acquiring N RGB images, and performing 1/4 downsampling on each RGB image to obtain N downsampled RGB images, wherein N is larger than or equal to 100;
(1b) respectively cutting N RGB images into image blocks of L multiplied by L size to obtain H image blocks in total, and simultaneously cutting the down-sampled RGB image corresponding to each RGB image into image blocks of L multiplied by L sizeObtaining H down-sampled image blocks, and using each clipped image block as a corresponding down-sampled and clipped imageSelecting M downsampled image blocks and labels corresponding to the image blocks to form a training sample set R, and forming other downsampled image blocks and labels corresponding to the image blocks to form a test sample set E, wherein L is larger than or equal to 192, and M is larger than or equal to 1/2H;
(2) constructing a channel-diffusion double-branch network O:
constructing a channel-diffusion double-branch network O comprising a first convolution layer, D channel-diffusion residual modules, a second convolution layer and an up-sampling module which are sequentially connected; the channel-diffusion residual block comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and are arranged in parallel; the adaptive convolution module comprises a plurality of convolution layers and a plurality of nonlinear activation layers; the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer; the channel attention branch comprises a pooling layer, a plurality of nonlinear active layers and a plurality of convolution layers, and D is more than or equal to 10;
(3) performing iterative training on the channel-diffusion double-branch network O:
(3a) the number of initialization iterations is S, the maximum number of iterations is S, S is more than or equal to 10000, and the channel-diffusion double branch network model of the S-th iteration is Os,OsThe weight and the bias parameter are w respectivelys、bsAnd let s be 1, Os=O;
(3b) Taking a training sample set R as an input of a channel-diffusion double-branch network O, and performing feature extraction on each training sample by using a first convolution layer; n extracted by D channel-diffusion residual module pairs1Performing feature mapping on the amplitude feature map; n obtained by second convolution layer to characteristic mapping1Extracting the characteristic of the nonlinear characteristic diagram, and extracting n from the first convolution layer1Adding the amplitude characteristic diagrams element by element; the up-sampling module adds the obtained n1Performing dimension transformation after up-sampling the characteristic map to obtain a super-resolution reconstruction image, wherein n is1The number of convolution kernels of the first convolution layer and the second convolution layer;
(3c) calculating a loss function by adopting an L1 norm, and calculating O through each reconstructed image and a corresponding training sample labelsLoss value L ofsSeparately calculating L by the chain rulesFor weight parameter omega in networksAnd bias parameter bsPartial derivatives ofAndand according toFor omegas、bsUpdating to obtain a channel-diffusion double-branch network model of the iteration;
(3d) judging whether S is more than or equal to S, if so, obtaining a trained channel-diffusion double-branch network model O, otherwise, making S equal to S +1, and executing the step (3 b);
(4) obtaining an image reconstruction result:
and (3) performing forward propagation by taking the test sample set E as the input of the trained channel-diffusion double-branch network model O to obtain reconstructed images corresponding to all the test samples.
2. The method for reconstructing an image with super resolution based on a channel-diffusion double branch network according to claim 1, wherein the channel-diffusion double branch network O in step (2), wherein:
constructing a channel-diffusion double-branch network O of a first convolution layer, D channel-diffusion residual modules, a second convolution layer and an up-sampling module which are connected in sequence; the channel-diffusion residual block comprises an adaptive convolution module and a channel attention branch and a diffusion branch which are connected with the adaptive convolution module and are arranged in parallel; the adaptive convolution module comprises a plurality of convolution layers and a plurality of nonlinear activation layers; the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer; the channel attention branch comprises a pooling layer, a plurality of nonlinear activation layers and a plurality of convolution layers;
the number n of convolution kernels of the first convolution layer and the second convolution layer164, the convolution kernel size is 3 x 3;
the number D of the channel-diffusion residual modules is 10; the number of convolution layers in the self-adaptive convolution module is 7, the number of the nonlinear activation function layers is 5, the self-adaptive convolution module comprises two branches which are arranged in parallel, the first branch comprises a third convolution layer, a first nonlinear activation function, a weighted image generation module, a fourth convolution layer, a fifth convolution layer and a second nonlinear activation function which are sequentially connected, the weighted image generation module comprises a sixth convolution layer and a seventh convolution layer which are arranged in parallel and a third nonlinear activation function connected with the sixth convolution layer and the seventh convolution layer, the second branch comprises an eighth convolution layer, a fourth nonlinear activation function, a ninth convolution layer and a fifth nonlinear activation function which are sequentially connected; the specific parameters are as follows: the third convolution layer, the eighth convolution layer, the fifth convolution layer, the sixth convolution layer, the seventh convolution layer, and the ninth convolution layer have a convolution kernel size of 1 × 1, and the fourth convolution layer, the fifth convolution layer, the sixth convolution layer, the seventh convolution layer, and the ninth convolution layer have a convolution kernel size of 3 × 3; the first nonlinear activation function, the second nonlinear activation function and the fifth nonlinear activation function are realized by a ReLU function, and the third nonlinear activation function is realized by a Sigmoid function;
the number of the convolutional layers contained in the channel attention branch is 2, the number of the nonlinear active layers is 2, and the specific structure of the channel attention branch is a pooling layer, a tenth convolutional layer, a sixth nonlinear active layer, an eleventh convolutional layer and a seventh nonlinear active layer which are sequentially cascaded; the specific parameters are as follows: the convolution kernels of the tenth convolution layer and the eleventh convolution layer are 1 x 1 in size, the pooling layer is set to be maximum pooling, the sixth nonlinear active layer is realized by a ReLU function, and the seventh nonlinear active layer is realized by a Sigmoid function;
the specific structure of the diffusion branch comprises a P-M diffusion layer, a convolution layer and a nonlinear activation layer which are sequentially connected, wherein the P-M diffusion layer comprises a twelfth convolution layer, a thirteenth convolution layer, a fourteenth convolution layer, a fifteenth convolution layer and a sixteenth convolution layer which are arranged in parallel; the specific parameters are as follows: the convolution kernel size of the convolution layer between the P-M diffusion layer and the nonlinear active layer is 1 x 1, the convolution kernel size of the twelfth convolution layer, the thirteenth convolution layer and the fourteenth convolution layer is 3 x 3, the convolution kernel size of the fifteenth convolution layer and the sixteenth convolution layer is 5 x 5, and the nonlinear active function is realized by a Sigmoid function;
the upsampling module is implemented by PixelShuffle, and the amplification parameter is 4.
3. The method for reconstructing image super resolution based on channel-diffusion double branch network of claim 1, wherein the L1 norm calculation loss function L in step (3c)sAnd according toFor omegas、bsThe updating formulas for updating are respectively as follows:
wherein ,representing the reconstructed image, I representing the labels of the samples in the training sample set, ws、bsRepresents OsWeight, bias parameter, w of all learnable parameterss'、bs' indicates updated learnable parameters,/rIndicates the learning rate, LsIs a function of the loss of the signal,representing a derivative operation.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210488529.7A CN114757830B (en) | 2022-05-06 | 2022-05-06 | Image super-resolution reconstruction method based on channel-diffusion double-branch network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202210488529.7A CN114757830B (en) | 2022-05-06 | 2022-05-06 | Image super-resolution reconstruction method based on channel-diffusion double-branch network |
Publications (2)
Publication Number | Publication Date |
---|---|
CN114757830A true CN114757830A (en) | 2022-07-15 |
CN114757830B CN114757830B (en) | 2023-09-08 |
Family
ID=82332334
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202210488529.7A Active CN114757830B (en) | 2022-05-06 | 2022-05-06 | Image super-resolution reconstruction method based on channel-diffusion double-branch network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN114757830B (en) |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112991181A (en) * | 2021-03-31 | 2021-06-18 | 武汉大学 | Image super-resolution reconstruction method based on reaction diffusion equation |
CN113177882A (en) * | 2021-04-29 | 2021-07-27 | 浙江大学 | Single-frame image super-resolution processing method based on diffusion model |
CN113222822A (en) * | 2021-06-02 | 2021-08-06 | 西安电子科技大学 | Hyperspectral image super-resolution reconstruction method based on multi-scale transformation |
US20210390338A1 (en) * | 2020-06-15 | 2021-12-16 | Dalian University Of Technology | Deep network lung texture recogniton method combined with multi-scale attention |
-
2022
- 2022-05-06 CN CN202210488529.7A patent/CN114757830B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20210390338A1 (en) * | 2020-06-15 | 2021-12-16 | Dalian University Of Technology | Deep network lung texture recogniton method combined with multi-scale attention |
CN112991181A (en) * | 2021-03-31 | 2021-06-18 | 武汉大学 | Image super-resolution reconstruction method based on reaction diffusion equation |
CN113177882A (en) * | 2021-04-29 | 2021-07-27 | 浙江大学 | Single-frame image super-resolution processing method based on diffusion model |
CN113222822A (en) * | 2021-06-02 | 2021-08-06 | 西安电子科技大学 | Hyperspectral image super-resolution reconstruction method based on multi-scale transformation |
Non-Patent Citations (1)
Title |
---|
彭宏京;侯文秀;: "基于各向异性扩散偏微分方程的图像去模糊", 信号处理, no. 05, pages 714 - 717 * |
Also Published As
Publication number | Publication date |
---|---|
CN114757830B (en) | 2023-09-08 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN110119780B (en) | Hyper-spectral image super-resolution reconstruction method based on generation countermeasure network | |
CN108090871B (en) | Multi-contrast magnetic resonance image reconstruction method based on convolutional neural network | |
CN109035142B (en) | Satellite image super-resolution method combining countermeasure network with aerial image prior | |
Chen et al. | Multi-attention augmented network for single image super-resolution | |
CN110660038A (en) | Multispectral image and panchromatic image fusion method based on generation countermeasure network | |
CN109003229B (en) | Magnetic resonance super-resolution reconstruction method based on three-dimensional enhanced depth residual error network | |
CN108492249B (en) | Single-frame super-resolution reconstruction method based on small convolution recurrent neural network | |
CN111275192B (en) | Auxiliary training method for improving accuracy and robustness of neural network simultaneously | |
Chen et al. | Single image super-resolution using deep CNN with dense skip connections and inception-resnet | |
CN114494015B (en) | Image reconstruction method based on blind super-resolution network | |
CN115953303B (en) | Multi-scale image compressed sensing reconstruction method and system combining channel attention | |
CN111738954B (en) | Single-frame turbulence degradation image distortion removal method based on double-layer cavity U-Net model | |
CN111696038A (en) | Image super-resolution method, device, equipment and computer-readable storage medium | |
CN113689517A (en) | Image texture synthesis method and system of multi-scale channel attention network | |
CN112991483B (en) | Non-local low-rank constraint self-calibration parallel magnetic resonance imaging reconstruction method | |
CN114723608A (en) | Image super-resolution reconstruction method based on fluid particle network | |
CN114626984A (en) | Super-resolution reconstruction method for Chinese text image | |
CN110675318B (en) | Sparse representation image super-resolution reconstruction method based on main structure separation | |
CN113379647B (en) | Multi-feature image restoration method for optimizing PSF estimation | |
CN112581626B (en) | Complex curved surface measurement system based on non-parametric and multi-attention force mechanism | |
Wen et al. | The power of complementary regularizers: Image recovery via transform learning and low-rank modeling | |
CN113096015A (en) | Image super-resolution reconstruction method based on progressive sensing and ultra-lightweight network | |
CN113240581A (en) | Real world image super-resolution method for unknown fuzzy kernel | |
Cheng et al. | Adaptive feature denoising based deep convolutional network for single image super-resolution | |
CN114757830B (en) | Image super-resolution reconstruction method based on channel-diffusion double-branch network |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |