CN112734643A - Lightweight image super-resolution reconstruction method based on cascade network - Google Patents

Lightweight image super-resolution reconstruction method based on cascade network Download PDF

Info

Publication number
CN112734643A
CN112734643A CN202110052039.8A CN202110052039A CN112734643A CN 112734643 A CN112734643 A CN 112734643A CN 202110052039 A CN202110052039 A CN 202110052039A CN 112734643 A CN112734643 A CN 112734643A
Authority
CN
China
Prior art keywords
network
image
resolution
channels
convolution
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202110052039.8A
Other languages
Chinese (zh)
Inventor
李浪
陶洋
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Chongqing University of Post and Telecommunications
Original Assignee
Chongqing University of Post and Telecommunications
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Chongqing University of Post and Telecommunications filed Critical Chongqing University of Post and Telecommunications
Priority to CN202110052039.8A priority Critical patent/CN112734643A/en
Publication of CN112734643A publication Critical patent/CN112734643A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4053Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration using two or more images, e.g. averaging or subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20112Image segmentation details
    • G06T2207/20132Image cropping

Landscapes

  • Physics & Mathematics (AREA)
  • General Physics & Mathematics (AREA)
  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention belongs to the technical field of image processing, and mainly relates to a light-weight image super-resolution reconstruction method based on a cascade network. The method mainly comprises the following steps: step 1: acquiring a high-resolution target image of a data set, and performing downsampling pretreatment; step 2: designing a light-weight image super-resolution reconstruction network based on an attention mechanism and a cascade network; and step 3: and designing a training strategy and training parameters and optimizing model parameters. After training and optimization, the invention can provide an image super-resolution reconstruction method with small volume and high reconstruction quality.

Description

Lightweight image super-resolution reconstruction method based on cascade network
The technical field is as follows:
the invention belongs to the technical field of image processing, and mainly relates to a light-weight image super-resolution reconstruction method based on a cascade network.
Background art:
generally, for an image, a higher resolution represents more and richer image details. But is limited by the factors of the size, the cost and the like of the imaging equipment, and the finally presented image often does not meet the actual requirements of people. Under such a background, an image super-resolution reconstruction technique for reconstructing a high-resolution image from a low-resolution image has attracted extensive attention in academic circles, and is widely used in the fields of video surveillance, medicine, and the like. Therefore, the super-resolution reconstruction algorithm with wide research application range and high reconstruction quality has important theoretical value and practical significance.
Most super-resolution reconstruction methods improve the reconstruction effect by continuously stacking the depth of the network, and although the algorithms can often obtain better effect, the network has larger volume and large calculation amount, so that the super-resolution reconstruction methods are difficult to apply to practical application. The conventional convolution is generally used for feature extraction, and with the development of a network model, technologies such as deep separable convolution and packet convolution have achieved prominent achievements in the field of deep learning. The existing light weight network in the field of image super-resolution reconstruction mainly solves the problems from several angles: an effective up-sampling mode and a feature extraction module are designed, and the depth of the network is reduced. Although a relatively good effect is obtained in the case of a relatively light weight, there is still room for further improvement. A lightweight image super-resolution reconstruction based on a cascade network is designed, firstly a lightweight multi-scale feature extraction module is designed, and then different features are weighted by using an attention mechanism. And finally, fusing the channel characteristics of different levels by using a cascading mechanism. The super-resolution reconstruction network designed based on the scheme can meet the requirement of real-time performance, and meanwhile, the quality of the super-resolution reconstruction network is superior to that of the super-resolution reconstruction network with the same level of network complexity.
Disclosure of Invention
The invention provides a light-weight image super-resolution reconstruction method based on a cascade network, which aims at the problems that the existing super-resolution reconstruction method is large in size and difficult to apply in practice, provides an unsupervised super-resolution reconstruction method, and optimizes a reconstruction result through a pyramid generation countermeasure network, so that a super-resolution reconstruction network conforming to image characteristics is obtained.
The technical scheme of the invention is as follows:
s1: and (6) image acquisition. The DIV2K data set is obtained and the PNG data is converted into the npy data format.
S2: on the basis of S1, the. npy data is cropped to crop a high resolution image HR of 128 × 128, and then down-sampled to obtain a low resolution image LR, forming a high and low resolution image pair.
S3: on the basis of S2, using convolution design low-level feature extraction network to extract features, wherein the extracted features are represented as FL
FL=Conv2d3x3(LR,3,64)
Where Conv2d3x3 represents a conventional convolution with a convolution kernel size of 3, in this way shallow features of the network are extracted. Where 3 and 64 represent the number of input and output channels, respectively.
S4: the feature F obtained in S3LAs an input of the depth feature extraction network, a channel 1 performs feature extraction through an FE module as shown in fig. 1, and finally performs channel dimensionality reduction using a standard convolution with a convolution kernel size of 1:
FFE1=Conv1x1(σ(GConv4(DWConv3x3(F,3,64),64,64)),64,32)
where DWConv3x3 represents a deep convolution with a convolution kernel size of 3, GConv4 represents a packet convolution with a packet number of 4, and σ represents the ReLU activation function. Conv1x1 represents a conventional convolution with a convolution kernel size of 1.
And the channel 2 firstly carries out feature extraction in a feature extraction mode of the channel 1, then carries out channel recombination, and then carries out secondary feature extraction through an FE module. The dimensionality reduction of the channels is then performed using the same 1x1 convolution.
FFE2=Conv1x1(σ(GConv4(DWConv3x3(Shufle(σ(GConv4(DWConv3x3(FL,3,64),64,64))),3,64),64,64)),64,32)
The characteristics of the outputs of channel 1 and channel 2 are then tiled to extend to a 64-channel profile and activated using ReLU, adding to the non-linearity of the network. Is shown as
FDF1=σ(Concat(FFE1+FFE2))
In DGSA, an enhanced channel attention mechanism network as shown in fig. 2 is introduced. First, the channel descriptor is validated. Let F (F ∈ i)W×H×CW, H, C is the width, height and channel number of the feature map) to obtain a1 × 1 × C feature map, and the result of pooling is used as the descriptor Z of the channel, and the descriptor of the C-th channel is expressed as:
Figure BDA0002899406930000021
subsequently, channel relationship modeling, with WkTo learn channel attention, expressed as follows:
Wc=σ(WkZ)
Figure BDA0002899406930000022
where k is expressed as the number of calculated adjacent channels, WkExpressed as:
Wk=[ω12,L,ωc]T
Figure BDA0002899406930000023
wherein the content of the first and second substances,
Figure BDA0002899406930000024
representing k adjacent channels
Figure BDA0002899406930000025
A collection of (a). Channel parameter sharing can further reduce parameter quantity, omegaiCan be expressed as:
Figure BDA0002899406930000026
finally, the information interaction between the channels is realized through the one-dimensional convolution with the convolution kernel size k, which is expressed as:
ω=σ(Conv1dkxk(Z))
wherein k is adaptively selected according to the number of channels, and is represented as:
C=φ(k)=22k-1
weighting the learned channel attention mechanism into the input features to obtain a final feature map:
Fout=Z*F
the final DGSA module may be represented as shown in fig. 3. And simultaneously, cascading the features of different layers, and sending the features into a conventional convolution with a convolution kernel size of 1 to perform feature information interaction and channel dimension reduction.
Fcon=Conv1x1(Concat(F1,F2,...),inputchannel,64)
In the formula, inputchanel represents the sum of the channels of each input feature map.
S5: feature F after feature extraction in S4conThe features are fed into an upsampling module as shown in fig. 4. The upsampling module first needs to expand the channel.
Fup=Conv3x3(Fcon,64,64*r*r)
Fup_f=PS(Fup,64*r*r,64)
Where r represents the multiple to be reconstructed and PS represents the Pixel Shuffle operation. If the amplification is four times, the above operation is performed twice, where r is 2.
S6: using the L1 loss as a loss function for network training, an Adam optimizer as an optimizer for the network, the error between SR and HR is calculated for back propagation, and 100 epochs are trained. And updating the parameters.
S7: after training is finished, parameters of the model are fixed, the image to be reconstructed is used as the input of the network, the reconstruction scale is selected, and the output is the super-resolution image of the target.
Drawings
FIG. 1 FE module
FIG. 2 attention mechanism network
FIG. 3 network model
FIG. 4 upsampling module
FIG. 5 comparison of reconstruction results
The specific implementation process comprises the following steps:
a lightweight image super-resolution reconstruction method based on a cascade network is characterized in that a network model is shown in figure 3, data preprocessing is firstly carried out, and then a super-resolution reconstruction network is constructed. And then setting training parameters and conditions to train the network and adjusting the network parameters. And finally, performing super-resolution reconstruction by using the trained model.
The invention is further illustrated by the following example of an embodiment, which is intended only for a better understanding of the subject matter of the invention and is not intended to limit the scope of the invention. The method comprises the following specific steps:
step S1: the example adopts a Set5, a Set14, a BSD100 and an Urban100 four-image super-resolution reconstruction data Set, the Set5 comprises 5 pairs of high-low resolution image pairs, the Set14 extends to 14 sheets, and the BSD100 and the Urban100 respectively comprise 100 sheets of high-low resolution image pairs with various features. The example selects quadrupling as the reference standard. And carrying out downsampling on the low-resolution image to obtain an ultra-low-resolution image pair.
Step S2: the present embodiment takes the low resolution image directly as the input to the network.
Step S3: and confirming the training environment and the strategy. The processor of the experimental device is
Figure BDA0002899406930000031
I99900K, wherein the display card is RTX2080Ti, the training is carried out in a Pythrch environment, and the acceleration model is trained by using CUDA10.0 and CuDNN7.1.
The epoch of the network trained in each scale is 100, the learning rate of a generator and a discriminator is set to be 0.001, the learning rate of the epoch is reduced to 10 percent of the original learning rate every 30, and the scale of the feature image is kept unchanged in the feature extraction process.
Step S4: the present example selects the low resolution images in Set5, Set14, and BSD100 as the reconstruction objects. And taking the low-resolution image to be reconstructed as an input, and directly inputting the low-resolution image into a network for calculation.
The example compares the algorithm proposed herein with the existing unsupervised image super-resolution reconstruction algorithm for 2, 3 and 4 times of reconstruction indexes, and mainly compares two indexes of peak signal-to-noise ratio (PSNR) and Structural Similarity (SSIM). The experimental results are shown in fig. 5 and table 1, taking 4-fold reconstruction as an example.
As can be seen from the observation of the graph, the method is superior to a comparison algorithm in terms of two indexes, and the quality of the reconstructed image is improved to a certain extent.
TABLE 1 reconstruction index comparison Table
Figure BDA0002899406930000041

Claims (8)

1. A light-weight image super-resolution reconstruction method based on a cascade network is characterized by comprising the following steps:
s1: the DIV2K data set is obtained and the PNG data is converted into the npy data format.
S2: on the basis of S1, the. npy data is cropped to crop a high resolution image HR of 128 × 128, and then down-sampled to obtain a low resolution image LR, forming a high and low resolution image pair.
S3: on the basis of S2, using convolution design low-level feature extraction network to extract features, wherein the extracted features are represented as FL
S4: on the basis of S3, a depth feature extraction module DGSA is designed, wherein the DGSA comprises a light-weight volume part and an attention mechanism part. And 4 DGSA cascades form a deep feature extraction network. Extracting a depth feature F using the low-level feature extracted in S3 as an input to a depth feature extraction networkH
S5: on the basis of S4, the extracted depth feature F is usedHThe input upsampling network enlarges the feature map to the size of the target reconstruction. And then, restoring the same channel number of the target image through a reconstruction module, and outputting a target SR.
S6: the error between the HR obtained at S1 and the SR obtained at S5 was calculated using the L1 loss function. And performing gradient negative feedback to update the parameters of the feature extraction layer, the up-sampling layer and the reconstruction layer.
S7: and (5) carrying out multiple iterations on the S2-S6, continuously updating network parameters, and fixing the model after the set training amount is reached. And reconstructing the image to be reconstructed as the input of the model to obtain the super-resolution image SR of the targetT
2. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S1:
the image file of the originally obtained PNG is converted into the npy data, and the file reading speed is accelerated.
3. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S2:
the resulting. npy file is cropped to a crop size of 128 x 128 resolution and downsampled for different reconstruction scales.
4. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S3:
and designing a low-level feature extraction network, wherein the low-level feature extraction network is formed by convolution with a convolution kernel size of 3, the number of input channels is 3, and the number of output channels is 64. The size of the characteristic diagram is unchanged.
5. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S4:
design depth feature extraction module DGSA (Depthwise-Group ShuffleNet with Attention). The DGSA is formed by two channels, namely a main channel 1 and an output channel after being directly subjected to feature extraction by an FE module. The channel 2 firstly carries out first feature extraction through an FE module, then carries out feature channel recombination, changes the position of the channel and enhances the information interaction between the channels. The FE module 64 consists of deep convolution of channels and packet convolution, and the number of packets is 4. And activated using the ReLU function. After the characteristics of the two channels are obtained, convolution with convolution kernel of 1 multiplied by 1 is carried out to carry out residual error connection on the information of the two channels, the information of the channels is fused, and finally, the characteristic F is outputH1=DGSA(FL). Then fusing low-level features and depth features, splicing channels, fusing the channels by using a standard convolution with a convolution kernel of 1 multiplied by 1, and finally outputting FH
6. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S5:
the channel expansion is carried out by convolution with convolution kernel of 3 × 3, then Pixel Shuffle is carried out for feature map amplification, and the number of channels is ensured to be 64, if the amplification factor is 2 and 3, the amplification is carried out only once, and if the amplification factor is 4, the above operation is carried out twice, and the amplification is respectively doubled.
7. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S6:
using the L1 loss as a loss function for network training, an Adam optimizer as an optimizer for the network, the error between SR and HR is calculated for back propagation, and 100 epochs are trained. And updating the parameters.
8. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S7:
after training is finished, parameters of the model are fixed, the image to be reconstructed is used as the input of the network, and the output is the super-resolution image of the target.
CN202110052039.8A 2021-01-15 2021-01-15 Lightweight image super-resolution reconstruction method based on cascade network Pending CN112734643A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202110052039.8A CN112734643A (en) 2021-01-15 2021-01-15 Lightweight image super-resolution reconstruction method based on cascade network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202110052039.8A CN112734643A (en) 2021-01-15 2021-01-15 Lightweight image super-resolution reconstruction method based on cascade network

Publications (1)

Publication Number Publication Date
CN112734643A true CN112734643A (en) 2021-04-30

Family

ID=75593193

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202110052039.8A Pending CN112734643A (en) 2021-01-15 2021-01-15 Lightweight image super-resolution reconstruction method based on cascade network

Country Status (1)

Country Link
CN (1) CN112734643A (en)

Cited By (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112966661A (en) * 2021-03-31 2021-06-15 东南大学 Construction method of face feature extraction network based on sparse feature reuse
CN113409191A (en) * 2021-06-02 2021-09-17 广东工业大学 Lightweight image super-resolution method and system based on attention feedback mechanism
CN114494022A (en) * 2022-03-31 2022-05-13 苏州浪潮智能科技有限公司 Model training method, super-resolution reconstruction method, device, equipment and medium
CN115601242A (en) * 2022-12-13 2023-01-13 电子科技大学(Cn) Lightweight image super-resolution reconstruction method suitable for hardware deployment

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110276721A (en) * 2019-04-28 2019-09-24 天津大学 Image super-resolution rebuilding method based on cascade residual error convolutional neural networks
US20200065940A1 (en) * 2018-08-23 2020-02-27 General Electric Company Patient-specific deep learning image denoising methods and systems
CN110909783A (en) * 2019-11-15 2020-03-24 重庆邮电大学 Blind domain image classification and reconstruction method based on enhanced reconstruction classification network
CN111192200A (en) * 2020-01-02 2020-05-22 南京邮电大学 Image super-resolution reconstruction method based on fusion attention mechanism residual error network
CN111489291A (en) * 2020-03-04 2020-08-04 浙江工业大学 Medical image super-resolution reconstruction method based on network cascade
CN111583109A (en) * 2020-04-23 2020-08-25 华南理工大学 Image super-resolution method based on generation countermeasure network
CN111754403A (en) * 2020-06-15 2020-10-09 南京邮电大学 Image super-resolution reconstruction method based on residual learning
CN111784582A (en) * 2020-07-08 2020-10-16 桂林电子科技大学 DEC-SE-based low-illumination image super-resolution reconstruction method
CN112203098A (en) * 2020-09-22 2021-01-08 广东启迪图卫科技股份有限公司 Mobile terminal image compression method based on edge feature fusion and super-resolution

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20200065940A1 (en) * 2018-08-23 2020-02-27 General Electric Company Patient-specific deep learning image denoising methods and systems
CN110276721A (en) * 2019-04-28 2019-09-24 天津大学 Image super-resolution rebuilding method based on cascade residual error convolutional neural networks
CN110909783A (en) * 2019-11-15 2020-03-24 重庆邮电大学 Blind domain image classification and reconstruction method based on enhanced reconstruction classification network
CN111192200A (en) * 2020-01-02 2020-05-22 南京邮电大学 Image super-resolution reconstruction method based on fusion attention mechanism residual error network
CN111489291A (en) * 2020-03-04 2020-08-04 浙江工业大学 Medical image super-resolution reconstruction method based on network cascade
CN111583109A (en) * 2020-04-23 2020-08-25 华南理工大学 Image super-resolution method based on generation countermeasure network
CN111754403A (en) * 2020-06-15 2020-10-09 南京邮电大学 Image super-resolution reconstruction method based on residual learning
CN111784582A (en) * 2020-07-08 2020-10-16 桂林电子科技大学 DEC-SE-based low-illumination image super-resolution reconstruction method
CN112203098A (en) * 2020-09-22 2021-01-08 广东启迪图卫科技股份有限公司 Mobile terminal image compression method based on edge feature fusion and super-resolution

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
JIAO XIE 等: "Training convolutional neural networks with cheap convolutions and online distillation", 《ARXIV PREPRINT ARXIV》 *

Cited By (8)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112966661A (en) * 2021-03-31 2021-06-15 东南大学 Construction method of face feature extraction network based on sparse feature reuse
CN113409191A (en) * 2021-06-02 2021-09-17 广东工业大学 Lightweight image super-resolution method and system based on attention feedback mechanism
CN113409191B (en) * 2021-06-02 2023-04-07 广东工业大学 Lightweight image super-resolution method and system based on attention feedback mechanism
CN114494022A (en) * 2022-03-31 2022-05-13 苏州浪潮智能科技有限公司 Model training method, super-resolution reconstruction method, device, equipment and medium
CN114494022B (en) * 2022-03-31 2022-07-29 苏州浪潮智能科技有限公司 Model training method, super-resolution reconstruction method, device, equipment and medium
WO2023184913A1 (en) * 2022-03-31 2023-10-05 苏州浪潮智能科技有限公司 Model training method and apparatus, super-resolution reconstruction method, device and medium
CN115601242A (en) * 2022-12-13 2023-01-13 电子科技大学(Cn) Lightweight image super-resolution reconstruction method suitable for hardware deployment
CN115601242B (en) * 2022-12-13 2023-04-18 电子科技大学 Lightweight image super-resolution reconstruction method suitable for hardware deployment

Similar Documents

Publication Publication Date Title
CN112734643A (en) Lightweight image super-resolution reconstruction method based on cascade network
CN110020989B (en) Depth image super-resolution reconstruction method based on deep learning
CN112750082B (en) Human face super-resolution method and system based on fusion attention mechanism
WO2021018163A1 (en) Neural network search method and apparatus
Wang et al. A review of image super-resolution approaches based on deep learning and applications in remote sensing
CN109064396A (en) A kind of single image super resolution ratio reconstruction method based on depth ingredient learning network
CN114972746B (en) Medical image segmentation method based on multi-resolution overlapping attention mechanism
CN110136067B (en) Real-time image generation method for super-resolution B-mode ultrasound image
Zheng et al. S-Net: a scalable convolutional neural network for JPEG compression artifact reduction
Gendy et al. Lightweight image super-resolution based on deep learning: State-of-the-art and future directions
CN113920043A (en) Double-current remote sensing image fusion method based on residual channel attention mechanism
Zhang et al. Deformable and residual convolutional network for image super-resolution
CN112365405A (en) Unsupervised super-resolution reconstruction method based on generation countermeasure network
CN116486074A (en) Medical image segmentation method based on local and global context information coding
CN114841859A (en) Single-image super-resolution reconstruction method based on lightweight neural network and Transformer
CN116168197A (en) Image segmentation method based on Transformer segmentation network and regularization training
Gendy et al. Balanced spatial feature distillation and pyramid attention network for lightweight image super-resolution
CN113627487B (en) Super-resolution reconstruction method based on deep attention mechanism
CN114359039A (en) Knowledge distillation-based image super-resolution method
CN110751271A (en) Image traceability feature characterization method based on deep neural network
CN116188272B (en) Two-stage depth network image super-resolution reconstruction method suitable for multiple fuzzy cores
CN113096015A (en) Image super-resolution reconstruction method based on progressive sensing and ultra-lightweight network
CN116597146A (en) Semantic segmentation method for laser radar sparse point cloud data
CN116128722A (en) Image super-resolution reconstruction method and system based on frequency domain-texture feature fusion
CN116485654A (en) Lightweight single-image super-resolution reconstruction method combining convolutional neural network and transducer

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
WD01 Invention patent application deemed withdrawn after publication

Application publication date: 20210430

WD01 Invention patent application deemed withdrawn after publication