CN112734643A - Lightweight image super-resolution reconstruction method based on cascade network - Google Patents
Lightweight image super-resolution reconstruction method based on cascade network Download PDFInfo
- Publication number
- CN112734643A CN112734643A CN202110052039.8A CN202110052039A CN112734643A CN 112734643 A CN112734643 A CN 112734643A CN 202110052039 A CN202110052039 A CN 202110052039A CN 112734643 A CN112734643 A CN 112734643A
- Authority
- CN
- China
- Prior art keywords
- network
- image
- resolution
- channels
- convolution
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Pending
Links
- 238000000034 method Methods 0.000 title claims abstract description 24
- 238000012549 training Methods 0.000 claims abstract description 11
- 230000007246 mechanism Effects 0.000 claims abstract description 7
- 238000000605 extraction Methods 0.000 claims description 21
- 230000003321 amplification Effects 0.000 claims description 6
- 238000003199 nucleic acid amplification method Methods 0.000 claims description 6
- 238000013461 design Methods 0.000 claims description 3
- 230000003993 interaction Effects 0.000 claims description 3
- ORILYTVJVMAKLC-UHFFFAOYSA-N Adamantane Natural products C1C(C2)CC3CC1CC2C3 ORILYTVJVMAKLC-UHFFFAOYSA-N 0.000 claims description 2
- 238000005215 recombination Methods 0.000 claims description 2
- 230000006798 recombination Effects 0.000 claims description 2
- 238000005070 sampling Methods 0.000 claims description 2
- 238000010586 diagram Methods 0.000 claims 1
- 238000012545 processing Methods 0.000 abstract description 2
- 238000005457 optimization Methods 0.000 abstract 1
- 108091006146 Channels Proteins 0.000 description 25
- 101100365548 Caenorhabditis elegans set-14 gene Proteins 0.000 description 3
- 230000000694 effects Effects 0.000 description 3
- 230000009467 reduction Effects 0.000 description 3
- 238000004364 calculation method Methods 0.000 description 2
- 230000006870 function Effects 0.000 description 2
- 230000001133 acceleration Effects 0.000 description 1
- 230000004913 activation Effects 0.000 description 1
- 238000013135 deep learning Methods 0.000 description 1
- 238000011161 development Methods 0.000 description 1
- 239000003814 drug Substances 0.000 description 1
- 238000005516 engineering process Methods 0.000 description 1
- 238000003384 imaging method Methods 0.000 description 1
- 230000006872 improvement Effects 0.000 description 1
- 238000011176 pooling Methods 0.000 description 1
- 238000007781 pre-processing Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000011160 research Methods 0.000 description 1
- 239000000126 substance Substances 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T3/00—Geometric image transformations in the plane of the image
- G06T3/40—Scaling of whole images or parts thereof, e.g. expanding or contracting
- G06T3/4053—Scaling of whole images or parts thereof, e.g. expanding or contracting based on super-resolution, i.e. the output image resolution being higher than the sensor resolution
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T5/00—Image enhancement or restoration
- G06T5/50—Image enhancement or restoration using two or more images, e.g. averaging or subtraction
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20081—Training; Learning
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06T—IMAGE DATA PROCESSING OR GENERATION, IN GENERAL
- G06T2207/00—Indexing scheme for image analysis or image enhancement
- G06T2207/20—Special algorithmic details
- G06T2207/20112—Image segmentation details
- G06T2207/20132—Image cropping
Landscapes
- Physics & Mathematics (AREA)
- General Physics & Mathematics (AREA)
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Image Analysis (AREA)
- Image Processing (AREA)
Abstract
The invention belongs to the technical field of image processing, and mainly relates to a light-weight image super-resolution reconstruction method based on a cascade network. The method mainly comprises the following steps: step 1: acquiring a high-resolution target image of a data set, and performing downsampling pretreatment; step 2: designing a light-weight image super-resolution reconstruction network based on an attention mechanism and a cascade network; and step 3: and designing a training strategy and training parameters and optimizing model parameters. After training and optimization, the invention can provide an image super-resolution reconstruction method with small volume and high reconstruction quality.
Description
The technical field is as follows:
the invention belongs to the technical field of image processing, and mainly relates to a light-weight image super-resolution reconstruction method based on a cascade network.
Background art:
generally, for an image, a higher resolution represents more and richer image details. But is limited by the factors of the size, the cost and the like of the imaging equipment, and the finally presented image often does not meet the actual requirements of people. Under such a background, an image super-resolution reconstruction technique for reconstructing a high-resolution image from a low-resolution image has attracted extensive attention in academic circles, and is widely used in the fields of video surveillance, medicine, and the like. Therefore, the super-resolution reconstruction algorithm with wide research application range and high reconstruction quality has important theoretical value and practical significance.
Most super-resolution reconstruction methods improve the reconstruction effect by continuously stacking the depth of the network, and although the algorithms can often obtain better effect, the network has larger volume and large calculation amount, so that the super-resolution reconstruction methods are difficult to apply to practical application. The conventional convolution is generally used for feature extraction, and with the development of a network model, technologies such as deep separable convolution and packet convolution have achieved prominent achievements in the field of deep learning. The existing light weight network in the field of image super-resolution reconstruction mainly solves the problems from several angles: an effective up-sampling mode and a feature extraction module are designed, and the depth of the network is reduced. Although a relatively good effect is obtained in the case of a relatively light weight, there is still room for further improvement. A lightweight image super-resolution reconstruction based on a cascade network is designed, firstly a lightweight multi-scale feature extraction module is designed, and then different features are weighted by using an attention mechanism. And finally, fusing the channel characteristics of different levels by using a cascading mechanism. The super-resolution reconstruction network designed based on the scheme can meet the requirement of real-time performance, and meanwhile, the quality of the super-resolution reconstruction network is superior to that of the super-resolution reconstruction network with the same level of network complexity.
Disclosure of Invention
The invention provides a light-weight image super-resolution reconstruction method based on a cascade network, which aims at the problems that the existing super-resolution reconstruction method is large in size and difficult to apply in practice, provides an unsupervised super-resolution reconstruction method, and optimizes a reconstruction result through a pyramid generation countermeasure network, so that a super-resolution reconstruction network conforming to image characteristics is obtained.
The technical scheme of the invention is as follows:
s1: and (6) image acquisition. The DIV2K data set is obtained and the PNG data is converted into the npy data format.
S2: on the basis of S1, the. npy data is cropped to crop a high resolution image HR of 128 × 128, and then down-sampled to obtain a low resolution image LR, forming a high and low resolution image pair.
S3: on the basis of S2, using convolution design low-level feature extraction network to extract features, wherein the extracted features are represented as FL。
FL=Conv2d3x3(LR,3,64)
Where Conv2d3x3 represents a conventional convolution with a convolution kernel size of 3, in this way shallow features of the network are extracted. Where 3 and 64 represent the number of input and output channels, respectively.
S4: the feature F obtained in S3LAs an input of the depth feature extraction network, a channel 1 performs feature extraction through an FE module as shown in fig. 1, and finally performs channel dimensionality reduction using a standard convolution with a convolution kernel size of 1:
FFE1=Conv1x1(σ(GConv4(DWConv3x3(F,3,64),64,64)),64,32)
where DWConv3x3 represents a deep convolution with a convolution kernel size of 3, GConv4 represents a packet convolution with a packet number of 4, and σ represents the ReLU activation function. Conv1x1 represents a conventional convolution with a convolution kernel size of 1.
And the channel 2 firstly carries out feature extraction in a feature extraction mode of the channel 1, then carries out channel recombination, and then carries out secondary feature extraction through an FE module. The dimensionality reduction of the channels is then performed using the same 1x1 convolution.
FFE2=Conv1x1(σ(GConv4(DWConv3x3(Shufle(σ(GConv4(DWConv3x3(FL,3,64),64,64))),3,64),64,64)),64,32)
The characteristics of the outputs of channel 1 and channel 2 are then tiled to extend to a 64-channel profile and activated using ReLU, adding to the non-linearity of the network. Is shown as
FDF1=σ(Concat(FFE1+FFE2))
In DGSA, an enhanced channel attention mechanism network as shown in fig. 2 is introduced. First, the channel descriptor is validated. Let F (F ∈ i)W×H×CW, H, C is the width, height and channel number of the feature map) to obtain a1 × 1 × C feature map, and the result of pooling is used as the descriptor Z of the channel, and the descriptor of the C-th channel is expressed as:
subsequently, channel relationship modeling, with WkTo learn channel attention, expressed as follows:
Wc=σ(WkZ)
where k is expressed as the number of calculated adjacent channels, WkExpressed as:
Wk=[ω1,ω2,L,ωc]T
wherein the content of the first and second substances,representing k adjacent channelsA collection of (a). Channel parameter sharing can further reduce parameter quantity, omegaiCan be expressed as:
finally, the information interaction between the channels is realized through the one-dimensional convolution with the convolution kernel size k, which is expressed as:
ω=σ(Conv1dkxk(Z))
wherein k is adaptively selected according to the number of channels, and is represented as:
C=φ(k)=22k-1
weighting the learned channel attention mechanism into the input features to obtain a final feature map:
Fout=Z*F
the final DGSA module may be represented as shown in fig. 3. And simultaneously, cascading the features of different layers, and sending the features into a conventional convolution with a convolution kernel size of 1 to perform feature information interaction and channel dimension reduction.
Fcon=Conv1x1(Concat(F1,F2,...),inputchannel,64)
In the formula, inputchanel represents the sum of the channels of each input feature map.
S5: feature F after feature extraction in S4conThe features are fed into an upsampling module as shown in fig. 4. The upsampling module first needs to expand the channel.
Fup=Conv3x3(Fcon,64,64*r*r)
Fup_f=PS(Fup,64*r*r,64)
Where r represents the multiple to be reconstructed and PS represents the Pixel Shuffle operation. If the amplification is four times, the above operation is performed twice, where r is 2.
S6: using the L1 loss as a loss function for network training, an Adam optimizer as an optimizer for the network, the error between SR and HR is calculated for back propagation, and 100 epochs are trained. And updating the parameters.
S7: after training is finished, parameters of the model are fixed, the image to be reconstructed is used as the input of the network, the reconstruction scale is selected, and the output is the super-resolution image of the target.
Drawings
FIG. 1 FE module
FIG. 2 attention mechanism network
FIG. 3 network model
FIG. 4 upsampling module
FIG. 5 comparison of reconstruction results
The specific implementation process comprises the following steps:
a lightweight image super-resolution reconstruction method based on a cascade network is characterized in that a network model is shown in figure 3, data preprocessing is firstly carried out, and then a super-resolution reconstruction network is constructed. And then setting training parameters and conditions to train the network and adjusting the network parameters. And finally, performing super-resolution reconstruction by using the trained model.
The invention is further illustrated by the following example of an embodiment, which is intended only for a better understanding of the subject matter of the invention and is not intended to limit the scope of the invention. The method comprises the following specific steps:
step S1: the example adopts a Set5, a Set14, a BSD100 and an Urban100 four-image super-resolution reconstruction data Set, the Set5 comprises 5 pairs of high-low resolution image pairs, the Set14 extends to 14 sheets, and the BSD100 and the Urban100 respectively comprise 100 sheets of high-low resolution image pairs with various features. The example selects quadrupling as the reference standard. And carrying out downsampling on the low-resolution image to obtain an ultra-low-resolution image pair.
Step S2: the present embodiment takes the low resolution image directly as the input to the network.
Step S3: and confirming the training environment and the strategy. The processor of the experimental device isI99900K, wherein the display card is RTX2080Ti, the training is carried out in a Pythrch environment, and the acceleration model is trained by using CUDA10.0 and CuDNN7.1.
The epoch of the network trained in each scale is 100, the learning rate of a generator and a discriminator is set to be 0.001, the learning rate of the epoch is reduced to 10 percent of the original learning rate every 30, and the scale of the feature image is kept unchanged in the feature extraction process.
Step S4: the present example selects the low resolution images in Set5, Set14, and BSD100 as the reconstruction objects. And taking the low-resolution image to be reconstructed as an input, and directly inputting the low-resolution image into a network for calculation.
The example compares the algorithm proposed herein with the existing unsupervised image super-resolution reconstruction algorithm for 2, 3 and 4 times of reconstruction indexes, and mainly compares two indexes of peak signal-to-noise ratio (PSNR) and Structural Similarity (SSIM). The experimental results are shown in fig. 5 and table 1, taking 4-fold reconstruction as an example.
As can be seen from the observation of the graph, the method is superior to a comparison algorithm in terms of two indexes, and the quality of the reconstructed image is improved to a certain extent.
TABLE 1 reconstruction index comparison Table
Claims (8)
1. A light-weight image super-resolution reconstruction method based on a cascade network is characterized by comprising the following steps:
s1: the DIV2K data set is obtained and the PNG data is converted into the npy data format.
S2: on the basis of S1, the. npy data is cropped to crop a high resolution image HR of 128 × 128, and then down-sampled to obtain a low resolution image LR, forming a high and low resolution image pair.
S3: on the basis of S2, using convolution design low-level feature extraction network to extract features, wherein the extracted features are represented as FL。
S4: on the basis of S3, a depth feature extraction module DGSA is designed, wherein the DGSA comprises a light-weight volume part and an attention mechanism part. And 4 DGSA cascades form a deep feature extraction network. Extracting a depth feature F using the low-level feature extracted in S3 as an input to a depth feature extraction networkH。
S5: on the basis of S4, the extracted depth feature F is usedHThe input upsampling network enlarges the feature map to the size of the target reconstruction. And then, restoring the same channel number of the target image through a reconstruction module, and outputting a target SR.
S6: the error between the HR obtained at S1 and the SR obtained at S5 was calculated using the L1 loss function. And performing gradient negative feedback to update the parameters of the feature extraction layer, the up-sampling layer and the reconstruction layer.
S7: and (5) carrying out multiple iterations on the S2-S6, continuously updating network parameters, and fixing the model after the set training amount is reached. And reconstructing the image to be reconstructed as the input of the model to obtain the super-resolution image SR of the targetT。
2. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S1:
the image file of the originally obtained PNG is converted into the npy data, and the file reading speed is accelerated.
3. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S2:
the resulting. npy file is cropped to a crop size of 128 x 128 resolution and downsampled for different reconstruction scales.
4. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S3:
and designing a low-level feature extraction network, wherein the low-level feature extraction network is formed by convolution with a convolution kernel size of 3, the number of input channels is 3, and the number of output channels is 64. The size of the characteristic diagram is unchanged.
5. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S4:
design depth feature extraction module DGSA (Depthwise-Group ShuffleNet with Attention). The DGSA is formed by two channels, namely a main channel 1 and an output channel after being directly subjected to feature extraction by an FE module. The channel 2 firstly carries out first feature extraction through an FE module, then carries out feature channel recombination, changes the position of the channel and enhances the information interaction between the channels. The FE module 64 consists of deep convolution of channels and packet convolution, and the number of packets is 4. And activated using the ReLU function. After the characteristics of the two channels are obtained, convolution with convolution kernel of 1 multiplied by 1 is carried out to carry out residual error connection on the information of the two channels, the information of the channels is fused, and finally, the characteristic F is outputH1=DGSA(FL). Then fusing low-level features and depth features, splicing channels, fusing the channels by using a standard convolution with a convolution kernel of 1 multiplied by 1, and finally outputting FH。
6. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S5:
the channel expansion is carried out by convolution with convolution kernel of 3 × 3, then Pixel Shuffle is carried out for feature map amplification, and the number of channels is ensured to be 64, if the amplification factor is 2 and 3, the amplification is carried out only once, and if the amplification factor is 4, the above operation is carried out twice, and the amplification is respectively doubled.
7. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S6:
using the L1 loss as a loss function for network training, an Adam optimizer as an optimizer for the network, the error between SR and HR is calculated for back propagation, and 100 epochs are trained. And updating the parameters.
8. A lightweight image super-resolution reconstruction method based on a cascade network is characterized in that in step S7:
after training is finished, parameters of the model are fixed, the image to be reconstructed is used as the input of the network, and the output is the super-resolution image of the target.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110052039.8A CN112734643A (en) | 2021-01-15 | 2021-01-15 | Lightweight image super-resolution reconstruction method based on cascade network |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110052039.8A CN112734643A (en) | 2021-01-15 | 2021-01-15 | Lightweight image super-resolution reconstruction method based on cascade network |
Publications (1)
Publication Number | Publication Date |
---|---|
CN112734643A true CN112734643A (en) | 2021-04-30 |
Family
ID=75593193
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110052039.8A Pending CN112734643A (en) | 2021-01-15 | 2021-01-15 | Lightweight image super-resolution reconstruction method based on cascade network |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112734643A (en) |
Cited By (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112966661A (en) * | 2021-03-31 | 2021-06-15 | 东南大学 | Construction method of face feature extraction network based on sparse feature reuse |
CN113409191A (en) * | 2021-06-02 | 2021-09-17 | 广东工业大学 | Lightweight image super-resolution method and system based on attention feedback mechanism |
CN114494022A (en) * | 2022-03-31 | 2022-05-13 | 苏州浪潮智能科技有限公司 | Model training method, super-resolution reconstruction method, device, equipment and medium |
CN115601242A (en) * | 2022-12-13 | 2023-01-13 | 电子科技大学(Cn) | Lightweight image super-resolution reconstruction method suitable for hardware deployment |
Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN110276721A (en) * | 2019-04-28 | 2019-09-24 | 天津大学 | Image super-resolution rebuilding method based on cascade residual error convolutional neural networks |
US20200065940A1 (en) * | 2018-08-23 | 2020-02-27 | General Electric Company | Patient-specific deep learning image denoising methods and systems |
CN110909783A (en) * | 2019-11-15 | 2020-03-24 | 重庆邮电大学 | Blind domain image classification and reconstruction method based on enhanced reconstruction classification network |
CN111192200A (en) * | 2020-01-02 | 2020-05-22 | 南京邮电大学 | Image super-resolution reconstruction method based on fusion attention mechanism residual error network |
CN111489291A (en) * | 2020-03-04 | 2020-08-04 | 浙江工业大学 | Medical image super-resolution reconstruction method based on network cascade |
CN111583109A (en) * | 2020-04-23 | 2020-08-25 | 华南理工大学 | Image super-resolution method based on generation countermeasure network |
CN111754403A (en) * | 2020-06-15 | 2020-10-09 | 南京邮电大学 | Image super-resolution reconstruction method based on residual learning |
CN111784582A (en) * | 2020-07-08 | 2020-10-16 | 桂林电子科技大学 | DEC-SE-based low-illumination image super-resolution reconstruction method |
CN112203098A (en) * | 2020-09-22 | 2021-01-08 | 广东启迪图卫科技股份有限公司 | Mobile terminal image compression method based on edge feature fusion and super-resolution |
-
2021
- 2021-01-15 CN CN202110052039.8A patent/CN112734643A/en active Pending
Patent Citations (9)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
US20200065940A1 (en) * | 2018-08-23 | 2020-02-27 | General Electric Company | Patient-specific deep learning image denoising methods and systems |
CN110276721A (en) * | 2019-04-28 | 2019-09-24 | 天津大学 | Image super-resolution rebuilding method based on cascade residual error convolutional neural networks |
CN110909783A (en) * | 2019-11-15 | 2020-03-24 | 重庆邮电大学 | Blind domain image classification and reconstruction method based on enhanced reconstruction classification network |
CN111192200A (en) * | 2020-01-02 | 2020-05-22 | 南京邮电大学 | Image super-resolution reconstruction method based on fusion attention mechanism residual error network |
CN111489291A (en) * | 2020-03-04 | 2020-08-04 | 浙江工业大学 | Medical image super-resolution reconstruction method based on network cascade |
CN111583109A (en) * | 2020-04-23 | 2020-08-25 | 华南理工大学 | Image super-resolution method based on generation countermeasure network |
CN111754403A (en) * | 2020-06-15 | 2020-10-09 | 南京邮电大学 | Image super-resolution reconstruction method based on residual learning |
CN111784582A (en) * | 2020-07-08 | 2020-10-16 | 桂林电子科技大学 | DEC-SE-based low-illumination image super-resolution reconstruction method |
CN112203098A (en) * | 2020-09-22 | 2021-01-08 | 广东启迪图卫科技股份有限公司 | Mobile terminal image compression method based on edge feature fusion and super-resolution |
Non-Patent Citations (1)
Title |
---|
JIAO XIE 等: "Training convolutional neural networks with cheap convolutions and online distillation", 《ARXIV PREPRINT ARXIV》 * |
Cited By (8)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN112966661A (en) * | 2021-03-31 | 2021-06-15 | 东南大学 | Construction method of face feature extraction network based on sparse feature reuse |
CN113409191A (en) * | 2021-06-02 | 2021-09-17 | 广东工业大学 | Lightweight image super-resolution method and system based on attention feedback mechanism |
CN113409191B (en) * | 2021-06-02 | 2023-04-07 | 广东工业大学 | Lightweight image super-resolution method and system based on attention feedback mechanism |
CN114494022A (en) * | 2022-03-31 | 2022-05-13 | 苏州浪潮智能科技有限公司 | Model training method, super-resolution reconstruction method, device, equipment and medium |
CN114494022B (en) * | 2022-03-31 | 2022-07-29 | 苏州浪潮智能科技有限公司 | Model training method, super-resolution reconstruction method, device, equipment and medium |
WO2023184913A1 (en) * | 2022-03-31 | 2023-10-05 | 苏州浪潮智能科技有限公司 | Model training method and apparatus, super-resolution reconstruction method, device and medium |
CN115601242A (en) * | 2022-12-13 | 2023-01-13 | 电子科技大学(Cn) | Lightweight image super-resolution reconstruction method suitable for hardware deployment |
CN115601242B (en) * | 2022-12-13 | 2023-04-18 | 电子科技大学 | Lightweight image super-resolution reconstruction method suitable for hardware deployment |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN112734643A (en) | Lightweight image super-resolution reconstruction method based on cascade network | |
CN110020989B (en) | Depth image super-resolution reconstruction method based on deep learning | |
CN112750082B (en) | Human face super-resolution method and system based on fusion attention mechanism | |
WO2021018163A1 (en) | Neural network search method and apparatus | |
Wang et al. | A review of image super-resolution approaches based on deep learning and applications in remote sensing | |
CN109064396A (en) | A kind of single image super resolution ratio reconstruction method based on depth ingredient learning network | |
CN114972746B (en) | Medical image segmentation method based on multi-resolution overlapping attention mechanism | |
CN110136067B (en) | Real-time image generation method for super-resolution B-mode ultrasound image | |
Zheng et al. | S-Net: a scalable convolutional neural network for JPEG compression artifact reduction | |
Gendy et al. | Lightweight image super-resolution based on deep learning: State-of-the-art and future directions | |
CN113920043A (en) | Double-current remote sensing image fusion method based on residual channel attention mechanism | |
Zhang et al. | Deformable and residual convolutional network for image super-resolution | |
CN112365405A (en) | Unsupervised super-resolution reconstruction method based on generation countermeasure network | |
CN116486074A (en) | Medical image segmentation method based on local and global context information coding | |
CN114841859A (en) | Single-image super-resolution reconstruction method based on lightweight neural network and Transformer | |
CN116168197A (en) | Image segmentation method based on Transformer segmentation network and regularization training | |
Gendy et al. | Balanced spatial feature distillation and pyramid attention network for lightweight image super-resolution | |
CN113627487B (en) | Super-resolution reconstruction method based on deep attention mechanism | |
CN114359039A (en) | Knowledge distillation-based image super-resolution method | |
CN110751271A (en) | Image traceability feature characterization method based on deep neural network | |
CN116188272B (en) | Two-stage depth network image super-resolution reconstruction method suitable for multiple fuzzy cores | |
CN113096015A (en) | Image super-resolution reconstruction method based on progressive sensing and ultra-lightweight network | |
CN116597146A (en) | Semantic segmentation method for laser radar sparse point cloud data | |
CN116128722A (en) | Image super-resolution reconstruction method and system based on frequency domain-texture feature fusion | |
CN116485654A (en) | Lightweight single-image super-resolution reconstruction method combining convolutional neural network and transducer |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
WD01 | Invention patent application deemed withdrawn after publication |
Application publication date: 20210430 |
|
WD01 | Invention patent application deemed withdrawn after publication |