CN113902658B - RGB image-to-hyperspectral image reconstruction method based on dense multiscale network - Google Patents

RGB image-to-hyperspectral image reconstruction method based on dense multiscale network Download PDF

Info

Publication number
CN113902658B
CN113902658B CN202111020790.6A CN202111020790A CN113902658B CN 113902658 B CN113902658 B CN 113902658B CN 202111020790 A CN202111020790 A CN 202111020790A CN 113902658 B CN113902658 B CN 113902658B
Authority
CN
China
Prior art keywords
multiplied
feature
image
network
feature map
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111020790.6A
Other languages
Chinese (zh)
Other versions
CN113902658A (en
Inventor
周慧鑫
李怡雨
宋江鲁奇
李宇燕
张嘉嘉
向培
滕翔
王瑛琨
李苗青
田成
王财顺
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Xidian University
Original Assignee
Xidian University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Xidian University filed Critical Xidian University
Priority to CN202111020790.6A priority Critical patent/CN113902658B/en
Publication of CN113902658A publication Critical patent/CN113902658A/en
Application granted granted Critical
Publication of CN113902658B publication Critical patent/CN113902658B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/50Image enhancement or restoration by the use of more than one image, e.g. averaging, subtraction
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T11/002D [Two Dimensional] image generation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • G06T2207/10036Multispectral image; Hyperspectral image
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20212Image combination
    • G06T2207/20221Image fusion; Image merging

Abstract

The invention discloses a reconstruction method of RGB image to hyperspectral image based on dense multiscale network, which constructs an improved residual error network model; performing feature extraction on an input image through a 1 × 1 convolutional layer to obtain a feature map; in the feature mapping part of the residual error network model, a conversion layer and two branches of a main network are adopted to respectively process a feature map; the conversion layer is characterized in that two groups of 3 x 3 convolutions are added on the basis of quick connection to directly extract the characteristics of the characteristic graph to obtain a first output characteristic graph; a cross-channel fusion receptive field module is added behind the last residual block of the main network, and feature extraction is carried out on the feature map to obtain a second output feature map; and adding the first output characteristic diagram and the second output characteristic diagram to obtain a reconstructed hyperspectral image. The method realizes the reconstruction from the RGB image to the hyperspectral image, has better reconstruction effect than the traditional algorithm, and can effectively enlarge the receptive field of the network, thereby obtaining good reconstruction effect.

Description

Method for reconstructing RGB image to hyperspectral image based on dense multiscale network
Technical Field
The invention belongs to the technical field of image processing, and particularly relates to a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network.
Background
The hyperspectral image contains spatial information and spectral information at the same time, so that not only can relevant operation be performed on the two-dimensional scene image, but also the material can be analyzed according to the spectral response curve, and the hyperspectral image has high application value. In order to reduce the difficulty in acquiring high-spatial-resolution hyperspectral images, reconstructing hyperspectral images by using low-cost RGB images has become an important research content in recent years.
The method for reconstructing the hyperspectral image by the RGB image can be mainly divided into a method for reconstructing by additional equipment and a method for reconstructing by directly using an algorithm. The method for reconstructing the RGB image to the hyperspectral image based on the equipment mainly comprises the steps of modifying an RGB camera system, adopting a plurality of cameras or a preset imaging environment and the like, and increasing spectral dimension information obtained from the RGB cameras so as to reduce reconstruction difficulty. But also because of the dependence on equipment and environment, the universality of the method is reduced, and meanwhile, a complex optical path system needs professional knowledge to construct and use, so that the use difficulty is increased. The method for reconstructing the image based on the algorithm mainly comprises a traditional algorithm and an algorithm based on deep learning. The hyperspectral image reconstruction network based on deep learning mainly comprises three parts of feature extraction, feature mapping and spectrum reconstruction.
In the algorithm for reconstructing the hyperspectral image from the RGB image based on deep learning, effective feature extraction is an important premise for reconstructing the hyperspectral image more accurately, and the receptive field is one of important factors influencing feature extraction. Therefore, the deep learning network architecture for researching and expanding the receptive field and effectively extracting the context information has important significance for reconstructing the RGB to hyperspectral images.
Disclosure of Invention
In view of this, the present invention provides a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network.
In order to achieve the purpose, the technical scheme of the invention is realized as follows:
the embodiment of the invention provides a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network, which comprises the following steps:
constructing an improved residual error network model;
performing feature extraction on an input image through a 1 × 1 convolutional layer to obtain a feature map;
in the feature mapping part of the residual error network model, a conversion layer and a main network branch are adopted to respectively process a feature map;
the conversion layer is characterized in that two groups of 3 x 3 convolutions are added on the basis of quick connection to directly extract the characteristics of the characteristic graph to obtain a first output characteristic graph;
a cross-channel fusion receptive field module is added behind the last residual block of the main network, and feature extraction is carried out on the feature map to obtain a second output feature map;
and adding the first output characteristic diagram and the second output characteristic diagram to obtain a reconstructed hyperspectral image.
In the scheme, the constructed improved residual error network model is based on a reconstruction algorithm based on a residual error block, and a traditional convolution layer is used as a conversion layer to extract basic characteristics; the main network is composed of a traditional convolution layer, a 1 x 1 convolution and a plurality of groups of residual blocks, wherein the 1 x 1 convolution is used for adjusting the number of channels.
In the above scheme, the residual block includes a main network and two branches for quick connection, the main network generates a feature matrix through a plurality of convolution layers, and the feature matrix is activated after being added with the feature matrix for quick connection.
In the above scheme, the main network adds a cross-channel fusion receptive field module behind the last residual block, which specifically comprises:
(1) The pooling part respectively adopts 2 multiplied by 2, 3 multiplied by 3 and 6 multiplied by 6 pooling windows to carry out down-sampling operation on the input feature map;
(2) The preprocessing part adopts a self-normalization activation function Selu, continuously down-samples feature graphs subjected to pooling operation by adopting 2 x 2 and 3 x 3 after self-normalization and channel quantity conversion until the feature graphs are consistent with the feature graphs of the 6 x 6 branches in size, and performs cross-layer splicing on the two feature graphs and the processed feature graphs on the 6 x 6 branches;
(3) An up-sampling part adopts pixel buffer to perform up-sampling treatment;
(4) And the characteristic splicing part splices the channel dimensions of the four groups of obtained characteristic graphs.
In the foregoing solution, the normalized activation function Selu specifically includes:
Figure BDA0003241873790000031
wherein the content of the first and second substances,
Figure BDA0003241873790000033
in the above scheme, the upsampling part performs upsampling processing by using pixel buffer, and specifically includes: the expression for performing the upsampling operation by using pixel shuffle is
I SR =f L (I LR )=PS(W L *f L-1 (I LR )+b L )
In which I SR Is a high resolution feature map, I LR Is a low resolution feature map, f L For non-linear operation on the L-th layer low resolution feature map, weight W L Is n l-1 ×r 2 C×k L ×k L ,b L Is an offset. PS is a periodic screening operation that combines and transforms spatial and channel information for upsampling purposes,it is calculated in the manner of
Figure BDA0003241873790000032
Wherein, T is a low-resolution image, r is an image expansion multiple, x and y are coordinates of pixel points on a high-resolution characteristic diagram, H multiplied by W multiplied by C multiplied by r 2 Is converted into a tensor of rH × rW × C.
Compared with the prior art, the method realizes the reconstruction from the RGB image to the hyperspectral image, and the reconstruction effect is superior to that of the traditional algorithm; the cross-channel fusion receptive field module is constructed by using the multi-scale pyramid module, the intensive connecting blocks, the pixel recombination module and the cross-channel connection, the main network part in the reconstruction algorithm based on the residual block is improved by using the cross-channel fusion receptive field module, a new side connection is constructed, multi-scale information is fused on the basis of deepening the network depth, the receptive field of the network can be effectively expanded, and a good reconstruction effect is achieved.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the invention and do not limit the invention. In the drawings:
FIG. 1 is a flowchart of an algorithm for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network according to an embodiment of the present invention;
FIG. 2 is a network structure diagram of a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network according to an embodiment of the present invention;
FIG. 3 is a cross-channel fusion receptive field module for a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network according to an embodiment of the present invention;
fig. 4 is an effect diagram of reconstructing a hyperspectral image from an RGB image to a hyperspectral image in a method for reconstructing an RGB image based on a dense multiscale network according to an embodiment of the present invention.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, the present invention is further described in detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the invention and do not limit the invention.
The embodiment of the invention provides a method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network, which comprises the following steps of:
step 1: the input image is subjected to feature extraction through a convolution layer having a size of 1 × 1.
Step 2: in the feature mapping part, a conversion layer and a main network branch are adopted to respectively process a feature map; the conversion layer part is used for directly extracting the features of the input image by adding two groups of 3 multiplied by 3 convolutions on the basis of quick connection.
And 3, step 3: in the main network part, the size of the first layer of the traditional convolution layer is 5 multiplied by 5, and the step length is 1; the second layer size is 1 × 1, step size is 1.
And 4, step 4: adding two layers of residual blocks with the size of 3 multiplied by 3, and then adding a cross-channel fusion receptive field module behind the last residual block of the residual modules to improve the feature extraction capability;
specifically, the cross-channel fusion receptive field module is realized by the following steps:
step 401: the pooling portion downsamples the input feature map using pooling windows of 2 × 2, 3 × 3, and 6 × 6, respectively.
Step 402: the preprocessing part adopts a self-normalization activation function Selu to replace the original batch normalization operation, the characteristic graphs which adopt 2 multiplied by 2 and 3 multiplied by 3 to carry out pooling operation are subjected to self-normalization and channel quantity conversion, then continuous down-sampling is carried out until the size of the characteristic graphs is consistent with that of the characteristic graphs of 6 multiplied by 6 branches, the two characteristic graphs are spliced in a cross-layer way with the characteristic graphs on the processed 6 multiplied by 6 branches, and the output x of the l layer is output l Contact is made with all previous output layers.
Step 403: the upsampling part adopts pixel buffer to perform upsampling processing, specifically, the upsampling part adopts pixel buffer
I SR =f L (I LR )=PS(W L *f L-1 (I LR )+b L )
In which I SR Is a high resolution feature map, I LR Is a low resolution feature map, f L For non-linear operation on the L-th layer low resolution feature map, weight W L Is of size n l-1 ×r 2 C×k L ×k L ,b L Is an offset. PS is a periodic screening operation that combines and transforms spatial and channel information to achieve upsampling by computing it as
Figure BDA0003241873790000051
Wherein T is a low-resolution image, r is an image expansion multiple, and x and y are coordinates of pixel points on a high-resolution characteristic diagram. The algorithm can convert H × W × C × r 2 Is converted into a tensor of rH × rW × C.
Step 404: and the characteristic splicing part splices the channel dimensions of the four groups of obtained characteristic graphs.
And 5: and adding a convolution layer with the size of 1 multiplied by 1, and completing the spectral reconstruction of the main network by the convolution layer with the size of 5 multiplied by 5 and the step length of 1.
And 6: and finally, adding the two groups of output characteristic graphs of the conversion layer and the output characteristic graph of the main network to realize characteristic fusion of multiple scales, thereby obtaining the reconstructed hyperspectral image.
Network model and parameter setting:
1. when training the network, an L2 normal form loss function is used, and the expression is
Figure BDA0003241873790000052
Wherein S is the total error, yi represents the value of the ith pixel in the target output characteristic diagram, and f (xi) represents the value of the ith pixel in the actual output characteristic diagram.
2. The network parameter setting of the RGB-hyperspectral image reconstruction algorithm based on the dense multiscale network is shown in table 1, wherein avgpool represents average pooling, pixel shuffle represents that pixel recombination is adopted to realize upsampling, and stride represents a step length.
Table 1 network parameter settings
Figure BDA0003241873790000061
Table 1 (continuous) network parameter settings
Figure BDA0003241873790000062
Figure BDA0003241873790000071
3. Some hyper-parameters during the experiment were set as: the initial learning rate is 0.005, the learning rate is dynamically reduced every 5000 training steps in the learning process, the attenuation coefficient is 0.93, and an Adam optimizer is adopted to optimize the algorithm.
The above description is only a preferred embodiment of the present invention, and is not intended to limit the scope of the present invention.

Claims (3)

1. A method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network is characterized by comprising the following steps:
step 1: performing feature extraction on an input image through a convolution layer with the size of 1 multiplied by 1;
step 2: in the feature mapping part, a conversion layer and a main network branch are adopted to process a feature map respectively; the conversion layer part is characterized in that two groups of 3 multiplied by 3 convolutions are added on the basis of quick connection to directly extract the characteristics of an input image;
and step 3: in the main network part, the size of the first layer of the traditional convolution layer is 5 multiplied by 5, and the step length is 1; the size of the second layer is 1 multiplied by 1, and the step length is 1;
and 4, step 4: adding two layers of residual blocks with the size of 3 multiplied by 3, and then adding a cross-channel fusion receptive field module behind the last residual block of the residual modules to improve the feature extraction capability;
the cross-channel fusion receptive field module specifically comprises: (1) The pooling part respectively adopts 2 multiplied by 2, 3 multiplied by 3 and 6 multiplied by 6 pooling windows to carry out down-sampling operation on the input feature map;
(2) The preprocessing part adopts a self-normalization activation function Selu, continuously down-samples feature graphs subjected to pooling operation by adopting 2 x 2 and 3 x 3 after self-normalization and channel quantity conversion until the feature graphs are consistent with the feature graphs of the 6 x 6 branches in size, and performs cross-layer splicing on the two feature graphs and the processed feature graphs on the 6 x 6 branches;
(3) An upsampling part adopts pixel buffer to perform upsampling processing;
(4) The characteristic splicing part splices the channel dimensions of the four groups of obtained characteristic graphs;
and 5: adding a convolution layer with the size of 1 multiplied by 1, and completing the spectrum reconstruction of the main network by the convolution layer with the size of 5 multiplied by 5 and the step length of 1;
step 6: and finally, adding the two groups of output characteristic graphs of the conversion layer and the output characteristic graph of the main network to realize characteristic fusion of multiple scales, thereby obtaining the reconstructed hyperspectral image.
2. The method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network according to claim 1, wherein the normalized activation function Selu specifically is:
Figure FDA0003891909920000011
wherein the content of the first and second substances,
Figure FDA0003891909920000021
3. the method for reconstructing an RGB image to a hyperspectral image based on a dense multiscale network according to claim 1, wherein the upsampling part performs upsampling processing by using pixel shuffle, and specifically comprises: the expression for performing the upsampling operation by using the pixel buffer is
I SR =f L (I LR )=PS(W L *f L-1 (I LR )+b L )
Wherein I SR Is a high resolution feature map, I LR Is a low resolution feature map, f L For non-linear operation on the L-th layer low resolution feature map, weight W L Is of size n l-1 ×r 2 C×k L ×k L ,b L For biasing, PS is a periodic screening operation that combines and transforms spatial and channel information to achieve upsampling by a calculation
Figure FDA0003891909920000022
Wherein, T is a low-resolution image, r is an image expansion multiple, x and y are coordinates of pixel points on a high-resolution characteristic diagram, H multiplied by W multiplied by C multiplied by r 2 Is converted into a tensor of rH × rW × C.
CN202111020790.6A 2021-09-01 2021-09-01 RGB image-to-hyperspectral image reconstruction method based on dense multiscale network Active CN113902658B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111020790.6A CN113902658B (en) 2021-09-01 2021-09-01 RGB image-to-hyperspectral image reconstruction method based on dense multiscale network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111020790.6A CN113902658B (en) 2021-09-01 2021-09-01 RGB image-to-hyperspectral image reconstruction method based on dense multiscale network

Publications (2)

Publication Number Publication Date
CN113902658A CN113902658A (en) 2022-01-07
CN113902658B true CN113902658B (en) 2023-02-10

Family

ID=79188294

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111020790.6A Active CN113902658B (en) 2021-09-01 2021-09-01 RGB image-to-hyperspectral image reconstruction method based on dense multiscale network

Country Status (1)

Country Link
CN (1) CN113902658B (en)

Families Citing this family (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN114580484B (en) * 2022-04-28 2022-08-12 西安电子科技大学 Small sample communication signal automatic modulation identification method based on incremental learning
CN114757831B (en) * 2022-06-13 2022-09-06 湖南大学 High-resolution video hyperspectral imaging method, device and medium based on intelligent space-spectrum fusion
CN117079105B (en) * 2023-08-04 2024-04-26 中国科学院空天信息创新研究院 Remote sensing image spatial spectrum fusion method and device, electronic equipment and storage medium

Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3382417A2 (en) * 2017-03-28 2018-10-03 Siemens Healthcare GmbH Magnetic resonance image reconstruction system and method
CN108734659A (en) * 2018-05-17 2018-11-02 华中科技大学 A kind of sub-pix convolved image super resolution ratio reconstruction method based on multiple dimensioned label
CN111402138A (en) * 2020-03-24 2020-07-10 天津城建大学 Image super-resolution reconstruction method of supervised convolutional neural network based on multi-scale feature extraction fusion
CN111598778A (en) * 2020-05-13 2020-08-28 云南电网有限责任公司电力科学研究院 Insulator image super-resolution reconstruction method
AU2021103300A4 (en) * 2021-06-11 2021-08-05 Nanjing University Of Aeronautics And Astronautics Unsupervised Monocular Depth Estimation Method Based On Multi- Scale Unification
CN113222822A (en) * 2021-06-02 2021-08-06 西安电子科技大学 Hyperspectral image super-resolution reconstruction method based on multi-scale transformation

Patent Citations (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
EP3382417A2 (en) * 2017-03-28 2018-10-03 Siemens Healthcare GmbH Magnetic resonance image reconstruction system and method
CN108734659A (en) * 2018-05-17 2018-11-02 华中科技大学 A kind of sub-pix convolved image super resolution ratio reconstruction method based on multiple dimensioned label
CN111402138A (en) * 2020-03-24 2020-07-10 天津城建大学 Image super-resolution reconstruction method of supervised convolutional neural network based on multi-scale feature extraction fusion
CN111598778A (en) * 2020-05-13 2020-08-28 云南电网有限责任公司电力科学研究院 Insulator image super-resolution reconstruction method
CN113222822A (en) * 2021-06-02 2021-08-06 西安电子科技大学 Hyperspectral image super-resolution reconstruction method based on multi-scale transformation
AU2021103300A4 (en) * 2021-06-11 2021-08-05 Nanjing University Of Aeronautics And Astronautics Unsupervised Monocular Depth Estimation Method Based On Multi- Scale Unification

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
HSCNN+: Advanced CNN-Based Hyperspectral Recovery from RGB Images;Zhan Shi 等;《2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)》;20181216;全文 *
对抗网络实现单幅RGB重建高光谱图像;刘鹏飞 等;《红外与激光工程》;20200731;第49卷(第S1期);全文 *

Also Published As

Publication number Publication date
CN113902658A (en) 2022-01-07

Similar Documents

Publication Publication Date Title
CN113902658B (en) RGB image-to-hyperspectral image reconstruction method based on dense multiscale network
CN109118432B (en) Image super-resolution reconstruction method based on rapid cyclic convolution network
CN108765296B (en) Image super-resolution reconstruction method based on recursive residual attention network
CN108537733B (en) Super-resolution reconstruction method based on multi-path deep convolutional neural network
CN109087273B (en) Image restoration method, storage medium and system based on enhanced neural network
CN110232661B (en) Low-illumination color image enhancement method based on Retinex and convolutional neural network
CN112750082B (en) Human face super-resolution method and system based on fusion attention mechanism
CN110210608B (en) Low-illumination image enhancement method based on attention mechanism and multi-level feature fusion
CN110599401A (en) Remote sensing image super-resolution reconstruction method, processing device and readable storage medium
CN111598778B (en) Super-resolution reconstruction method for insulator image
CN109064396A (en) A kind of single image super resolution ratio reconstruction method based on depth ingredient learning network
CN112435191B (en) Low-illumination image enhancement method based on fusion of multiple neural network structures
CN109389667B (en) High-efficiency global illumination drawing method based on deep learning
CN111178499B (en) Medical image super-resolution method based on generation countermeasure network improvement
CN111986092B (en) Dual-network-based image super-resolution reconstruction method and system
CN113421187B (en) Super-resolution reconstruction method, system, storage medium and equipment
CN114331831A (en) Light-weight single-image super-resolution reconstruction method
CN112017116B (en) Image super-resolution reconstruction network based on asymmetric convolution and construction method thereof
CN115205147A (en) Multi-scale optimization low-illumination image enhancement method based on Transformer
CN111654621B (en) Dual-focus camera continuous digital zooming method based on convolutional neural network model
CN115100039B (en) Lightweight image super-resolution reconstruction method based on deep learning
CN111951164A (en) Image super-resolution reconstruction network structure and image reconstruction effect analysis method
Hui et al. Two-stage convolutional network for image super-resolution
CN114565539B (en) Image defogging method based on online knowledge distillation
Yang et al. Image super-resolution reconstruction based on improved Dirac residual network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant