CN111797712A - Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network - Google Patents

Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network Download PDF

Info

Publication number
CN111797712A
CN111797712A CN202010545643.XA CN202010545643A CN111797712A CN 111797712 A CN111797712 A CN 111797712A CN 202010545643 A CN202010545643 A CN 202010545643A CN 111797712 A CN111797712 A CN 111797712A
Authority
CN
China
Prior art keywords
cloud
remote sensing
sensing image
network
detection
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202010545643.XA
Other languages
Chinese (zh)
Other versions
CN111797712B (en
Inventor
张秀再
胡敬锋
周丽娟
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Nanjing University of Information Science and Technology
Original Assignee
Nanjing University of Information Science and Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Nanjing University of Information Science and Technology filed Critical Nanjing University of Information Science and Technology
Priority to CN202010545643.XA priority Critical patent/CN111797712B/en
Publication of CN111797712A publication Critical patent/CN111797712A/en
Application granted granted Critical
Publication of CN111797712B publication Critical patent/CN111797712B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V20/00Scenes; Scene-specific elements
    • G06V20/10Terrestrial scenes
    • G06V20/13Satellite images
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/21Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
    • G06F18/214Generating training patterns; Bootstrap methods, e.g. bagging or boosting
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/25Fusion techniques
    • G06F18/253Fusion techniques of extracted features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/0002Inspection of images, e.g. flaw detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/10Image acquisition modality
    • G06T2207/10032Satellite or aerial image; Remote sensing
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20081Training; Learning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/20Special algorithmic details
    • G06T2207/20084Artificial neural networks [ANN]
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30192Weather; Meteorology

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Computation (AREA)
  • General Engineering & Computer Science (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • General Health & Medical Sciences (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Software Systems (AREA)
  • Astronomy & Astrophysics (AREA)
  • Remote Sensing (AREA)
  • Multimedia (AREA)
  • Quality & Reliability (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network, which comprises the steps of constructing a training set comprising a plurality of remote sensing image pictures, manufacturing labels of all pixel points in the remote sensing image pictures of the training set, enabling the labels to form mask images corresponding to the remote sensing image pictures, inputting the mask images and the remote sensing image pictures into a supervision network for training, storing optimal models through 50 epochs and automatically stopping training when cost function loss of the supervision network is convergent and tends to be stable and all indexes of the training set obtain the highest values, obtaining detection models, and adopting the images to be detected in a detection model test set to carry out cloud and cloud shadow detection so as to quickly and efficiently carry out corresponding detection on cloud and cloud shadows in the images to be detected, and improving the accuracy of corresponding detection results.

Description

Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network
Technical Field
The invention relates to the technical field of deep learning and image recognition, in particular to a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network.
Background
In recent years, with the rapid development of remote sensing technology, information contained in various remote sensing satellite images is more and more abundant, and the remote sensing satellite images are more and more widely applied to the fields of agricultural production, natural disaster prediction, military science and technology, geographical mapping, change detection, water conservancy and traffic and the like. The cloud and the cloud shadow blur the spectral information of the optical remote sensing sensor, and are not beneficial to observing the ground object; in addition, long-term observation and practice has shown that cloud generation and dissipation, and evolution and transformation between various types of clouds, are performed under certain conditions of moisture and atmospheric movement. The evolution of the cloud can be seen as the movement of water vapor and atmospheric motion, so that the cloud is closely and inseparably connected with the occurrence of various weather phenomena. Therefore, the detection of cloud and cloud shadow in the remote sensing image is of great significance to research in many fields. At present, various cloud detection methods have been proposed, and these methods can be mainly classified into two types, a multi-image based method and a single-image based method. The multi-image based method requires a set of images to be taken at different times in the same background and requires more images to be acquired in a short time to ensure that the underground surface does not change too much, in which process it is very difficult to acquire a clear reference image. For single image detection methods, early cloud detection methods were based primarily on thresholds, by extracting various spectral features of each pixel, and then using several thresholds to determine the cloud pixels. The threshold-based cloud detection method usually only utilizes low-level spectral information, but ignores spatial information, and has poor cloud detection effect under different imaging conditions due to sensitivity to the underlying surface and sensitivity to the cloud coverage; with the rapid development of big data and machine learning technology, the complex work in cloud detection can be realized by a computer, so that the labor cost can be greatly saved, and the cloud detection method based on machine learning is more and more widely applied. Such as Xu et al extract cloud boundaries using a decision tree. Hu et al combine computer vision with a random forest algorithm to obtain a cloud-overlaid image map. In general, machine learning methods yield more accurate cloud detection results than threshold methods. However, these methods based on machine learning use a hand-made feature and a simple classifier, and do not use semantic-level information. These artificially designed features rely on expert a priori knowledge and sensors, it is difficult to accurately segment the features of the cloud with a complex underlying surface, and their detection performance depends mainly on the choice of manual features.
Recently, deep neural networks have been rapidly developed in many fields, the powerful segmentation capability thereof has been widely recognized, and cloud detection methods based on deep learning have attracted attention. Shi and S et al used superpixel segmentation and Full Convolution Neural Networks (FCNN) to detect clouds in Quickbird, *** earth images, and Landsat8 images. Chai et al used an adaptive SegNet to detect clouds and cloud shadows in Landsat images. Sorour et al use a full convolution neural network (FCN) for cloud detection, the algorithm includes a pre-processing thresholding step for ice and snow retrieval, and uses the existing landsat8 cloud mask as a premise for improved performance. Xu et al use a depth residual network to detect clouds and cloud shadows in remote sensing images. Wu and the like acquire a cloud probability map by using a deep convolutional neural network, and then obtain a refined high-resolution first-cloud mask by using a composite image filtering technology. The traditional technology only uses spectral features to hardly distinguish the cloud from some bright ground objects (such as snow and white buildings), so that the detection accuracy of the cloud and cloud shadow in the remote sensing image is low.
Disclosure of Invention
Aiming at the problems, the invention provides a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network.
In order to realize the purpose of the invention, the invention provides a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network, which comprises the following steps:
s10, constructing a training set comprising a plurality of remote sensing image pictures, making labels of all pixel points in the remote sensing image pictures of the training set, enabling the labels to form mask images corresponding to the remote sensing image pictures, and inputting the mask images and the remote sensing image pictures into a supervision network for training;
s20, when the loss of the cost function of the supervision network is convergent and tends to be stable, and each index of the training set obtains the highest value, storing the optimal model through 50 epochs and automatically stopping training to obtain a detection model;
and S30, carrying out cloud and cloud shadow detection by using the to-be-detected images in the detection model test set.
In one embodiment, the supervisory network includes a res.block module to prevent network degradation, a multi-scale convolution module to increase the network's receptive field, and a multi-scale feature module to extract different scale information.
Specifically, the experimental parameters of the supervision network during training include: learning rate of 0.0001, batch size of 8, first momentum size beta10.9 and a second momentum magnitude beta2At 0.999, the learning rate is set to 0.0001, and the minimum learning rate is set to 0.0000001.
Specifically, in the res.block module, a convolution of 1 × 1 is added before and after the convolution of 3 × 3.
Specifically, the multi-scale convolution module includes 3 convolution kernels, each convolution kernel having a size of 1 × 1, 3 × 3, and 5 × 5, respectively, and each convolution kernel uses an expansion rate of 1, 2, and 4, respectively.
Specifically, in the training process of the supervision network, the information extracted by the encoder is processed by each convolution kernel to obtain a plurality of sampling features, the plurality of sampling features are subjected to feature fusion with the decoder, and the fusion features are extracted by each convolution kernel.
In an embodiment, the method for detecting cloud and cloud shadow of remote sensing image based on multi-scale feature fusion network further includes:
obtaining average cross-over ratio, precision rate, recall rate, precision rate and harmonic mean value generated in the process of detecting each image to be detected of the detection model, determining the detection performance of the detection model according to the average cross-over ratio, precision rate, recall rate, precision rate and harmonic mean value,
specifically, the average intersection ratio includes:
Figure BDA0002540613440000031
the essenceThe accuracy rate comprises:
Figure BDA0002540613440000032
the recall rate includes:
Figure BDA0002540613440000033
the accuracy rate comprises:
Figure BDA0002540613440000034
the blended mean includes:
Figure BDA0002540613440000035
wherein mIoU represents the average cross-over ratio, CCCNumber of correctly detected cloud pixels, CNCNumber of cloud pixels representing false detection as non-cloud pixels, CNNRepresenting the number of correctly detected non-cloud pixels, Precison representing Accuracy, Recall representing Recall, Accuracy representing Accuracy, F1scoreRepresenting the harmonic mean.
According to the remote sensing image cloud and cloud shadow detection method based on the multi-scale feature fusion network, a training set comprising a plurality of remote sensing image pictures is constructed, labels of all pixel points in the remote sensing image pictures of the training set are made, the labels form mask images corresponding to the remote sensing image pictures, the mask images and the remote sensing image pictures are input into a monitoring network for training, when the loss of cost functions of the monitoring network is convergent and tends to be stable, and all indexes of the training set obtain the highest values, the optimal model is stored through 50 epochs and the training is automatically stopped, a detection model is obtained, the detection model is adopted to test images to be detected in the set for cloud and cloud shadow detection, so that the cloud and cloud shadows in the images to be detected are correspondingly detected quickly and efficiently, and the accuracy of corresponding detection results can be improved.
Drawings
Fig. 1 is a flowchart of a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network according to an embodiment;
fig. 2 is a flowchart of a remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network according to another embodiment;
fig. 3 is a block module schematic diagram of an embodiment;
FIG. 4 is a schematic diagram of a multi-scale convolution module of an embodiment;
FIG. 5 is a multi-scale feature module diagram of an embodiment.
Detailed Description
In order to make the objects, technical solutions and advantages of the present application more apparent, the present application is described in further detail below with reference to the accompanying drawings and embodiments. It should be understood that the specific embodiments described herein are merely illustrative of the present application and are not intended to limit the present application.
Reference herein to "an embodiment" means that a particular feature, structure, or characteristic described in connection with the embodiment can be included in at least one embodiment of the application. The appearances of the phrase in various places in the specification are not necessarily all referring to the same embodiment, nor are separate or alternative embodiments mutually exclusive of other embodiments. It is explicitly and implicitly understood by one skilled in the art that the embodiments described herein can be combined with other embodiments.
The technical problem that this application will solve is to overcome prior art and only use spectral feature to be difficult to distinguish cloud and some bright ground thing (e.g. snow, white building). Due to the limited spectral range (including blue, green, red, near infrared bands) of clouds and cloud shadows, the complexity of the underlying surface, accurate detection of clouds and cloud shadows is difficult for optical high resolution images. The existence of clouds and cloud shadows has a significant impact on the efficient late-stage application of optical high-resolution images, and manual labeling of cloud and cloud shadow pixels requires a great deal of time and human resources. To solve this problem, a detection algorithm based on a Multi-scale Feature Fusion Network (MFFN) model is proposed. The algorithm mainly comprises a Res.block module, a multi-scale convolution (MCM) and a multi-scale feature module (MFM), can extract rich spatial information and semantic information, can detect the cloud and cloud shadow in the remote sensing image at a pixel level, and can obtain a relatively fine edge.
Referring to fig. 1, fig. 1 is a flowchart of a method for detecting cloud and cloud shadow of a remote sensing image based on a multi-scale feature fusion network according to an embodiment, and includes the following steps:
s10, constructing a training set comprising a plurality of remote sensing image pictures, making labels of all pixel points in the remote sensing image pictures of the training set, enabling the labels to form mask images corresponding to the remote sensing image pictures, and inputting the mask images and the remote sensing image pictures into a supervision network for training.
The training set may include a plurality of remote sensing imagery pictures. In consideration of the video memory of a computer and the calculation speed, the picture (remote sensing image picture) is cut into 256 multiplied by 256 pixels, so that a training set is constructed, and the training speed can be improved. In the label manufacturing process, whether each pixel point is 'cloud', 'cloud shadow' or 'non-cloud' is respectively represented by 255, 128 and 0, and a mask image corresponding to the original image is formed for supervising the training of the network.
And S20, when the loss of the cost function of the supervision network is convergent and tends to be stable, and each index of the training set obtains the highest value, storing the optimal model through 5 epochs and automatically stopping training to obtain the detection model.
In one embodiment, the supervisory network includes a res.block (residual module) to prevent network degradation, a multi-scale convolution module to increase the network's receptive field, and a multi-scale feature module to extract different scale information.
In one example, the experimental parameters of the supervising network while training include: learning rate of 0.0001, batch size of 8, first momentum size beta10.9 and a second momentum magnitude beta2At 0.999, the learning rate is set to 0.0001, and the minimum learning rate is set to 0.0000001.
In this example, the experimental parameter settings are: learning rate of 0.0001, batch size of 8, set momentum size beta10.9 and β2The learning rate was attenuated to 0.999 to prevent overfitting, and was set to0.0001. the learning rate is reduced to one tenth of the original one every 10 epochs of iteration (one epoch after all data sets are iterated), and the minimum learning rate is set to 0.0000001. When the cost function loss is convergent and tends to be stable, and each index (accuracy rate and average cross-over ratio) of the training set and the verification set obtains the highest value, the optimal model is stored through 50 epochs, and the training is automatically stopped.
In one example, the res.block module adds a 1 × 1 convolution before and after a 3 × 3 convolution.
In res.block used in the constructed network model constructed in this example, one convolution of 1 × 1 is added before and after a convolution of 3 × 3, which can effectively reduce network parameters and suppress degradation with the depth of the network.
In one example, the multi-scale convolution module includes 3 convolution kernels, each convolution kernel being 1 × 1, 3 × 3, and 5 × 5 in size, respectively, and each convolution kernel using a dilation rate of 1, 2, and 4, respectively.
The model obtains different sampling characteristics from information extracted by an encoder through different convolution kernels, then performs characteristic fusion with a decoder, and extracts characteristics through three different convolution kernels, so that richer spatial information and semantic information are obtained, and the boundary of a segmentation result can be finer.
In one example, in the training process of the supervision network, the information extracted by the encoder is processed by each convolution kernel to obtain a plurality of sampling features, the plurality of sampling features are subjected to feature fusion with the decoder, and the fusion features are extracted by each convolution kernel.
The example adopts a multi-scale feature module similar to a PPM (pyramid pooling model) to extract multi-scale features, performs down-sampling through four parallel average pooling, obtains feature maps with sizes of 8 × 8, 6 × 6, 2 × 2 and 1 × 1 respectively, and then performs up-sampling on the pooled feature maps to the same size through a bilinear interpolation method to realize the function of multi-scale feature fusion.
Further, in the process of making the label, "cloud", "cloud shadow" and "non-cloud" are respectively represented by 255, 128 and 0, so that when the model is called, 2, 1 and 0 corresponding to each pixel recognized by the model are respectively converted into 255, 128 and 0, that is, the cloud is white, the cloud shadow is gray and the non-cloud is black in the recognized image.
And S30, carrying out cloud and cloud shadow detection by using the to-be-detected images in the detection model test set.
In the above steps, the trained semantic segmentation model (detection model) can be used for performing cloud and cloud shadow detection on the images in the test set.
According to the remote sensing image cloud and cloud shadow detection method based on the multi-scale feature fusion network, a training set comprising a plurality of remote sensing image pictures is constructed, labels of all pixel points in the remote sensing image pictures of the training set are made, the labels form mask images corresponding to the remote sensing image pictures, the mask images and the remote sensing image pictures are input into a monitoring network for training, when the loss of cost functions of the monitoring network is convergent and tends to be stable, and all indexes of the training set obtain the highest values, the optimal model is stored through 50 epochs and the training is automatically stopped, a detection model is obtained, the detection model is adopted to test images to be detected in the set for cloud and cloud shadow detection, so that the cloud and cloud shadows in the images to be detected are correspondingly detected quickly and efficiently, and the accuracy of corresponding detection results can be improved.
In an embodiment, the method for detecting cloud and cloud shadow of remote sensing image based on multi-scale feature fusion network further includes:
obtaining average cross-over ratio, precision rate, recall rate, precision rate and harmonic mean value generated in the process of detecting each image to be detected of the detection model, determining the detection performance of the detection model according to the average cross-over ratio, precision rate, recall rate, precision rate and harmonic mean value,
specifically, the average intersection ratio includes:
Figure BDA0002540613440000061
the accuracy rate includes:
Figure BDA0002540613440000062
the recall rate includes:
Figure BDA0002540613440000071
the accuracy rate comprises:
Figure BDA0002540613440000072
the blended mean includes:
Figure BDA0002540613440000073
wherein mIoU represents the average cross-over ratio, CCCNumber of correctly detected cloud pixels, CNCNumber of cloud pixels representing false detection as non-cloud pixels, CNNRepresenting the number of correctly detected non-cloud pixels, Precison representing Accuracy, Recall representing Recall, Accuracy representing Accuracy, F1scoreRepresenting the harmonic mean.
Compared with the traditional scheme, the remote sensing image cloud and cloud shadow detection method based on the multi-scale feature fusion network has the following technical effects:
(1) the Res.block modules used in the constructed network model respectively increase a convolution of 1 multiplied by 1 before and after a convolution of 3 multiplied by 3, and an activation function used after three parallel convolutions is leak _ ReLU, so that compared with the use of two continuous convolutions of 3 multiplied by 3, the network model can effectively reduce network parameters and can inhibit the phenomenon of degradation along with the deepening of the network depth;
(2) the multi-scale convolution (MCM) module used in the model comprises three convolutions, the sizes of convolution kernels are respectively 1 × 1, 3 × 3 and 5 × 5, the receptive field of a neural network can be increased by expanding the convolutions, the expansion rates used by the three convolutions are respectively 1, 2 and 4, the model obtains different sampling characteristics from information extracted by an encoder through different convolution kernels, then the sampling characteristics are fused with a decoder, and the characteristics are extracted by using the three different convolution kernels, so that richer space information and semantic information are obtained, and the boundary of a segmentation result can be finer;
(3) the multi-scale feature module adopted in the model is similar to a PPM (pyramid pooling model) to extract multi-scale features, down-sampling is carried out through four parallel average pooling, the sizes of obtained feature maps are respectively 8 multiplied by 8, 6 multiplied by 6, 2 multiplied by 2 and 1 multiplied by 1, and then the pooled feature maps are up-sampled to the same size through a bilinear interpolation method so as to realize the function of multi-scale feature fusion.
(4) The cloud and cloud shadow with the complex underlying surface can be well identified, and the robustness is good.
In an embodiment, the method for detecting cloud and cloud shadow of remote sensing images based on the multi-scale feature fusion network can be as shown in fig. 2, firstly, inputting a data set and a corresponding label into a network, wherein the label is used for supervising network training, calculating a probability value of distribution of each pixel class by using a feature input loss function extracted finally, calculating a loss value according to the probability value, and stopping training when the loss value tends to be stable and an mlio u obtains a highest value. And after the network training is finished, calling the trained model, and inputting the model into a classifier to perform pixel-level cloud, cloud shadow and clear sky recognition. In one example, fig. 3 is a res.block block schematic; FIG. 4 is a schematic diagram of a multi-scale convolution module; FIG. 5 is a schematic diagram of a multi-scale feature module.
The data used in the experiment were Landsat8 remote sensing satellite images. Because the data contains limited information, the data is enhanced firstly, and the data set is expanded in modes of cutting, rotating, turning, adding noise and the like. And cutting the enhanced Landsat8 remote sensing data into 256 multiplied by 256 image blocks, wherein 6800 images are obtained, 6000 images are used as a training set, and 800 images are used as a test set. In the process of making the label image, the label image and the training data are subjected to the same manufacturing operations of rotation, overturning and the like so as to realize supervision training on each pixel, and gray values of 0, 128 and 255 are respectively used for representing a clear sky pixel, a cloud shadow pixel and a cloud pixel, namely three labels corresponding to three categories.
In the obtained deep semantic segmentation network model (detection), the overall framework of the multi-feature fusion segmentation network comprises four Res.block modules, four MCM modules, an MFM module and a decoder. The input of the network is 256 × 256 × N (the first two numbers are the dimensionality of data, N is the number of channels of data, e.g., RGB image N is 3) Landsat8 remote sensing image. Two convolution kernels with step size 1, size 3 × 3 and 1 × 1 are used in extracting image features, and an activation function ReLU is used after each convolution. The average pooling with a pooling step of 2 and a filter size of 2 is used in the network. The last layer of the encoder is input into an MFM module, the dimensions of four downsampling features passing through the MFM are reduced to 1/4 when the MFM is input through convolution of 1 multiplied by 1, then upsampling is carried out through a bilinear interpolation method, the sizes of four feature mapping graphs are the same, and then feature fusion is carried out on the output of the last Res.block module and the output feature graph of the MFM module, so that multi-scale context information is obtained. The output of each Res.block module passes through an MCM module to obtain the characteristics after multi-scale convolution, so that the cloud and cloud shadow boundary information can be more completely extracted. The feature graph obtained by the encoder module is subjected to up-sampling through bilinear interpolation, then is subjected to dimensionality reduction through 1 x 1 convolution, and then is subjected to feature fusion with the output feature of the MCM module. And finally, performing feature fusion on the output of each layer of the decoder through rich jump connection, and inputting the fused features into a classifier for 3-class classification. The end-to-end pixel-to-pixel semantic segmentation function can be realized, namely the input data and the output characteristic diagram are the same in size, and each pixel in the image is divided into a cloud pixel, a cloud shadow pixel and a clear sky pixel.
As the algorithm needs to solve the problem of multi-classification, a Softmax function is used when the probability value of the distribution of each pixel class is judged, a loss value is calculated by using the Softmax function, and a loss function J (theta) formula is defined as
Figure BDA0002540613440000081
In the formula, yiIs asThe identification labels have K different values, and the remote sensing image is divided into cloud, cloud shadow and clear sky, which belong to the three classification problems, so that K is 3, hiObserving vector H ═ H for input image pixel1,h2,h3...hmThe element in the equation, theta is a model parameter, m is the pixel number of the image, and 1 is an explicit function.
In the visual comparison of the detection of Landsat8 remote sensing image cloud and cloud shadow by the algorithm of the embodiment, the K-means algorithm and the reset algorithm, it can be found that more accumulated clouds in the original image accompany with the cloud shadow, the shape is irregular, and partial high-brightness objects exist in the underlying surface, so that the detection result is easily interfered. The K-means algorithm is easily interfered by a highlight object on the underlying surface, the highlight object is falsely detected as cloud, the cloud and cloud shadow edge detection is not fine enough, and the detail loss is serious in the cloud shadow detection process; the resuret algorithm is susceptible to the influence of a small amount of highlight ground objects, and partial details of cloud shadow are lost; the algorithm can effectively detect the cloud and the cloud shadow, has a fine detection result on the edge of the cloud and the cloud shadow, is not easily interfered by a white object on an underlying surface, can obtain a better detection result on a fine cloud area, and has a lower final false detection rate. The algorithm can extract more effective features, so that the visual detection result is closer to a true value.
In the visual comparison process of the algorithm of the embodiment with the K-means algorithm and the reset algorithm for the detection of Landsat8 and cloud shadow, the original image contains a large amount of ice/snow which easily interferes with the detection of a cloud area, and the detection difficulty of cloud and cloud shadow is increased. The K-means algorithm cannot distinguish ice/snow from clouds, and details in cloud shadow results are seriously lost; the resurset can better distinguish the cloud from the ice/snow, but the cloud and cloud shadow detection result has smooth edges and more lost details, and partial cloud shadow false detection and a small amount of thin cloud missing detection exist in the image. The algorithm can effectively distinguish the cloud from the ice/snow, is not easy to be interfered, obtains a better result for detecting the cloud shadow, and is generally superior to a K-means algorithm and a reset algorithm.
In the process of visually comparing the algorithm of the embodiment with the K-means algorithm and the reset to the cloud shadow detection of the Landsat8 image, it can be known that the K-means algorithm has a large error in the result of the cloud shadow detection, and the details are seriously lost; the algorithm of the embodiment obtains a better result for the detection of the cloud shadow, the detection result is obviously superior to the K-means algorithm and the reset algorithm, the edge details of the cloud shadow can be accurately detected, and the edge details are closer to the true value visually.
Table 1 is an average value of quantitative comparison indexes of cloud detection of Landsat8 remote sensing images by three algorithms. As can be seen from Table 1, the K-means algorithm is easily interfered by the objects on the underlying surface, and all detection indexes are the lowest; compared with a reset algorithm, the MFFN has obvious advantages for cloud detection, and each index is far higher than that of the reset algorithm, so that the MFFN is more suitable for cloud detection. In the process of extracting information, the MFFN algorithm uses multi-scale sampling, hole convolution, residual modules and rich jump connection to ensure that less useful information is lost, and a network structure is richer than spatial information and semantic information fused by a reset structure.
TABLE 1
Figure BDA0002540613440000091
Table 2 is an average value of the cloud shadow detection quantitative evaluation indexes of the Landsat8 remote sensing image by the three algorithms. As can be seen from table 2, the MFFN algorithm can obtain a better result for detecting cloud shadows, which proves that the model can well extract cloud and cloud shadow information; due to the lack of multi-scale feature extraction in the reset algorithm and the fact that multi-scale convolution is not used in the process of jump connection, although feature fusion exists, a large amount of information is still lost, and all indexes of cloud shadow detection are lower than those of the MFFN algorithm. Compared with the two algorithms, the cloud shadow detection indexes are lower than those of the cloud shadow detection, the cloud shadow learning of the model is insufficient due to relatively less cloud shadow information in the data set, and meanwhile, the cloud shadow is not irregular relative to the cloud and is easily interfered by black objects (such as shadows of buildings, trees, oceans and the like) on the underlying surface, so that the cloud shadow detection result is not ideal.
TABLE 2
Figure BDA0002540613440000101
The technical features of the above embodiments can be arbitrarily combined, and for the sake of brevity, all possible combinations of the technical features in the above embodiments are not described, but should be considered as the scope of the present specification as long as there is no contradiction between the combinations of the technical features.
It should be noted that the terms "first \ second \ third" referred to in the embodiments of the present application merely distinguish similar objects, and do not represent a specific ordering for the objects, and it should be understood that "first \ second \ third" may exchange a specific order or sequence when allowed. It should be understood that "first \ second \ third" distinct objects may be interchanged under appropriate circumstances such that the embodiments of the application described herein may be implemented in an order other than those illustrated or described herein.
The terms "comprising" and "having" and any variations thereof in the embodiments of the present application are intended to cover non-exclusive inclusions. For example, a process, method, apparatus, product, or device that comprises a list of steps or modules is not limited to the listed steps or modules but may alternatively include other steps or modules not listed or inherent to such process, method, product, or device.
The above-mentioned embodiments only express several embodiments of the present application, and the description thereof is more specific and detailed, but not construed as limiting the scope of the invention. It should be noted that, for a person skilled in the art, several variations and modifications can be made without departing from the concept of the present application, which falls within the scope of protection of the present application. Therefore, the protection scope of the present patent shall be subject to the appended claims.

Claims (8)

1. A remote sensing image cloud and cloud shadow detection method based on a multi-scale feature fusion network is characterized by comprising the following steps:
s10, constructing a training set comprising a plurality of remote sensing image pictures, making labels of all pixel points in the remote sensing image pictures of the training set, enabling the labels to form mask images corresponding to the remote sensing image pictures, and inputting the mask images and the remote sensing image pictures into a supervision network for training;
s20, when the loss of the cost function of the supervision network is convergent and tends to be stable, and each index of the training set obtains the highest value, storing the optimal model through 50 epochs and automatically stopping training to obtain a detection model;
and S30, carrying out cloud and cloud shadow detection by using the to-be-detected images in the detection model test set.
2. The method for detecting the cloud and cloud shadow of the remote sensing image based on the multi-scale feature fusion network as claimed in claim 1, wherein the supervision network comprises a Res.block module for preventing network degradation, a multi-scale convolution module for increasing a network reception field, and a multi-scale feature module for extracting different scale information.
3. The method for detecting cloud and cloud shadow of remote sensing images based on the multi-scale feature fusion network as claimed in claim 2, wherein experimental parameters of the supervision network during training comprise: learning rate of 0.0001, batch size of 8, first momentum size beta10.9 and a second momentum magnitude beta2At 0.999, the learning rate is set to 0.0001, and the minimum learning rate is set to 0.0000001.
4. The method for detecting the cloud and cloud shadow of the remote sensing image based on the multi-scale feature fusion network as claimed in claim 2, wherein a convolution of 1 x 1 is respectively added before and after a convolution of 3 x 3 in the Res.
5. The method for detecting the cloud and the cloud shadow of the remote sensing image based on the multi-scale feature fusion network as claimed in claim 2, wherein the multi-scale convolution module comprises 3 convolution kernels, the sizes of the convolution kernels are 1 x 1, 3 x 3 and 5 x 5 respectively, and the expansion rates used by the convolution kernels are 1, 2 and 4 respectively.
6. The method for detecting the cloud and the cloud shadow of the remote sensing image based on the multi-scale feature fusion network as claimed in claim 4, wherein in the training process of the supervision network, the information extracted by the encoder is processed by each convolution kernel to obtain a plurality of sampling features, the plurality of sampling features are subjected to feature fusion with the decoder, and the features after fusion are extracted by each convolution kernel.
7. The method for detecting the cloud and cloud shadow of the remote sensing image based on the multi-scale feature fusion network according to any one of claims 1 to 6, further comprising:
and obtaining an average intersection ratio, an accuracy rate, a recall rate, an accuracy rate and a harmonic mean value generated in the detection process of each image to be detected of the detection model, and determining the detection performance of the detection model according to the average intersection ratio, the accuracy rate, the recall rate, the accuracy rate and the harmonic mean value.
8. The method for detecting cloud and cloud shadow of remote sensing image based on multi-scale feature fusion network according to claim 7,
the average intersection ratio comprises:
Figure FDA0002540613430000021
the accuracy rate includes:
Figure FDA0002540613430000022
the recall rate includes:
Figure FDA0002540613430000023
the accuracy rate comprises:
Figure FDA0002540613430000024
the blended mean includes:
Figure FDA0002540613430000025
wherein mIoU represents the average cross-over ratio, CCCNumber of correctly detected cloud pixels, CNCNumber of cloud pixels representing false detection as non-cloud pixels, CNNRepresenting the number of correctly detected non-cloud pixels, Precison representing Accuracy, Recall representing Recall, Accuracy representing Accuracy, F1scoreRepresenting the harmonic mean.
CN202010545643.XA 2020-06-16 2020-06-16 Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network Active CN111797712B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010545643.XA CN111797712B (en) 2020-06-16 2020-06-16 Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010545643.XA CN111797712B (en) 2020-06-16 2020-06-16 Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network

Publications (2)

Publication Number Publication Date
CN111797712A true CN111797712A (en) 2020-10-20
CN111797712B CN111797712B (en) 2023-09-15

Family

ID=72803338

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010545643.XA Active CN111797712B (en) 2020-06-16 2020-06-16 Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network

Country Status (1)

Country Link
CN (1) CN111797712B (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113223040A (en) * 2021-05-17 2021-08-06 中国农业大学 Remote sensing-based banana yield estimation method and device, electronic equipment and storage medium
CN113436115A (en) * 2021-07-30 2021-09-24 西安热工研究院有限公司 Image shadow detection method based on depth unsupervised learning
CN113870124A (en) * 2021-08-25 2021-12-31 西北工业大学 Dual-network mutual excitation learning shadow removing method based on weak supervision
CN114494821A (en) * 2021-12-16 2022-05-13 广西壮族自治区自然资源遥感院 Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation
CN114693670A (en) * 2022-04-24 2022-07-01 西京学院 Ultrasonic detection method for weld defects of longitudinal submerged arc welded pipe based on multi-scale U-Net
CN114943963A (en) * 2022-04-29 2022-08-26 南京信息工程大学 Remote sensing image cloud and cloud shadow segmentation method based on double-branch fusion network
CN115410074A (en) * 2022-07-19 2022-11-29 中国科学院空天信息创新研究院 Remote sensing image cloud detection method and device
CN116340733A (en) * 2023-05-29 2023-06-27 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) On-orbit cloud detection method and remote sensing data processing method based on 0-level remote sensing signals
CN116740569A (en) * 2023-06-15 2023-09-12 安徽理工大学 Deep learning-based snowfall area cloud detection system

Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180260956A1 (en) * 2017-03-10 2018-09-13 TuSimple System and method for semantic segmentation using hybrid dilated convolution (hdc)
US20190057507A1 (en) * 2017-08-18 2019-02-21 Samsung Electronics Co., Ltd. System and method for semantic segmentation of images
CN109584246A (en) * 2018-11-16 2019-04-05 成都信息工程大学 Based on the pyramidal DCM cardiac muscle diagnosis and treatment irradiation image dividing method of Analysis On Multi-scale Features
CN109934200A (en) * 2019-03-22 2019-06-25 南京信息工程大学 A kind of RGB color remote sensing images cloud detection method of optic and system based on improvement M-Net
CN110119728A (en) * 2019-05-23 2019-08-13 哈尔滨工业大学 Remote sensing images cloud detection method of optic based on Multiscale Fusion semantic segmentation network
CN110264484A (en) * 2019-06-27 2019-09-20 上海海洋大学 A kind of improvement island water front segmenting system and dividing method towards remotely-sensed data
CN110826596A (en) * 2019-10-09 2020-02-21 天津大学 Semantic segmentation method based on multi-scale deformable convolution
CN111079739A (en) * 2019-11-28 2020-04-28 长沙理工大学 Multi-scale attention feature detection method
CN111127493A (en) * 2019-11-12 2020-05-08 中国矿业大学 Remote sensing image semantic segmentation method based on attention multi-scale feature fusion

Patent Citations (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20180260956A1 (en) * 2017-03-10 2018-09-13 TuSimple System and method for semantic segmentation using hybrid dilated convolution (hdc)
US20190057507A1 (en) * 2017-08-18 2019-02-21 Samsung Electronics Co., Ltd. System and method for semantic segmentation of images
CN109584246A (en) * 2018-11-16 2019-04-05 成都信息工程大学 Based on the pyramidal DCM cardiac muscle diagnosis and treatment irradiation image dividing method of Analysis On Multi-scale Features
CN109934200A (en) * 2019-03-22 2019-06-25 南京信息工程大学 A kind of RGB color remote sensing images cloud detection method of optic and system based on improvement M-Net
CN110119728A (en) * 2019-05-23 2019-08-13 哈尔滨工业大学 Remote sensing images cloud detection method of optic based on Multiscale Fusion semantic segmentation network
CN110264484A (en) * 2019-06-27 2019-09-20 上海海洋大学 A kind of improvement island water front segmenting system and dividing method towards remotely-sensed data
CN110826596A (en) * 2019-10-09 2020-02-21 天津大学 Semantic segmentation method based on multi-scale deformable convolution
CN111127493A (en) * 2019-11-12 2020-05-08 中国矿业大学 Remote sensing image semantic segmentation method based on attention multi-scale feature fusion
CN111079739A (en) * 2019-11-28 2020-04-28 长沙理工大学 Multi-scale attention feature detection method

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
JIN CHEN 等: "AtICNet: semantic segmentation with atrous spatial pyramid pooling in image cascade network", 《EURASIP JOURNAL ON WIRELESS COMMUNICATIONS AND NETWORKING》 *
LIANG-CHIEH CHEN 等: "DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs", 《ARXIV》 *
RONGHUA SHANG 等: "Multi-scale Adaptive Feature Fusion Network for Semantic Segmentation in Remote Sensing Images", 《REMOTE SENSING》 *
姚群力 等: "基于多尺度融合特征卷积神经网络的遥感图像飞机目标检测", 《测绘学报》 *
杨昌军 等: "基于多尺度特征融合网络的云和云阴影检测试验", 《大气科学》 *
邓志鹏 等: "基于多尺度形变特征卷积网络的高分辨率遥感影像目标检测", 《测绘学报》 *

Cited By (17)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN113223040A (en) * 2021-05-17 2021-08-06 中国农业大学 Remote sensing-based banana yield estimation method and device, electronic equipment and storage medium
CN113223040B (en) * 2021-05-17 2024-05-14 中国农业大学 Banana estimated yield method and device based on remote sensing, electronic equipment and storage medium
CN113436115A (en) * 2021-07-30 2021-09-24 西安热工研究院有限公司 Image shadow detection method based on depth unsupervised learning
CN113436115B (en) * 2021-07-30 2023-09-19 西安热工研究院有限公司 Image shadow detection method based on depth unsupervised learning
CN113870124B (en) * 2021-08-25 2023-06-06 西北工业大学 Weak supervision-based double-network mutual excitation learning shadow removing method
CN113870124A (en) * 2021-08-25 2021-12-31 西北工业大学 Dual-network mutual excitation learning shadow removing method based on weak supervision
CN114494821A (en) * 2021-12-16 2022-05-13 广西壮族自治区自然资源遥感院 Remote sensing image cloud detection method based on feature multi-scale perception and self-adaptive aggregation
CN114693670A (en) * 2022-04-24 2022-07-01 西京学院 Ultrasonic detection method for weld defects of longitudinal submerged arc welded pipe based on multi-scale U-Net
CN114693670B (en) * 2022-04-24 2023-05-23 西京学院 Ultrasonic detection method for weld defects of longitudinal submerged arc welded pipe based on multi-scale U-Net
CN114943963B (en) * 2022-04-29 2023-07-04 南京信息工程大学 Remote sensing image cloud and cloud shadow segmentation method based on double-branch fusion network
CN114943963A (en) * 2022-04-29 2022-08-26 南京信息工程大学 Remote sensing image cloud and cloud shadow segmentation method based on double-branch fusion network
CN115410074B (en) * 2022-07-19 2023-08-29 中国科学院空天信息创新研究院 Remote sensing image cloud detection method and device
CN115410074A (en) * 2022-07-19 2022-11-29 中国科学院空天信息创新研究院 Remote sensing image cloud detection method and device
CN116340733A (en) * 2023-05-29 2023-06-27 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) On-orbit cloud detection method and remote sensing data processing method based on 0-level remote sensing signals
CN116340733B (en) * 2023-05-29 2023-09-01 哈尔滨工业大学(深圳)(哈尔滨工业大学深圳科技创新研究院) On-orbit cloud detection method and remote sensing data processing method based on 0-level remote sensing signals
CN116740569A (en) * 2023-06-15 2023-09-12 安徽理工大学 Deep learning-based snowfall area cloud detection system
CN116740569B (en) * 2023-06-15 2024-01-16 安徽理工大学 Deep learning-based snowfall area cloud detection system

Also Published As

Publication number Publication date
CN111797712B (en) 2023-09-15

Similar Documents

Publication Publication Date Title
CN111797712B (en) Remote sensing image cloud and cloud shadow detection method based on multi-scale feature fusion network
CN109934200B (en) RGB color remote sensing image cloud detection method and system based on improved M-Net
CN110119728B (en) Remote sensing image cloud detection method based on multi-scale fusion semantic segmentation network
CN108573276B (en) Change detection method based on high-resolution remote sensing image
CN111986099B (en) Tillage monitoring method and system based on convolutional neural network with residual error correction fused
CN108510504B (en) Image segmentation method and device
CN111915592B (en) Remote sensing image cloud detection method based on deep learning
CN109145872B (en) CFAR and Fast-RCNN fusion-based SAR image ship target detection method
Isikdogan et al. Seeing through the clouds with deepwatermap
CN109558806B (en) Method for detecting high-resolution remote sensing image change
CN111914686B (en) SAR remote sensing image water area extraction method, device and system based on surrounding area association and pattern recognition
CN111274865A (en) Remote sensing image cloud detection method and device based on full convolution neural network
CN113936217A (en) Priori semantic knowledge guided high-resolution remote sensing image weakly supervised building change detection method
CN111783523B (en) Remote sensing image rotating target detection method
Zheng et al. Single image cloud removal using U-Net and generative adversarial networks
CN111046880A (en) Infrared target image segmentation method and system, electronic device and storage medium
CN110796009A (en) Method and system for detecting marine vessel based on multi-scale convolution neural network model
CN112614136B (en) Infrared small target real-time instance segmentation method and device
CN110751075A (en) Remote sensing image culture pond detection method based on example segmentation
CN109829423B (en) Infrared imaging detection method for frozen lake
CN110717921B (en) Full convolution neural network semantic segmentation method of improved coding and decoding structure
CN114943893B (en) Feature enhancement method for land coverage classification
CN114612769A (en) Integrated sensing infrared imaging ship detection method integrated with local structure information
CN114943876A (en) Cloud and cloud shadow detection method and device for multi-level semantic fusion and storage medium
CN113887472A (en) Remote sensing image cloud detection method based on cascade color and texture feature attention

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant