CN111462124A - Remote sensing satellite cloud detection method based on Deep L abV3+ - Google Patents

Remote sensing satellite cloud detection method based on Deep L abV3+ Download PDF

Info

Publication number
CN111462124A
CN111462124A CN202010241130.XA CN202010241130A CN111462124A CN 111462124 A CN111462124 A CN 111462124A CN 202010241130 A CN202010241130 A CN 202010241130A CN 111462124 A CN111462124 A CN 111462124A
Authority
CN
China
Prior art keywords
convolution
deep
size
feature maps
abv3
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN202010241130.XA
Other languages
Chinese (zh)
Inventor
黄焱
杜飞飞
倪彦朝
姜炳强
王天玮
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing Mechanical And Electrical Engineering General Design Department
Wuhan Zmvision Technology Co ltd
Original Assignee
Beijing Mechanical And Electrical Engineering General Design Department
Wuhan Zmvision Technology Co ltd
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing Mechanical And Electrical Engineering General Design Department, Wuhan Zmvision Technology Co ltd filed Critical Beijing Mechanical And Electrical Engineering General Design Department
Priority to CN202010241130.XA priority Critical patent/CN111462124A/en
Publication of CN111462124A publication Critical patent/CN111462124A/en
Pending legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T7/00Image analysis
    • G06T7/10Segmentation; Edge detection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T3/00Geometric image transformations in the plane of the image
    • G06T3/40Scaling of whole images or parts thereof, e.g. expanding or contracting
    • G06T3/4007Scaling of whole images or parts thereof, e.g. expanding or contracting based on interpolation, e.g. bilinear interpolation
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T5/00Image enhancement or restoration
    • G06T5/20Image enhancement or restoration using local operators
    • G06T5/30Erosion or dilatation, e.g. thinning
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06TIMAGE DATA PROCESSING OR GENERATION, IN GENERAL
    • G06T2207/00Indexing scheme for image analysis or image enhancement
    • G06T2207/30Subject of image; Context of image processing
    • G06T2207/30181Earth observation
    • G06T2207/30192Weather; Meteorology

Landscapes

  • Engineering & Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Theoretical Computer Science (AREA)
  • General Physics & Mathematics (AREA)
  • General Health & Medical Sciences (AREA)
  • General Engineering & Computer Science (AREA)
  • Biophysics (AREA)
  • Computational Linguistics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Artificial Intelligence (AREA)
  • Molecular Biology (AREA)
  • Computing Systems (AREA)
  • Biomedical Technology (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Mathematical Physics (AREA)
  • Software Systems (AREA)
  • Health & Medical Sciences (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Image Analysis (AREA)
  • Image Processing (AREA)

Abstract

A remote sensing satellite cloud detection method of Deep L abV3+ includes the steps of S1, obtaining satellite load simulation images of different scenes, S2, inputting image data into a Deep convolution neural network to obtain semantic information feature maps a and semantic information feature maps b of different levels, S3, directly transmitting the semantic information feature maps a into a decoding area, enabling the semantic information feature maps b to be subjected to parallel cavity convolution, S4, obtaining output feature maps b of a coding process through the feature maps of the levels after the parallel cavity convolution in S3, S5, changing the number of channels of the bottom layer semantic feature maps a obtained in the coding process, enlarging the feature maps b in the coding process to be in the size, S6, splicing the feature maps a and the feature maps b obtained in S5 to extract features again, S7, conducting amplification and up-sampling on the feature maps obtained in S6 again to obtain segmented final feature maps, and solving the problem that the segmentation effect of the segmented images is poor in the image decoding-coding process of the existing CNN image segmentation cloud detection method.

Description

Remote sensing satellite cloud detection method based on Deep L abV3+
Technical Field
The invention relates to the field of Deep learning and computer vision application semantic segmentation, in particular to a remote sensing satellite cloud detection method based on Deep L abV3 +.
Background
Semantic segmentation and tracking is an important research direction in computer vision, and has extremely wide application in a plurality of fields such as land satellite image division, unmanned vehicle automatic driving, unmanned vehicle aerial image entity identification, face identification, medical image and the like, however, the semantic segmentation faces a plurality of challenges such as non-uniform image object picture, different image shooting angles, multi-posture or multi-view of an object, external illumination change, complex image semantic information and the like, so that the reliable and accurate image segmentation method has very important practical significance.
The conventional segmentation method is mainly based on digital image processing, topology, mathematics and other methods to perform image segmentation, such as a threshold-based segmentation method, and the basic idea of the threshold method is to calculate one or more gray threshold values based on gray features of an image, compare the gray value of each pixel in the image with the threshold values, and finally classify the pixels into proper classes according to the comparison results. Therefore, the most critical step of the method is to solve the optimal gray threshold according to some criterion function. The thresholding method is particularly suitable for maps where the object and the background occupy different gray level ranges. If the image only has two categories of a target and a background, only one threshold value needs to be selected for segmentation, and the method becomes single-threshold segmentation; however, if there are multiple objects in the image to be extracted, the segmentation with a single threshold value will be wrong, in this case, multiple threshold values need to be selected to separate each object, and this segmentation method is correspondingly called multi-threshold segmentation. The traditional segmentation method mainly has the characteristics of simple calculation and high efficiency, but only considers the characteristics of the gray value of a pixel point, and does not generally consider the spatial characteristics, so the traditional segmentation method is sensitive to noise and has low robustness.
In recent years, with the increase of computer power and the continuous development of deep learning, some conventional segmentation methods have been unable to compare in effect with the segmentation methods based on deep learning. Convolutional neural networks have a remarkable effect in the field of image segmentation, such as the well-known FCN full convolutional neural network, where the FCN replaces the fully-connected network in the conventional CNN with a convolutional network, so that the network can accept pictures of any size and output segmentation maps of the same size as the original pictures. Only then can a classification be made for each pixel. Meanwhile, in order to solve the influence of convolution and pooling on the image size, a method for recovering the image size by using an upsampling mode is provided. Even though the FCN solves the problems of any image input size, local feature extraction, and the like, in the image decoding-encoding process, the conventional CNN image segmentation method has the problems that the image segmentation result is poor in effect under the conditions that the internal data structure is lost, the semantic bottom layer information is lost, the small object image information cannot be reconstructed, and the like, due to the problems that the learning cannot be performed, the spatial hierarchy information is lost, and the small object information cannot be reconstructed in the sampling and pooling layer.
Disclosure of Invention
In view of the above, the present invention has been made to provide a method for remote sensing satellite cloud detection based on Deep L abV3+, which overcomes or at least partially solves the above problems.
The technical scheme provided by the invention is as follows:
a remote sensing satellite cloud detection method of Deep L abV3+, the method comprises the following steps:
s1, acquiring satellite load simulation images of different scenes, reading images of a multi-dimensional channel, normalizing the images, and converting the data format of the images;
s2, inputting the image data into a deep convolution neural network, setting the size of a convolution kernel by using serial hole convolution, and obtaining semantic information characteristic graphs a and b of different levels according to the size of the convolution kernel;
s3, directly transmitting the semantic information feature graph a into a decoding area, setting the size of cavity convolution expansion through parallel cavity convolution on the semantic information feature graph b, extracting features, combining the feature, performing 1x1 convolution compression features, and finally adding an average pooling layer to obtain 5 level feature graphs;
s4, splicing the feature maps of all levels after the parallel hole convolution in the S3 on a channel dimension to obtain an output feature map b in the encoding process;
s5, carrying out convolution on the bottom-layer semantic feature map a obtained in the encoding process by 1x1, changing the number of channels, outputting a feature map b in the encoding process, carrying out up-sampling by adopting a bilinear interpolation method, and enlarging the size of a picture;
s6, splicing the characteristic graph a and the characteristic graph b obtained in the step S5, performing convolution with a convolution kernel of 3x3, and re-extracting characteristics;
and S7, performing up-sampling on the feature map obtained in the step S6 in an enlarged size again to obtain a final feature map after segmentation.
Further, in S1, the input image is processed with a keras function using a keras frame to convert the image into a tensor format.
Further, in S2, the convolution kernels have sizes of 32, 64, 128, and 256, respectively, and 256x256x32, 128x128x64, 64x64x128, and 32x32x256 semantic information feature maps of different levels are sequentially obtained.
Further, the size of the semantic information feature map a is 128x128x64, and the size of the semantic information feature map b is 32x32x 256.
Further, in S3, the hole convolution expansion size is set to [1, 6, 12, 18], the pooling layer is a depth separable convolution layer, wherein the step size is 2;
further, in S5, the bilinear interpolation method includes:
the known function f is in Q11=(x1,y1),Q12=(x1,y2),Q21(x2, y1), and Q22The pixel values of four points (x2, y2) are required to have a size of (x, y) point,
and performing linear interpolation in the x direction of the coordinate axis to obtain:
Figure BDA0002432610650000031
Figure BDA0002432610650000032
and performing linear interpolation in the y direction of the coordinate axis to obtain:
Figure BDA0002432610650000033
the size of the pixel value at the point P ═ x, y is obtained.
Further, a loss function in the neural network model training process adopts a binary _ cross entropy loss function, and a calculation formula is as follows:
Figure BDA0002432610650000034
wherein i represents the position of the ith point, y'iThe predicted pixel value, y, representing the ith pointiRepresenting the real pixel value of the ith point.
Further, Adam function is adopted as an optimization function to reduce the size of the loss function value.
Further, the intermediate convolutional layer of the Deeplab V3 model was repeated 16 times, followed by the batch normalization and Relu activation functions.
Further, the neural network model is evaluated by adopting an image segmentation evaluation index Miou. Wherein, the calculation formula is as follows:
Figure BDA0002432610650000041
FN is false negative, which means that the reality is 1 and the prediction is 0;
TP is true, which means that the reality is 1, and the prediction is 1;
FP is false positive, indicating a true 0, predicted to be 1.
Compared with the prior art, the invention at least has the following beneficial effects:
the remote sensing satellite cloud detection method based on Deep L abV3+ provided by the invention introduces a hollow convolution and hollow space convolution pooling pyramid structure, increases the sense field of convolution under the condition of no information loss, and enables each convolution output to contain information in a larger range, so that the image segmentation result is more accurate and accords with space semantic logic, and the cloud detection direction has better cloud segmentation effect than the traditional cloud segmentation direction.
Drawings
The accompanying drawings, which are included to provide a further understanding of the invention and are incorporated in and constitute a part of this specification, illustrate embodiments of the invention and together with the description serve to explain the principles of the invention and not to limit the invention. In the drawings:
fig. 1 is a flowchart of a remote sensing satellite cloud detection method based on Deep L abV3+, in an embodiment of the present invention;
fig. 2 is a Deep neural network structure diagram based on Deep L abv3+ in the second embodiment of the invention.
Detailed Description
Example 1
A remote sensing satellite cloud detection method of Deep L abV3+, the method comprises the following steps:
s1, acquiring satellite load simulation images of different scenes, reading images of a multi-dimensional channel, normalizing the images, and converting the data format of the images;
specifically, an OPENCV function library is adopted to read an image of a multi-dimensional channel, a keras framework is adopted after the image is normalized, a keras framework is adopted, an input image is processed by a keras.
In some preferred embodiments, 15530 satellite load simulation data are acquired, 530 invalid data are deleted, the remaining 15000 valid data are divided according to 7:2:1, 10500 data are used for model training, 3000 data are used for model verification, and 1500 data are used for model testing, as shown in table one below:
watch 1
Training data Verification data Test data
10500(70%) 3000(20%) 1500(10%)
Aiming at the cloud layer characteristics of different scenes, the invention respectively trains 4 types of models, namely a full-landform all-day-time infrared model, a marine daytime model, a land daytime model and a polar all-day-time infrared model. To simplify the model, the model is a two-class model, i.e. the truth mask is the dichotomy result: "cloud" and "not cloud".
The satellite cloud data includes 19 visible and near-infrared channels, 6 mid-infrared and far-infrared channels, for a total of 25 channels. The different data channels used by each model are shown in table two below:
watch two
Model name Data channel
All-landform all-time infrared model 6 infrared band data, the channel number is 20 ~ 25
Ocean daytime model 6 infrared band data, the channel number is 20 ~ 25
Land day model Visible light and infrared, the channel numbers are 1-5, 7, 16-25
Polar region all-time infrared model 6 infrared band data, the channel number is 20 ~ 25
Different models aim at different scenes, and the application scene effect has better accuracy than the training of a full landform model.
S2, inputting the image data into a deep convolution neural network, setting the size of a convolution kernel by using serial hole convolution, and obtaining semantic information characteristic graphs a and b of different levels according to the size of the convolution kernel;
specifically, the format of the input image in S2 is hxwxc (height width xchanchannel), and taking the image size as 512x512x25 as an example, the input image is subjected to a DCNN deep convolution neural network, and is convolved by using serial AtrousConvolution holes, the sizes of convolution kernels are 32, 64, 128 and 256, so as to sequentially obtain semantic information feature maps of 256x256x32, 128x128x64, 64x64x128 and 32x32x256 at different levels;
s3, directly transmitting the semantic information feature graph a into a decoding area, setting the size of cavity convolution expansion through parallel cavity convolution on the semantic information feature graph b, extracting features, combining the feature, performing 1x1 convolution compression features, and finally adding an average pooling layer to obtain 5 level feature graphs;
specifically, the semantic information feature map a is 128x128x64, the semantic information feature map b is 32x32x256, and the feature map is directly transmitted into a decoding area; and (b) performing parallel cavity convolution on the feature graph b, performing feature extraction by respectively using cavity convolutions with different cavity convolution expansion sizes, merging, and performing 1x1 convolution compression features. The convolution expansion sizes of the parallel AtrousConvolition holes are [1, 6, 12 and 18], and finally an average pooling layer is added, wherein the pooling layer is a depth separable convolution layer, the step length is 2, and 5 level characteristic graphs are obtained;
s4, splicing the feature maps of all levels after the parallel hole convolution in the S3 on a channel dimension to obtain an output feature map b in the encoding process;
in some preferred embodiments, the output signature b of the encoding process is obtained in a size of 32x32x 1280.
S5, carrying out convolution on the bottom-layer semantic feature map a obtained in the encoding process by 1x1, changing the number of channels, outputting a feature map b in the encoding process, carrying out up-sampling by adopting a bilinear interpolation method, and enlarging the size of a picture;
in some preferred embodiments, the underlying semantic feature map b is subjected to 1x1 convolution, the number of channels is changed, a feature map with a size of 32x32x256 is obtained, then, bilinear interpolation is adopted for upsampling, and the amplification size is 4 times that of the original feature map, so that a feature map with a size of 128x128x256 is obtained.
Specifically, the bilinear interpolation method includes:
the known function f is in Q11=(x1,y1),Q12=(x1,y2),Q21(x2, y1), and Q22The pixel values of four points (x2, y2) are required to have a size of (x, y) point,
and performing linear interpolation in the x direction of the coordinate axis to obtain:
Figure BDA0002432610650000071
Figure BDA0002432610650000072
and performing linear interpolation in the y direction of the coordinate axis to obtain:
Figure BDA0002432610650000073
the size of the pixel value at the point P ═ x, y is obtained.
S6, splicing the characteristic graph a and the characteristic graph b obtained in the step S5, performing convolution with a convolution kernel of 3x3, and re-extracting characteristics;
in some preferred embodiments, the feature map a and the feature map b obtained in S5 are spliced to obtain a 128x128x320 feature map, and then the feature is re-extracted through convolution kernel convolution of 3x3 to obtain a feature map of 128x128x256 size;
and S7, performing up-sampling on the feature map obtained in the step S6 in an enlarged size again to obtain a final feature map after segmentation.
In some preferred embodiments, the feature map obtained in S6 is subjected to up-sampling of 4 times size again, and dimension reduction is performed on the convolutional layer with the size of 1x1 and the size of the convolutional kernel of 1, so as to obtain a feature map of 512x512x1, which is the same as the size of the original map, wherein the feature map channel dimension 1 is a grayscale map, and the segmentation result is the final prediction result.
The real cloud detection label map is a gray scale map, the image size is 512x512x1, that is, the value set of the pixel values is {0,1 }. 0 represents that the point is a non-cloud part, and 1 represents that the point is a cloud.
In some preferred embodiments, the loss function in the neural network model training process adopts a binary _ cross entropy loss function, and the calculation formula is as follows:
Figure BDA0002432610650000074
where i represents the position of the ith point, yiiThe predicted pixel value, y, representing the ith pointiRepresenting the real pixel value of the ith point.
In some preferred embodiments, Adam's function is used as the optimization function to reduce the magnitude of the loss function value.
In some preferred embodiments, the intermediate convolutional layer of the Deeplab V3 model is repeated 16 times, followed by the batch normalization and Relu activation functions.
In some preferred embodiments, the neural network model is evaluated using an image segmentation evaluation index Miou. Wherein, the calculation formula is as follows:
Figure BDA0002432610650000081
FN is false negative, which means that the reality is 1 and the prediction is 0;
TP is true, which means that the reality is 1, and the prediction is 1;
FP is false positive, indicating a true 0, predicted to be 1.
The remote sensing satellite cloud detection method based on Deep L abV3+ provided by the invention introduces a hollow convolution and hollow space convolution pooling pyramid structure, increases the sense field of convolution under the condition of no information loss, and enables each convolution output to contain information in a larger range, so that the image segmentation result is more accurate and accords with space semantic logic, and the cloud detection direction has better cloud segmentation effect than the traditional cloud segmentation direction.
It should be understood that the specific order or hierarchy of steps in the processes disclosed is an example of exemplary approaches. Based upon design preferences, it is understood that the specific order or hierarchy of steps in the processes may be rearranged without departing from the scope of the present disclosure. The accompanying method claims present elements of the various steps in a sample order, and are not intended to be limited to the specific order or hierarchy presented.
In the foregoing detailed description, various features are grouped together in a single embodiment for the purpose of streamlining the disclosure. This method of disclosure is not to be interpreted as reflecting an intention that the claimed embodiments of the subject matter require more features than are expressly recited in each claim. Rather, as the following claims reflect, invention lies in less than all features of a single disclosed embodiment. Thus, the following claims are hereby expressly incorporated into the detailed description, with each claim standing on its own as a separate preferred embodiment of the invention.
Those of skill would further appreciate that the various illustrative logical blocks, modules, circuits, and algorithm steps described in connection with the embodiments disclosed herein may be implemented as electronic hardware, computer software, or combinations of both. To clearly illustrate this interchangeability of hardware and software, various illustrative components, blocks, modules, circuits, and steps have been described above generally in terms of their functionality. Whether such functionality is implemented as hardware or software depends upon the particular application and design constraints imposed on the overall system. Skilled artisans may implement the described functionality in varying ways for each particular application, but such implementation decisions should not be interpreted as causing a departure from the scope of the present disclosure.
The steps of a method or algorithm described in connection with the embodiments disclosed herein may be embodied directly in hardware, in a software module executed by a processor, or in a combination of the two. A software module may reside in RAM memory, flash memory, ROM memory, EPROM memory, EEPROM memory, registers, hard disk, a removable disk, a CD-ROM, or any other form of storage medium known in the art. An exemplary storage medium is coupled to the processor such the processor can read information from, and write information to, the storage medium. Of course, the storage medium may also be integral to the processor. The processor and the storage medium may reside in an ASIC. The ASIC may reside in a user terminal. Of course, the processor and the storage medium may reside as discrete components in a user terminal.
For a software implementation, the techniques described herein may be implemented with modules (e.g., procedures, functions, and so on) that perform the functions described herein. The software codes may be stored in memory units and executed by processors. The memory unit may be implemented within the processor or external to the processor, in which case it can be communicatively coupled to the processor via various means as is known in the art.
What has been described above includes examples of one or more embodiments. It is, of course, not possible to describe every conceivable combination of components or methodologies for purposes of describing the aforementioned embodiments, but one of ordinary skill in the art may recognize that many further combinations and permutations of various embodiments are possible. Accordingly, the embodiments described herein are intended to embrace all such alterations, modifications and variations that fall within the scope of the appended claims. Furthermore, to the extent that the term "includes" is used in either the detailed description or the claims, such term is intended to be inclusive in a manner similar to the term "comprising" as "comprising" is interpreted when employed as a transitional word in a claim. Furthermore, any use of the term "or" in the specification of the claims is intended to mean a "non-exclusive or".

Claims (10)

1. A remote sensing satellite cloud detection method of Deep L abV3+, which is characterized by comprising the following steps:
s1, acquiring satellite load simulation images of different scenes, reading images of a multi-dimensional channel, normalizing the images, and converting the data format of the images;
s2, inputting the image data into a deep convolution neural network, setting the size of a convolution kernel by using serial hole convolution, and obtaining semantic information characteristic graphs a and b of different levels according to the size of the convolution kernel;
s3, directly transmitting the semantic information feature graph a into a decoding area, setting the size of cavity convolution expansion through parallel cavity convolution on the semantic information feature graph b, extracting features, combining the feature, performing 1x1 convolution compression features, and finally adding an average pooling layer to obtain 5 level feature graphs;
s4, splicing the feature maps of all levels after the parallel hole convolution in the S3 on a channel dimension to obtain an output feature map b in the encoding process;
s5, carrying out convolution on the bottom-layer semantic feature map a obtained in the encoding process by 1x1, changing the number of channels, outputting a feature map b in the encoding process, carrying out up-sampling by adopting a bilinear interpolation method, and enlarging the size of a picture;
s6, splicing the characteristic graph a and the characteristic graph b obtained in the step S5, performing convolution with a convolution kernel of 3x3, and re-extracting characteristics;
and S7, performing up-sampling on the feature map obtained in the step S6 in an enlarged size again to obtain a final feature map after segmentation.
2. The method for remotely sensing the satellite cloud by Deep L abV3+ as claimed in claim 1, wherein in S1, a keras frame is adopted, and a keras.
3. The method for remotely sensing satellite cloud detection through Deep L abV3+ as claimed in claim 1, wherein in S2, the sizes of convolution kernels are 32, 64, 128 and 256 respectively, and semantic information feature maps of 256x256x32, 128x128x64, 64x64x128 and 32x32x256 at different levels are sequentially obtained.
4. The method for remotely sensing satellite cloud by Deep L abV3+, according to claim 3, wherein the size of semantic information feature map a is 128x128x64, and the size of semantic information feature map b is 32x32x 256.
5. The method for remotely sensing the satellite cloud by Deep L abV3+, according to claim 1, wherein in S3, the size of the hole convolution expansion is set to [1, 6, 12, 18], the pooling layer is a depth separable convolution layer, and the step size is 2.
6. The remote sensing satellite cloud detection method of Deep L abV3+, according to claim 1, wherein in S5, the bilinear interpolation method is:
the known function f is in Q11=(x1,y1),Q12=(x1,y2),Q21(x2, y1), and Q22The pixel values of four points (x2, y2) are required to have a size of (x, y) point,
and performing linear interpolation in the x direction of the coordinate axis to obtain:
Figure FDA0002432610640000021
Figure FDA0002432610640000022
and performing linear interpolation in the y direction of the coordinate axis to obtain:
Figure FDA0002432610640000023
the size of the pixel value at the point P ═ x, y is obtained.
7. The method for remotely sensing the satellite cloud by Deep L abV3+, according to claim 1, wherein a loss function in a neural network model training process is a binary _ cross entropy loss function, and a calculation formula is as follows:
Figure FDA0002432610640000024
wherein i represents the position of the ith point, y'iThe predicted pixel value, y, representing the ith pointiRepresenting the real pixel value of the ith point.
8. The method for remotely sensing the satellite cloud through Deep L abV3+, according to claim 1, wherein an Adam function is adopted as an optimization function to reduce the size of the loss function value.
9. The method for remotely sensing the cloud of Deep L abV3+ as recited in claim 1, wherein the number of times of repetition of the intermediate convolutional layer of the Deep v3 model is 16, and the convolutional layer is followed by batch normalization and Relu activation functions.
10. The method for remotely sensing satellite cloud detection through Deep L abV3+ as claimed in claim 1, wherein the evaluation of the neural network model is performed by using an image segmentation evaluation index Miou, wherein the calculation formula is as follows:
Figure FDA0002432610640000031
FN is false negative, which means that the reality is 1 and the prediction is 0;
TP is true, which means that the reality is 1, and the prediction is 1;
FP is false positive, indicating a true 0, predicted to be 1.
CN202010241130.XA 2020-03-31 2020-03-31 Remote sensing satellite cloud detection method based on Deep L abV3+ Pending CN111462124A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202010241130.XA CN111462124A (en) 2020-03-31 2020-03-31 Remote sensing satellite cloud detection method based on Deep L abV3+

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202010241130.XA CN111462124A (en) 2020-03-31 2020-03-31 Remote sensing satellite cloud detection method based on Deep L abV3+

Publications (1)

Publication Number Publication Date
CN111462124A true CN111462124A (en) 2020-07-28

Family

ID=71685093

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202010241130.XA Pending CN111462124A (en) 2020-03-31 2020-03-31 Remote sensing satellite cloud detection method based on Deep L abV3+

Country Status (1)

Country Link
CN (1) CN111462124A (en)

Cited By (9)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112258431A (en) * 2020-09-27 2021-01-22 成都东方天呈智能科技有限公司 Image classification model based on mixed depth separable expansion convolution and classification method thereof
CN112347976A (en) * 2020-11-23 2021-02-09 腾讯科技(深圳)有限公司 Region extraction method and device for remote sensing satellite image, electronic equipment and medium
CN112487184A (en) * 2020-11-26 2021-03-12 北京智源人工智能研究院 User character judging method and device, memory and electronic equipment
CN112489054A (en) * 2020-11-27 2021-03-12 中北大学 Remote sensing image semantic segmentation method based on deep learning
CN113361662A (en) * 2021-07-22 2021-09-07 全图通位置网络有限公司 System and method for processing remote sensing image data of urban rail transit
CN113724260A (en) * 2021-08-03 2021-11-30 南京邮电大学 Satellite capturing method based on deep reinforcement learning
CN114549535A (en) * 2022-01-28 2022-05-27 北京百度网讯科技有限公司 Image segmentation method, device, equipment, storage medium and product
CN114996488A (en) * 2022-08-08 2022-09-02 北京道达天际科技股份有限公司 Skynet big data decision-level fusion method
CN117237644A (en) * 2023-11-10 2023-12-15 广东工业大学 Forest residual fire detection method and system based on infrared small target detection

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110097544A (en) * 2019-04-25 2019-08-06 武汉精立电子技术有限公司 A kind of display panel open defect detection method
CN110781775A (en) * 2019-10-10 2020-02-11 武汉大学 Remote sensing image water body information accurate segmentation method supported by multi-scale features

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN110097544A (en) * 2019-04-25 2019-08-06 武汉精立电子技术有限公司 A kind of display panel open defect detection method
CN110781775A (en) * 2019-10-10 2020-02-11 武汉大学 Remote sensing image water body information accurate segmentation method supported by multi-scale features

Non-Patent Citations (6)

* Cited by examiner, † Cited by third party
Title
LIANG-CHIEH CHEN等: "Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation", pages 1 - 4 *
张永宏;蔡朋艳;陶润喆;王剑庚;田伟;: "基于改进U-Net网络的遥感图像云检测", no. 03 *
张永宏等: "基于改进U-Net 网络的遥感图像云检测", no. 03, pages 18 *
张鑫禄;张崇涛;戴晨光;季虹良;王映雪;: "基于DeepLabv3架构的高分辨率遥感图像分类", no. 02, pages 40 - 100 *
桑红石等: "《基于深度学习的图像语义分割技术》", 武汉:华中科技大学出版社, pages: 63 - 100 *
裴亮;刘阳;高琳;: "结合全卷积神经网络与条件随机场的资源3号遥感影像云检测", no. 10, pages 269 - 275 *

Cited By (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN112258431A (en) * 2020-09-27 2021-01-22 成都东方天呈智能科技有限公司 Image classification model based on mixed depth separable expansion convolution and classification method thereof
CN112258431B (en) * 2020-09-27 2021-07-20 成都东方天呈智能科技有限公司 Image classification model based on mixed depth separable expansion convolution and classification method thereof
CN112347976A (en) * 2020-11-23 2021-02-09 腾讯科技(深圳)有限公司 Region extraction method and device for remote sensing satellite image, electronic equipment and medium
CN112487184A (en) * 2020-11-26 2021-03-12 北京智源人工智能研究院 User character judging method and device, memory and electronic equipment
CN112489054A (en) * 2020-11-27 2021-03-12 中北大学 Remote sensing image semantic segmentation method based on deep learning
CN113361662B (en) * 2021-07-22 2023-08-29 全图通位置网络有限公司 Urban rail transit remote sensing image data processing system and method
CN113361662A (en) * 2021-07-22 2021-09-07 全图通位置网络有限公司 System and method for processing remote sensing image data of urban rail transit
CN113724260A (en) * 2021-08-03 2021-11-30 南京邮电大学 Satellite capturing method based on deep reinforcement learning
CN113724260B (en) * 2021-08-03 2023-10-17 南京邮电大学 Satellite grabbing method based on deep reinforcement learning
CN114549535A (en) * 2022-01-28 2022-05-27 北京百度网讯科技有限公司 Image segmentation method, device, equipment, storage medium and product
CN114996488A (en) * 2022-08-08 2022-09-02 北京道达天际科技股份有限公司 Skynet big data decision-level fusion method
CN117237644A (en) * 2023-11-10 2023-12-15 广东工业大学 Forest residual fire detection method and system based on infrared small target detection
CN117237644B (en) * 2023-11-10 2024-02-13 广东工业大学 Forest residual fire detection method and system based on infrared small target detection

Similar Documents

Publication Publication Date Title
CN111462124A (en) Remote sensing satellite cloud detection method based on Deep L abV3+
CN112446270B (en) Training method of pedestrian re-recognition network, pedestrian re-recognition method and device
CN113033570B (en) Image semantic segmentation method for improving void convolution and multilevel characteristic information fusion
CN110781756A (en) Urban road extraction method and device based on remote sensing image
CN113743269B (en) Method for recognizing human body gesture of video in lightweight manner
CN115661144A (en) Self-adaptive medical image segmentation method based on deformable U-Net
CN117079139B (en) Remote sensing image target detection method and system based on multi-scale semantic features
CN107766864B (en) Method and device for extracting features and method and device for object recognition
CN112836625A (en) Face living body detection method and device and electronic equipment
CN113326930A (en) Data processing method, neural network training method, related device and equipment
CN112364873A (en) Character recognition method and device for curved text image and computer equipment
CN116453121B (en) Training method and device for lane line recognition model
CN116469100A (en) Dual-band image semantic segmentation method based on Transformer
CN116310850B (en) Remote sensing image target detection method based on improved RetinaNet
CN115578615A (en) Night traffic sign image detection model establishing method based on deep learning
CN115410030A (en) Target detection method, target detection device, computer equipment and storage medium
CN114359297A (en) Attention pyramid-based multi-resolution semantic segmentation method and device
CN111524232A (en) Three-dimensional modeling method and device and server
CN110969164A (en) Low-illumination imaging license plate recognition method and device based on deep learning end-to-end
CN114299383A (en) Remote sensing image target detection method based on integration of density map and attention mechanism
CN113762039A (en) Information matching method and related device for traffic sign board
CN116863194A (en) Foot ulcer image classification method, system, equipment and medium
CN116645598A (en) Remote sensing image semantic segmentation method based on channel attention feature fusion
CN116524189A (en) High-resolution remote sensing image semantic segmentation method based on coding and decoding indexing edge characterization
CN115861756A (en) Earth background small target identification method based on cascade combination network

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
CB03 Change of inventor or designer information
CB03 Change of inventor or designer information

Inventor after: Huang Yan

Inventor after: Du Feifei

Inventor after: Ni Yanchao

Inventor after: Jiang Bingqiang

Inventor after: Wang Tianwei

Inventor before: Huang Yan

Inventor before: Du Feifei

Inventor before: Ni Yanchao

Inventor before: Jiang Bingqiang

Inventor before: Wang Tianwei