CN114037896B - PolSAR ground object fine classification method based on multi-index convolution self-encoder - Google Patents

PolSAR ground object fine classification method based on multi-index convolution self-encoder Download PDF

Info

Publication number
CN114037896B
CN114037896B CN202111319995.4A CN202111319995A CN114037896B CN 114037896 B CN114037896 B CN 114037896B CN 202111319995 A CN202111319995 A CN 202111319995A CN 114037896 B CN114037896 B CN 114037896B
Authority
CN
China
Prior art keywords
encoder
polsar
index
matrix
polarization
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN202111319995.4A
Other languages
Chinese (zh)
Other versions
CN114037896A (en
Inventor
艾加秋
黄默
王非凡
范高伟
裴志林
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Hefei University of Technology
Original Assignee
Hefei University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Hefei University of Technology filed Critical Hefei University of Technology
Priority to CN202111319995.4A priority Critical patent/CN114037896B/en
Publication of CN114037896A publication Critical patent/CN114037896A/en
Application granted granted Critical
Publication of CN114037896B publication Critical patent/CN114037896B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2415Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on parametric or probabilistic models, e.g. based on likelihood ratio or false acceptance rate versus a false rejection rate
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/04Architecture, e.g. interconnection topology
    • G06N3/045Combinations of networks
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06NCOMPUTING ARRANGEMENTS BASED ON SPECIFIC COMPUTATIONAL MODELS
    • G06N3/00Computing arrangements based on biological models
    • G06N3/02Neural networks
    • G06N3/08Learning methods
    • G06N3/084Backpropagation, e.g. using gradient descent

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • Evolutionary Computation (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Artificial Intelligence (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Computing Systems (AREA)
  • Software Systems (AREA)
  • Molecular Biology (AREA)
  • Computational Linguistics (AREA)
  • Biophysics (AREA)
  • Biomedical Technology (AREA)
  • Mathematical Physics (AREA)
  • General Health & Medical Sciences (AREA)
  • Health & Medical Sciences (AREA)
  • Evolutionary Biology (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Probability & Statistics with Applications (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a PolSAR ground object fine classification method based on a multi-index convolution self-encoder, which comprises the following steps: 1. preprocessing and slicing the PolSAR data; 2. extracting a feature map of the PolSAR image through convolution and pooling operation, inputting the obtained feature map into an MI-SE module to calculate multiple indexes, and obtaining the weight of each dimension of the input feature map; 3. weighting the importance of the feature map according to the weight distributed to the input feature map to obtain a final feature; 4. inputting the final characteristics into a classifier to obtain a predicted result, and comparing the predicted result with a real result to complete the training process of the model; 5. and inputting the PolSAR image to be classified into a trained multi-index convolution self-encoder model to obtain a classification result. The method can improve the integrality and the fineness of the feature representation of the PolSAR, obtain higher classification precision and classification efficiency, and have better engineering application value.

Description

PolSAR ground object fine classification method based on multi-index convolution self-encoder
Technical Field
The invention relates to the technical field of PolSAR ground object fine classification, in particular to a PolSAR ground object fine classification method based on a multi-index convolution self-encoder.
Background
Polarized synthetic aperture radar (Polarimetric synthetic aperture radar, polSAR) as a high resolution imaging sensor can provide detailed ground object information. Compared with an optical image, the PolSAR can obtain more effective and comprehensive ground feature information under all-weather and full-time conditions. Therefore, classification of ground features by using PolSAR data can effectively improve classification accuracy.
The conventional PolSAR terrain classification method can be summarized into an unsupervised classification without training sample tags and a supervised classification with training sample tags. Although a lot of human resources are not consumed in the non-supervision classification process, the operation time is long, and the classification precision is low. The supervised classification based on machine learning mainly comprises two processes of feature extraction and classification, the performance of which is mainly limited by the representation capability of the features, and the classification performance is extremely reduced when the extracted features are difficult to highlight the differences among different feature classes. Supervised classification algorithms based on deep learning have been proposed in recent years, but as the complexity of deep learning networks increases, the training process generally requires a large number of samples and a sufficient amount of time. However the problem of sample lack of labels is just a short plate in the PolSAR clutter classification problem. To address this problem, convolutional self-encoders (convolutional auto-encoders, CAE) are of greater interest because CAE can train network parameters through the decoder under unsupervised conditions, thereby reducing the amount of samples required for final classification. However, because the network depth is low, the extracted depth features cannot fully represent the information of the target and are also easily affected by the interference information, so that the classification result is poor.
There are many existing methods for improving the problems of the CAE network, which improve the conventional CAE network in terms of sample data size, network depth, loss function, etc. For example: stacked convolutional self-encoders (stacked convolutional Autoencoder, SCAE) deepen the entire network by stacking multiple CAE networks and learn network parameters using hierarchical training to obtain deeper ground feature information; there is also an algorithm that uses the wishart distance as the reconstruction loss of the CAE network. However, these improved algorithms merely stack adjacent features into different feature map channels, which is not complete and fine enough for different classes of feature representation.
Disclosure of Invention
The invention aims to solve the defects in the prior art, and provides a PolSAR ground object fine classification method based on a multi-index convolution self-encoder, so that the completeness and the fineness of the feature characterization of the PolSAR ground object can be improved, and higher classification precision and classification efficiency are obtained.
In order to achieve the aim of the invention, the invention adopts the following technical scheme:
the invention relates to a PolSAR ground object fine classification method based on a multi-index convolution self-encoder, which is characterized by comprising the following steps:
step 1: preprocessing and clipping operation of PolSAR data:
step 1.1: the scattering characteristic of any minimum resolution unit in the PolSAR data is represented by using a polarized scattering matrix S; according to the reciprocity principle and Pauli decomposition principle, converting any one polarized scattering matrix S in PolSAR data into a vectorWherein S is HH Representing the complex scattering coefficient when the transmitting polarization mode is horizontal polarization and the receiving polarization mode is horizontal polarization, S VV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is vertical polarization, S HV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is horizontal polarization;
step 1.2: based on the loss K, a polarization coherence matrix T is obtained using equation (1):
in the formula (1) (. Cndot. H Represents conjugate transpose, T ij Elements representing the ith row and jth column of the polarization coherence matrix T; and i, j=1, 2,3;
step 1.3: extracting a six-dimensional eigenvector f= [ a, B, C, D, E, F ] from the polarization coherence matrix T using formula (2):
in equation (2), span represents the total scattered power of all polarized channels, and span=t 11 +T 22 +T 33 A represents the decibel form of the total scattered power Span, B and C represent the 2 nd row and 2 nd column elements T, respectively 22 And line 3, column 3 element T 33 D, E and F represent three relative correlation coefficients;
step 1.4: normalizing the six-dimensional feature vector F to obtain a normalized six-dimensional feature vector F', so that preprocessed PolSAR data are formed by the normalized six-dimensional feature vectors of all minimum resolution units;
step 1.5: cutting the preprocessed PolSAR data into slices with the size of L multiplied by 6, thereby obtaining a slice set { s } of the PolSAR data 1 ,...,s n ,...,s N -wherein s n Represents an nth slice, N represents the total number of slices; n is E [1, N];
Step 2: constructing a multi-index convolutional self-encoder based classification network, comprising: an encoder, a decoder, and a classifier; and aggregating the slices { s } 1 ,...,s n ,...,s N -input into the classification network;
step 2.1: the encoder consists of m convolution layers, m pooling layers, m MI-SE modules, m weighting modules and a full connection layer;
the nth slice s n Inputting the characteristic diagram U into the encoder, and obtaining a characteristic diagram U with a scale of H multiplied by W multiplied by D after the treatment of a first convolution layer and a pooling layer, wherein H and W respectively represent the height and the width of the characteristic diagram U, and D represents the channel number of the characteristic diagram U;
the feature map U is input into a first MI-SE module, and a multi-index is calculated by using a formula (3), and the method comprises the following steps: first moment S 1 Second order central moment S 2 And coefficient of variation S 3
In the formula (3), U (a, b, i) represents all elements of the a-th row and the b-th column of the characteristic diagram U;
the first moment S 1 Second order central moment S 2 And coefficient of variation S 3 Splicing the matrix into a multi-index matrix V with the size of 3 xD, and processing the multi-index matrix V by a convolution layer and a full connection layer in the first MI-SE module to obtain a weight matrix Y with the size of 1 xD;
the first weighting module multiplies the weight matrix Y with the feature map U to obtain a weighted feature map U' with the size of H multiplied by W multiplied by D;
the weighted feature map U' sequentially passes through the rest m-1 convolution layers, the pooling layer, the MI-SE module and the weighting module of the encoder, and finally the output features pass through a full connection layer to obtain the feature vector alpha output by the encoder;
step 2.2: the decoder is formed by sequentially connecting k convolution layers and k up-sampling layers in a staggered manner;
the feature vector alpha is input into a decoder to obtain an nth slice s n Is a reconstructed slice of (a)Re-use of (4) to establish a counter-propagating loss function L MSE
In equation (4), MSE represents the minimum mean square error; p is the number of elements per slice,for the nth slice s n The c-th element in the slice reconstructed after decoder,>for the nth slice s n C element of (a);
the encoder and decoder are trained using back propagation when the loss function L MSE When the minimum value is taken, training is stopped, and the optimal feature vector alpha output by the trained encoder is obtained *
Step 2.3: the classifier consists of t full-connection layers and one softmax layer;
the optimal feature vector alpha * In the input classifier, firstly, reducing the dimension through t full-connection layers, then inputting the feature vector after the dimension reduction into a softmax layer, and calculating an nth slice s n The center element of (a) corresponds to the posterior probability of each category, and the category corresponding to the maximum posterior probability is selected as the nth slice s n Is a predictive category of (2);
step 2.4: obtaining prediction categories of N slices according to the modes of the steps 2.1-2.3, comparing the prediction categories with corresponding real categories, and optimizing the classifier through back propagation, so that training of the encoder, the decoder and the classifier is completed, and a trained multi-index convolution self-encoder model is obtained and used for classifying PolSAR data to be classified.
Compared with the prior art, the invention has the beneficial effects that:
1. the PolSAR ground object fine classification method based on the multi-index convolution self-encoder effectively improves the integrity and the fineness of feature representation in the PolSAR ground object classification process, reduces the influence of individual extremum of images on the integral feature, and realizes the PolSAR ground object fine classification.
2. The invention provides the MI-SE module, solves the problem that in the traditional CAE network, the result of the feature map output by the convolution layer is simply and averagely input to the next layer, and performs effective weighting operation on the output of the convolution layer by utilizing the information in the feature map, thereby better extracting and characterizing the feature of the ground object. The characteristics carrying different information are weighted with different weights, and the more complete and fine characteristic information is obtained.
3. The invention provides three indexes for measuring the importance of different features, namely a first moment, a second central moment and a variation coefficient. The first moment can reflect the average value of the features, the second-order center distance can describe the fluctuation in each feature, the content of the edge and texture information can be reflected to a certain extent, and the variation coefficient can eliminate the influence of the average value on the dispersion, so that the fluctuation is reflected better. The three indexes complement each other, and each feature is weighted by different angles.
4. The convolution result is weighted through the three indexes, so that the difference in the ground object categories is obviously reduced, the difference between the ground object categories is increased, and in order to illustrate the capability of improving the separability between the different categories by the three different indexes, the t-distribution random neighborhood embedding algorithm is used for visualizing the middle characteristic diagram, and the result after 100 times of network iteration is shown in figure 3. So that the final softmax can more easily separate each category of ground objects.
Drawings
FIG. 1 is a MI-SE block structure of the invention;
FIG. 2 is a flow chart of the PolSAR image ground object target classification method of the present invention;
FIG. 3 is an illustration of a visual intermediate feature map of the present invention employing a t-distributed random neighborhood embedding algorithm;
FIG. 4 is an experimental data set of the invention San Francisco;
FIG. 5 is a graph of the classification results of the method of the present invention and other methods on the San Francisco dataset.
Detailed Description
In this embodiment, a multi-index compression and excitation (MI-SE) module shown in fig. 1 is added to a conventional CAE network, which proposes a multi-index convolutional self-encoder based classification network shown in fig. 2, so that the detail features and depth features of the PolSAR ground object target can be effectively considered, different attention degrees of different features can be realized when the model is guided to train, the completeness and fineness of feature characterization of the PolSAR ground object target are ensured, the gaps in various ground objects are reduced, and the gaps among various ground objects are increased, thereby obtaining a better classification result. Specifically, the PolSAR terrain fine classification method based on the multi-index convolution self-encoder comprises the following steps:
step 1: preprocessing and clipping operation of PolSAR data:
step 1.1: the scattering characteristic of any minimum resolution unit in the PolSAR data is represented by using a polarized scattering matrix S; according to the reciprocity principle and Pauli decomposition principle, converting any one polarized scattering matrix S in PolSAR data into a vectorWherein S is HH Representing the complex scattering coefficient when the transmitting polarization mode is horizontal polarization and the receiving polarization mode is horizontal polarization, S VV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is vertical polarization, S HV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is horizontal polarization;
step 1.2: based on the loss K, a polarization coherence matrix T is obtained using equation (1):
in the formula (1) (. Cndot. H Represents conjugate transpose, T ij Elements representing the ith row and jth column of the polarization coherence matrix T; and i, j=1, 2,3;
step 1.3: extracting six-dimensional eigenvectors f= [ a, B, C, D, E, F ] from the polarization coherence matrix T using equation (2):
in equation (2), span represents the total scattered power of all polarized channels, and span=t 11 +T 22 +T 33 A represents the decibel form of the total scattered power Span, B and C represent the 2 nd row and 2 nd column elements T, respectively 22 And line 3, column 3 element T 33 D, E and F represent three relative correlation coefficients;
step 1.4: normalizing the six-dimensional feature vector F to obtain a normalized six-dimensional feature vector F', so that preprocessed PolSAR data are formed by the normalized six-dimensional feature vectors of all minimum resolution units;
step 1.5: cutting the preprocessed PolSAR data into slices with the size of L multiplied by 6, thereby obtaining a slice set { s } of the PolSAR data 1 ,...,s n ,...,s N -wherein s n Represents the nth slice, N represents the total number of slices, n.epsilon.1, N]In this embodiment, the data is cut into slices of size l×l×6=15×15×6, the total number of slices being 0.5% of the amount of tag data in the data set;
step 2: constructing a multi-index convolutional self-encoder based classification network, comprising: an encoder, a decoder, and a classifier; and to aggregate the slices { s } 1 ,...,s n ,...,s N The input into the classification network,
step 2.1: the encoder consists of m convolution layers, m pooling layers, m MI-SE modules, m weighting modules and a full connection layer, in this embodiment, two convolution layers, two pooling layers, two MI-SE modules, two weighting modules and a full connection layer are constructed;
nth slice s n Inputting the characteristic diagram U into an encoder, and obtaining a characteristic diagram U with a scale of H multiplied by W multiplied by D after processing of a first convolution layer and a pooling layer, wherein H and W respectively represent the height and width of the characteristic diagram U, and D represents the channel number of the characteristic diagram U;
the profile U is input into a first MI-SE module and a multi-index is calculated using equation (3), comprising: first moment S 1 Second order central moment S 2 And coefficient of variation S 3
In the formula (3), U (a, b, i) represents all elements of the a-th row and the b-th column of the characteristic diagram U;
will first moment S 1 Second order central moment S 2 And coefficient of variation S 3 Splicing the matrix into a multi-index matrix V with the size of 3 xD, and processing the multi-index matrix V by a convolution layer and a full connection layer in a first MI-SE module to obtain a weight matrix Y with the size of 1 xD;
the first weighting module multiplies the weight matrix Y with the feature map U to obtain a weighted feature map U' with the size of H multiplied by W multiplied by D;
the weighted characteristic diagram U' sequentially passes through the rest m-1 convolution layers, the pooling layer, the MI-SE module and the weighting module of the encoder, and finally the output characteristic passes through a full connection layer to obtain the characteristic vector alpha output by the encoder;
step 2.2: the decoder is formed by sequentially connecting k convolution layers and k up-sampling layers in a staggered manner;
the feature vector alpha is input into the decoder to obtain the nth slice s n Is a reconstructed slice of (a)Re-use of (4) to establish a counter-propagating loss function L MSE
In equation (4), MSE represents the minimum mean square error; p is the number of elements per slice,for the nth slice s n The c-th element in the slice reconstructed after decoder,>for the nth slice s n C element of (a);
encoder and decoder training with back propagation when the loss function L MSE When the minimum value is taken, training is stopped, and the optimal feature vector alpha output by the trained encoder is obtained * In this embodiment, three convolution layers and are provided in the decoderThree upsampling layers;
step 2.3: the classifier consists of t full-connection layers and one softmax layer;
optimal feature vector alpha * In the input classifier, firstly, reducing the dimension through t full-connection layers, then inputting the feature vector after the dimension reduction into a softmax layer, and calculating an nth slice s n The center element of (a) corresponds to the posterior probability of each category, and the category corresponding to the maximum posterior probability is selected as the nth slice s n In this embodiment, two fully connected layers and one softmax layer are built in the classifier;
step 2.4: obtaining prediction categories of N slices according to the modes of the steps 2.1-2.3, comparing the prediction categories with corresponding real categories, optimizing the classifier through back propagation, and training the encoder, the decoder and the classifier, so that a trained multi-index convolution self-encoder model is obtained and used for classifying PolSAR data to be classified.
So far, the PolSAR terrain fine classification method based on the multi-index convolution self-encoder is basically completed.
The effectiveness of the invention is further illustrated by the San Francisco dataset experiments below.
San Francisco dataset PolSAR image target classification experiment:
1. experiment setting:
the experimental dataset was an L-band full polarized synthetic aperture radar image collected from san francisco by the air synthetic aperture radar platform of the united states space agency jet propulsion laboratory (Jet Propulsion Laboratory, JPL). The data set has a size of 1300 x 1200 and comprises five ground objects of low-density cities, high-density cities, developed cities, water bodies and vegetation. In order to input the data set into the multi-index convolutional self-encoder model, the size of the data slice was set to 15×15 in the experiment, and furthermore, the number of training samples in the data set was randomly selected to be 0.5% of the total samples. The dataset is shown in fig. 3 and the final classification result is shown in fig. 4. Fig. 3 (a) shows original data. Fig. 3 (b) shows the case where data can be separated when conventional CAE is used. Fig. 3 (c) shows the case where data can be separated after CAE is added to the first moment. Fig. 3 (d) shows the case where data can be separated after CAE has been added to the second order central moment. FIG. 3 (e) shows how the data can be separated after CAE is added to the coefficient of variation. FIG. 3 (f) shows the case where data can be separated when the method of the present invention is employed;
2. analysis of results:
the experiment in this example uses the Overall Accuracy (OA) and Kappa coefficient (Kappa coefficient) to quantitatively analyze the performance of the proposed method. To illustrate the superiority of the proposed methods, several common PolSAR image ground object classification methods are selected for comparison, and most of the methods are based on improvement of convolutional neural networks (convolutional neural networks, CNN), wherein the methods comprise methods based on complete local binary pattern fusion of CNN (completed local binary patterns feature integrated CNN, CLBP-CNN), combination of CNN and Markov random fields (combination of CNN and Markov random field, MRF-CNN), polarization compression and excitation networks (polarimetric squeeze-and-excitation network, PSE-NET), complex CNN (complex-valued CNN, CV-CNN), VGG-Net, residual neural networks (residual neural network, res-Net) and CAE. Batch processing size is set to be 64 in the training process of all methods, iteration times are 100, learning rate is 0.001, and optimization algorithm is Adam algorithm. The comparison results are shown in Table 1, wherein:
t and N in equation (5) represent the number of categories and the total number of samples in the classification task, c ii Corresponding to diagonal elements in the confusion matrix. C in formula (6) ij Corresponding to all elements in the confusion matrix.
TABLE 1 PolSAR terrain fine classification method Performance evaluation index based on Multi-index convolutional self-encoder
Analysis is performed in conjunction with fig. 5 and table 1, wherein (a) in fig. 5 represents an original label map, (b) in fig. 5-fig. 5 (h) represents a classification result of the comparison algorithm, and (i) in fig. 5 represents a classification result of the present invention; compared with other common PolSAR ground feature classification algorithms, the PolSAR ground feature fine classification algorithm based on the multi-index convolution self-encoder provided by the invention has better continuity in the classification result, namely fewer isolated points in the classification result image. In several other comparison methods, CLBP-CNN fuses description edge and texture feature CLBP, but the time cost spent in extracting features and training process in practical application is high due to the high CLBP data dimension. MRF-CNN is limited by the underlying network, and when the classification task is complex, it is difficult to remove the misclassified area, and even the classification performance is degraded again. PSE-NET increases the correlation between information through the proposed SE model, thus improving classification performance, but texture features are still lost, and more sample data is needed to participate in training due to the increase of parameters, otherwise the classification result is poor. The internal network parameters in CV-CNN also belong to a complex form, so the number of parameters that need to be trained is large. VGG-Net and Res-Net have very deep network depths, and also have very high requirements on the data volume of training samples, otherwise the final classification is less effective. The method of the invention embeds the newly proposed MI-SE module on the basis of CAE, effectively inhibits the interference information in the data set, and simultaneously carries out multi-index weighting on the characteristics extracted by the convolution and pooling layers, thereby greatly improving the characteristic separation degree. At the same time, three metrics are presented in the MI-SE module: the first moment may reflect the average of the features; the second order center distance may describe the fluctuations in each feature, reflecting the content of the edge and texture information to some extent; the coefficient of variation can eliminate the influence of the average on the dispersion, thereby better reflecting the fluctuation. The three indexes complement each other to realize weighting of each feature at different angles.
In conclusion, the PolSAR ground object fine classification method based on the multi-index convolution self-encoder has the characteristics of simple structure, accurate classification and high calculation efficiency, and has high application value in practical engineering.

Claims (1)

1. A PolSAR ground object fine classification method based on a multi-index convolution self-encoder is characterized by comprising the following steps:
step 1: preprocessing and clipping operation of PolSAR data:
step 1.1: the scattering characteristic of any minimum resolution unit in the PolSAR data is represented by using a polarized scattering matrix S; according to the reciprocity principle and Pauli decomposition principle, converting any one polarized scattering matrix S in PolSAR data into a vectorWherein S is HH Representing the complex scattering coefficient when the transmitting polarization mode is horizontal polarization and the receiving polarization mode is horizontal polarization, S VV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is vertical polarization, S HV Representing the complex scattering coefficient when the transmitting polarization mode is vertical polarization and the receiving polarization mode is horizontal polarization;
step 1.2: based on the loss K, a polarization coherence matrix T is obtained using equation (1):
in the formula (1) (. Cndot. H Represents conjugate transpose, T ij Elements representing the ith row and jth column of the polarization coherence matrix T; and i, j=1, 2,3;
step 1.3: extracting a six-dimensional eigenvector f= [ a, B, C, D, E, F ] from the polarization coherence matrix T using formula (2):
in equation (2), span represents the total scattered power of all polarized channels, and span=t 11 +T 22 +T 33 A represents the decibel form of the total scattered power Span, B and C represent the 2 nd row and 2 nd column elements T, respectively 22 And line 3, column 3 element T 33 D, E and F represent three relative correlation coefficients;
step 1.4: normalizing the six-dimensional feature vector F to obtain a normalized six-dimensional feature vector F', so that preprocessed PolSAR data are formed by the normalized six-dimensional feature vectors of all minimum resolution units;
step 1.5: cutting the preprocessed PolSAR data into slices with the size of L multiplied by 6, thereby obtaining a slice set { s } of the PolSAR data 1 ,...,s n ,...,s N -wherein s n Represents an nth slice, N represents the total number of slices; n is E [1, N];
Step 2: constructing a multi-index convolutional self-encoder based classification network, comprising: an encoder, a decoder, and a classifier; and aggregating the slices { s } 1 ,...,s n ,...,s N -input into the classification network;
step 2.1: the encoder consists of m convolution layers, m pooling layers, m MI-SE modules, m weighting modules and a full connection layer;
the nth slice s n Inputting the characteristic diagram U into the encoder, and obtaining a characteristic diagram U with a scale of H multiplied by W multiplied by D after the treatment of a first convolution layer and a pooling layer, wherein H and W respectively represent the height and the width of the characteristic diagram U, and D represents the channel number of the characteristic diagram U;
the feature map U is input into a first MI-SE module, and a multi-index is calculated by using a formula (3), and the method comprises the following steps: first moment S 1 Second order central moment S 2 And coefficient of variation S 3
In the formula (3), U (a, b, i) represents all elements of the a-th row and the b-th column of the characteristic diagram U;
the first moment S 1 Second order central moment S 2 And coefficient of variation S 3 Splicing the matrix into a multi-index matrix V with the size of 3 xD, and processing the multi-index matrix V by a convolution layer and a full connection layer in the first MI-SE module to obtain a weight matrix Y with the size of 1 xD;
the first weighting module multiplies the weight matrix Y with the feature map U to obtain a weighted feature map U' with the size of H multiplied by W multiplied by D;
the weighted feature map U' sequentially passes through the rest m-1 convolution layers, the pooling layer, the MI-SE module and the weighting module of the encoder, and finally the output features pass through a full connection layer to obtain the feature vector alpha output by the encoder;
step 2.2: the decoder is formed by sequentially connecting k convolution layers and k up-sampling layers in a staggered manner;
the feature vector alpha is input into a decoder to obtain an nth slice s n Is a reconstructed slice of (a)Re-use of (4) to establish a counter-propagating loss function L MSE
In equation (4), MSE represents the minimum mean square error; p is the number of elements per slice,for the nth slice s n The c-th element in the slice reconstructed after decoder,>for the nth slice s n C element of (a);
the encoder and decoder are trained using back propagation whenThe loss function L MSE When the minimum value is taken, training is stopped, and the optimal feature vector alpha output by the trained encoder is obtained *
Step 2.3: the classifier consists of t full-connection layers and one softmax layer;
the optimal feature vector alpha * In the input classifier, firstly, reducing the dimension through t full-connection layers, then inputting the feature vector after the dimension reduction into a softmax layer, and calculating an nth slice s n The center element of (a) corresponds to the posterior probability of each category, and the category corresponding to the maximum posterior probability is selected as the nth slice s n Is a predictive category of (2);
step 2.4: obtaining prediction categories of N slices according to the modes of the steps 2.1-2.3, comparing the prediction categories with corresponding real categories, and optimizing the classifier through back propagation, so that training of the encoder, the decoder and the classifier is completed, and a trained multi-index convolution self-encoder model is obtained and used for classifying PolSAR data to be classified.
CN202111319995.4A 2021-11-09 2021-11-09 PolSAR ground object fine classification method based on multi-index convolution self-encoder Active CN114037896B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202111319995.4A CN114037896B (en) 2021-11-09 2021-11-09 PolSAR ground object fine classification method based on multi-index convolution self-encoder

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202111319995.4A CN114037896B (en) 2021-11-09 2021-11-09 PolSAR ground object fine classification method based on multi-index convolution self-encoder

Publications (2)

Publication Number Publication Date
CN114037896A CN114037896A (en) 2022-02-11
CN114037896B true CN114037896B (en) 2024-02-13

Family

ID=80143590

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202111319995.4A Active CN114037896B (en) 2021-11-09 2021-11-09 PolSAR ground object fine classification method based on multi-index convolution self-encoder

Country Status (1)

Country Link
CN (1) CN114037896B (en)

Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020162834A1 (en) * 2019-02-08 2020-08-13 Singapore Health Services Pte Ltd Method and system for classification and visualisation of 3d images
CN112434628A (en) * 2020-11-30 2021-03-02 西安理工大学 Small sample polarization SAR image classification method based on active learning and collaborative representation
CN112966779A (en) * 2021-03-29 2021-06-15 安徽大学 PolSAR image semi-supervised classification method
CN113240040A (en) * 2021-05-27 2021-08-10 西安理工大学 Polarized SAR image classification method based on channel attention depth network
CN113392871A (en) * 2021-04-06 2021-09-14 北京化工大学 Polarized SAR terrain classification method based on scattering mechanism multichannel expansion convolutional neural network

Patent Citations (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
WO2020162834A1 (en) * 2019-02-08 2020-08-13 Singapore Health Services Pte Ltd Method and system for classification and visualisation of 3d images
CN112434628A (en) * 2020-11-30 2021-03-02 西安理工大学 Small sample polarization SAR image classification method based on active learning and collaborative representation
CN112966779A (en) * 2021-03-29 2021-06-15 安徽大学 PolSAR image semi-supervised classification method
CN113392871A (en) * 2021-04-06 2021-09-14 北京化工大学 Polarized SAR terrain classification method based on scattering mechanism multichannel expansion convolutional neural network
CN113240040A (en) * 2021-05-27 2021-08-10 西安理工大学 Polarized SAR image classification method based on channel attention depth network

Non-Patent Citations (1)

* Cited by examiner, † Cited by third party
Title
全卷积网络和条件随机场相结合的全极化SAR土地覆盖分类;赵泉华;谢凯浪;王光辉;李玉;;测绘学报;20200115(01);全文 *

Also Published As

Publication number Publication date
CN114037896A (en) 2022-02-11

Similar Documents

Publication Publication Date Title
CN110135267B (en) Large-scene SAR image fine target detection method
CN110533631B (en) SAR image change detection method based on pyramid pooling twin network
CN109214452B (en) HRRP target identification method based on attention depth bidirectional cyclic neural network
CN109190491B (en) Sea ice classification method for residual convolutional neural network SAR (synthetic Aperture Radar) image
CN106529721B (en) A kind of ad click rate forecasting system and its prediction technique that depth characteristic is extracted
CN100383805C (en) Method for sorting characters of ground object through interfusion of satellite carried microwave and infrared remote sensing
CN111639587B (en) Hyperspectral image classification method based on multi-scale spectrum space convolution neural network
CN113240040B (en) Polarized SAR image classification method based on channel attention depth network
Yang et al. Aa-transunet: Attention augmented transunet for nowcasting tasks
CN113343563B (en) Landslide susceptibility evaluation method based on automatic sample selection and surface deformation rate
CN111709487B (en) Underwater multi-source acoustic image substrate classification method and system based on decision-level fusion
CN104732244A (en) Wavelet transform, multi-strategy PSO (particle swarm optimization) and SVM (support vector machine) integrated based remote sensing image classification method
CN104252625A (en) Sample adaptive multi-feature weighted remote sensing image method
CN110674858A (en) Traffic jam detection method based on space-time association and big data mining
CN103366371A (en) K distribution and texture feature-based SAR (Synthetic Aperture Radar) image segmentation method
CN115222994A (en) Hyperspectral image classification method based on hybrid spectrum network and multi-head self-attention mechanism
CN114913434B (en) High-resolution remote sensing image change detection method based on global relation reasoning
CN106251375A (en) A kind of degree of depth study stacking-type automatic coding of general steganalysis
CN106446935A (en) Kernel sparse representation and spatial constraint-based polarimetric SAR image classification method
CN114037001A (en) Mechanical pump small sample fault diagnosis method based on WGAN-GP-C and metric learning
CN114821340A (en) Land utilization classification method and system
CN104809471A (en) Hyperspectral image residual error fusion classification method based on space spectrum information
CN116580322A (en) Unmanned aerial vehicle infrared small target detection method under ground background
CN102073875B (en) Sparse representation-based background clutter quantification method
CN103955711A (en) Mode recognition method in imaging spectrum object recognition analysis

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant