CN112417188A - Hyperspectral image classification method based on graph model - Google Patents

Hyperspectral image classification method based on graph model Download PDF

Info

Publication number
CN112417188A
CN112417188A CN202011436246.5A CN202011436246A CN112417188A CN 112417188 A CN112417188 A CN 112417188A CN 202011436246 A CN202011436246 A CN 202011436246A CN 112417188 A CN112417188 A CN 112417188A
Authority
CN
China
Prior art keywords
node
formula
matrix
equation
value
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Granted
Application number
CN202011436246.5A
Other languages
Chinese (zh)
Other versions
CN112417188B (en
Inventor
蒋俊正
黄炟鑫
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Guilin University of Electronic Technology
Original Assignee
Guilin University of Electronic Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Guilin University of Electronic Technology filed Critical Guilin University of Electronic Technology
Priority to CN202011436246.5A priority Critical patent/CN112417188B/en
Publication of CN112417188A publication Critical patent/CN112417188A/en
Application granted granted Critical
Publication of CN112417188B publication Critical patent/CN112417188B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/55Clustering; Classification
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/583Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using metadata automatically derived from the content
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/50Information retrieval; Database structures therefor; File system structures therefor of still image data
    • G06F16/58Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually
    • G06F16/5866Retrieval characterised by using metadata, e.g. metadata not derived from the content or metadata generated manually using information manually generated, e.g. tags, keywords, comments, manually generated location and time information
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F16/00Information retrieval; Database structures therefor; File system structures therefor
    • G06F16/90Details of database functions independent of the retrieved data types
    • G06F16/901Indexing; Data structures therefor; Storage structures
    • G06F16/9024Graphs; Linked lists
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques
    • G06F18/241Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
    • G06F18/2413Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
    • G06F18/24147Distances to closest patterns, e.g. nearest neighbour classification

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Data Mining & Analysis (AREA)
  • Physics & Mathematics (AREA)
  • Databases & Information Systems (AREA)
  • General Physics & Mathematics (AREA)
  • General Engineering & Computer Science (AREA)
  • Library & Information Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Evolutionary Computation (AREA)
  • Evolutionary Biology (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • Artificial Intelligence (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Software Systems (AREA)
  • Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
  • Image Analysis (AREA)

Abstract

The invention discloses a hyperspectral image classification method based on a graph model, which is characterized by comprising the following steps of: 1) modeling a graph; 2) resolving an optimization problem; 3) decomposing a Hessian matrix; 4) an approximation of newton step size; 5) carrying out iterative solution; 6) distributed solving; 7) and solving the classification result. The method has low calculation complexity under large-scale data and can finish classification with higher precision.

Description

Hyperspectral image classification method based on graph model
Technical Field
The invention relates to the technical field of hyperspectral images, graph models and machine learning, in particular to a hyperspectral image classification method based on a graph model.
Background
The hyperspectral image data is a special image, is firstly applied to the satellite remote sensing technology, and is widely applied to the fields of military affairs, medicine, geology, agriculture and the like. The common image is subjected to broadband imaging during imaging, the imaging result hardly has spectral characteristics, and the related research is extremely dependent on the spatial resolution of the image. The hyperspectral image is the result of imaging of multiple narrow bands, and compared with a common image, the hyperspectral image carries abundant spectrum information. The research on the hyperspectral data is very extensive, and classification is an important item. The hyperspectral image classification aims at marking pixel points, and the marking result can well reflect the relevant characteristics of the target to be detected. However, the hyperspectral image classification has the problems of small known sample size, high data dimensionality and the like.
The graph is a common data structure and is commonly used in wireless sensor networks, social networks and other scenarios. In addition, graphs are also commonly used for dimensionality reduction of data with higher feature dimensions.
Machine learning is a generic term for a class of task solution methods that mathematically model related tasks and then use tools such as computers to solve the modeled optimization problem to obtain results. Machine learning has been widely used in industry, traffic, IC design, etc. through decades of development.
Disclosure of Invention
The invention aims to provide a hyperspectral image classification method based on a graph model, aiming at the problems of large data size, high data dimensionality, small known sample size and the like in a hyperspectral image classification task. The method has low calculation complexity under large-scale data and can finish classification with higher precision.
The technical scheme for realizing the purpose of the invention is as follows:
a hyperspectral image classification method based on a graph model comprises the following steps:
1) graph modeling: suppose the hyperspectral image data is X ═ X1,x2,…,xm×n]∈Rm×n×dTotal of m × n pixel points, xi∈RdIs the feature vector of the ith pixel,
Figure BDA0002828363860000011
labels on pixels in the data are all from L ═ L1,l2,...,lcAnd all the types of the C are replaced by C,
Figure BDA0002828363860000012
is a tag of a known moiety in X,
Figure BDA0002828363860000013
is an unknown tag, the values are all 0, and
Figure BDA0002828363860000014
feature vector x of different pixel points of hyperspectral image dataiThe graph G is constructed as { V, E, W }, where V is a node set of the graph, corresponding to each pixel in the data and connected by an edge, E is an edge set, the edge is used to describe similarity and adjacency relation between nodes, W is a weight matrix, and an internal element represents a similarity degree corresponding to two nodes and is defined as formula (1):
Figure BDA0002828363860000021
Figure BDA0002828363860000022
is a neighbor set of the node i;
2) the optimization problem is summarized as follows: the one-hot encoding matrix Y of the label Y is defined as shown in formula (2):
Figure BDA0002828363860000023
yifor the ith element in the label Y, the classification problem of the hyperspectral image can be summarized as shown in formula (3) by Y:
Figure BDA0002828363860000024
and (3) solving the formula (3) to obtain a one-hot coding matrix F for classification, wherein in the formula (3), alpha is a weight factor, and L isn=I-D-1/2WD-1/2Is a normalized graph Laplace matrix, I is a unit matrix, yjAnd fjThe j-th columns of Y and F respectively,
Figure BDA0002828363860000025
is fjD is a degree matrix, defined as shown in formula (4):
Figure BDA0002828363860000026
will D-1/2WD-1/2Is marked as WnThe problem in the formula (3) is composed of c optimization problems, which are independent from each other and similar in solving process, and only one of the optimization problems is analyzed, so that f is equal to fjThe method comprises the following steps:
Figure BDA0002828363860000027
in the formula (5)
Figure BDA0002828363860000028
In order to match the terms of the search,
Figure BDA0002828363860000029
is a regular term;
3) decomposition of the Hessian matrix: the problem in the formula (5) is solved by using a Newton method, and the information of the first order and the second order gradient of the problem is respectively shown in formulas (6) and (7):
Figure BDA00028283638600000210
Figure BDA00028283638600000211
thus obtaining the Hessian matrix H ═ alpha I + L of the objective functionnLet Wn=Wnd+WnuWherein W isnd=diag(Wn) From Ln=I-WnThe method comprises the following steps:
A=αI+(1+θ)(I-Wnd) (8),
B=Wnu+θ(I-Wnd) (9),
wherein theta is a parameter for controlling decomposition, and the Hessian matrix can be decomposed into a combination of A and B as shown in formula (10):
H=A-B (10);
4) approximation of newton step size: when solving by the Newton method, the Newton step length of the k step
Figure BDA0002828363860000031
From equation (10) there is:
Figure BDA0002828363860000032
let P equal Wn-αI,Q=(I-P)-1Q ≈ I + P by Taylor expansion, the Newton step size of equation (11) is approximated by equation (12):
Figure BDA0002828363860000033
5) and (3) iterative solution: equation (5) is solved according to equation (13) using the approximate newton step size in equation (12):
Figure BDA0002828363860000034
in equation (13), k in the upper right corner of the variable is the number of iterations, and f0=0,f1When 1 is equal to
Figure BDA0002828363860000035
When the result is true, terminating the iteration;
6) distributed solution: the matching terms in the formula (5) are respectively expressed as
Figure BDA0002828363860000036
The first order gradient information at the ith node is shown in equation (14):
Figure BDA0002828363860000037
wijis WnThe element in the ith row and j column corresponds to the weight value between the node i and the node j on the graph, so that for the node i, the following elements exist:
6-1) with
Figure BDA0002828363860000038
The node in (1) communicates to obtain fkIn that
Figure BDA0002828363860000039
A value of (d) above;
6-2) calculation
Figure BDA00028283638600000310
Figure BDA00028283638600000311
Is ekValue at node i, AiiIs the ith element on the diagonal of A;
6-3) with
Figure BDA0002828363860000041
The node in (1) communicates to obtain e at
Figure BDA0002828363860000042
A value of (d) above;
6-4) calculation
Figure BDA0002828363860000043
Figure BDA0002828363860000044
Is ukThe value at node i;
6-5) with
Figure BDA0002828363860000045
The node in (1) communicates to obtain u is
Figure BDA0002828363860000046
A value of (d);
6-6) calculating the step size
Figure BDA0002828363860000047
Figure BDA0002828363860000048
Is the step length skThe value at node i;
6-7) calculation
Figure BDA0002828363860000049
To fkUpdating the value at the node i;
6-8) on all nodes
Figure BDA00028283638600000410
Are spliced into fk+1Judging the termination condition
Figure BDA00028283638600000411
Whether or not: if the condition is terminated
Figure BDA00028283638600000412
If it is true, fk+1Setting a corresponding column in the one-hot matrix, and starting to solve the next column of the one-hot matrix; if the condition is terminated
Figure BDA00028283638600000413
If not, f is usedk+1Performing the next iteration with the initial value;
7) solving a classification result: obtaining a one-hot coding matrix F through the step 6), and finally obtaining a classification result through a formula (15)
Figure BDA00028283638600000414
Figure BDA00028283638600000415
The technical scheme includes that a graph model is adopted to carry out dimensionality reduction on hyperspectral image data, then a learning task is subjected to problem regression, then a Hessian matrix is approximated through matrix decomposition, and the approximated Hessian matrix is used for solving through a quasi-Newton method.
The method has the advantages of low calculation complexity and high classification precision, and is more suitable for the large-scale hyperspectral image classification problem.
Drawings
FIG. 1 is a schematic diagram of iterative convergence in KSC data for implementing the method;
FIG. 2 is a schematic diagram of iterative convergence in IndianPines data by the implementation method.
Detailed Description
The invention will be further described with reference to the following drawings and examples, but is not limited thereto.
Example (b):
a hyperspectral image classification method based on a graph model comprises the following steps:
1) graph modeling: suppose the hyperspectral image data is X ═ X1,x2,...,xm×n]∈Rm×n×dTotal of m × n pixel points, xi∈RdIs the feature vector of the ith pixel,
Figure BDA00028283638600000416
labels on pixels in the data are all from L ═ L1,l2,...,lcAnd all the types of the C are replaced by C,
Figure BDA00028283638600000417
is a tag of a known moiety in X,
Figure BDA00028283638600000418
is an unknown tag, the values are all 0, and
Figure BDA00028283638600000419
feature vector x of different pixel points of hyperspectral image dataiAccording to the similar characteristics, a graph G is constructed, wherein the graph G is { V, E, W }, V is a node set of the graph, corresponding to each pixel in the data and connected by an edge, E is an edge set, and the edge is used for describing similarity and adjacent relation between nodesW is a weight matrix, and the internal elements represent the degree of similarity between two nodes, and are defined as shown in formula (1):
Figure BDA0002828363860000051
Figure BDA0002828363860000052
is a neighbor set of the node i;
2) the optimization problem is summarized as follows: the one-hot encoding matrix Y of the label Y is defined as shown in formula (2):
Figure BDA0002828363860000053
yifor the ith element in the label Y, the classification problem of the hyperspectral image can be summarized as shown in formula (3) by Y:
Figure BDA0002828363860000054
and (3) solving the formula (3) to obtain a one-hot coding matrix F for classification, wherein in the formula (3), alpha is a weight factor, and L isn=I-D-1/2WD-1/2Is a normalized graph Laplace matrix, I is a unit matrix, yjAnd fjThe j-th columns of Y and F respectively,
Figure BDA0002828363860000055
is fjD is a degree matrix, defined as shown in formula (4):
Figure BDA0002828363860000056
will D-1/2WD-1/2Is marked as WnThe problem in the formula (3) consists of c optimization problems which are independent from each other and similar in solving process, and only one of the optimization problems is analyzed to enable the problem to be solvedf=fjThe method comprises the following steps:
Figure BDA0002828363860000057
in the formula (5)
Figure BDA0002828363860000058
In order to match the terms of the search,
Figure BDA0002828363860000059
is a regular term;
3) decomposition of the Hessian matrix: the problem in the formula (5) is solved by using a Newton method, and the information of the first order and the second order gradient of the problem is respectively shown in formulas (6) and (7):
Figure BDA00028283638600000510
Figure BDA00028283638600000511
thus obtaining the Hessian matrix H ═ alpha I + L of the objective functionnLet Wn=Wnd+WnuWherein W isnd=diag(Wn) From Ln=I-WnThe method comprises the following steps:
A=αI+(1+θ)(I-Wnd) (8),
B=Wnu+θ(I-Wnd) (9),
wherein theta is a parameter for controlling decomposition, and the Hessian matrix can be decomposed into a combination of A and B as shown in formula (10):
H=A-B (10);
4) approximation of newton step size: when solving by the Newton method, the Newton step length of the k step
Figure BDA0002828363860000061
From equation (10) there is:
Figure BDA0002828363860000062
let P equal Wn-αI,Q=(I-P)-1Q ≈ I + P by Taylor expansion, the Newton step size of equation (11) is approximated by equation (12):
Figure BDA0002828363860000063
5) and (3) iterative solution: equation (5) is solved according to equation (13) using the approximate newton step size in equation (12):
Figure BDA0002828363860000064
in equation (13), k in the upper right corner of the variable is the number of iterations, and f0=0,f1When 1 is equal to
Figure BDA0002828363860000065
When the result is true, terminating the iteration;
6) distributed solution: the matching terms in the formula (5) are respectively expressed as
Figure BDA0002828363860000066
The first order gradient information at the ith node is shown in equation (14):
Figure BDA0002828363860000067
wijis WnThe element in the ith row and j column corresponds to the weight value between the node i and the node j on the graph, so that for the node i, the following elements exist:
6-1) with
Figure BDA0002828363860000068
The node in (1) communicates to obtain fkIn that
Figure BDA0002828363860000069
A value of (d) above;
6-2) calculation
Figure BDA0002828363860000071
Figure BDA0002828363860000072
Is ekValue at node i, AiiIs the ith element on the diagonal of A;
6-3) with
Figure BDA0002828363860000073
The node in (1) communicates to obtain e at
Figure BDA0002828363860000074
A value of (d) above;
6-4) calculation
Figure BDA0002828363860000075
Figure BDA0002828363860000076
Is ukThe value at node i;
6-5) with
Figure BDA0002828363860000077
The node in (1) communicates to obtain u is
Figure BDA0002828363860000078
A value of (d);
6-6) calculating the step size
Figure BDA0002828363860000079
Figure BDA00028283638600000710
Is the step length skThe value at node i;
6-7) calculation
Figure BDA00028283638600000711
To fkAt one sectionUpdating the value at the point i;
6-8) on all nodes
Figure BDA00028283638600000712
Are spliced into fk+1Judging the termination condition
Figure BDA00028283638600000713
Whether or not: if the condition is terminated
Figure BDA00028283638600000714
If it is true, fk+1Setting a corresponding column in the one-hot matrix, and starting to solve the next column of the one-hot matrix; if the condition is terminated
Figure BDA00028283638600000715
If not, f is usedk+1Performing the next iteration with the initial value;
7) solving a classification result: obtaining a one-hot coding matrix F through the step 6), and finally obtaining a classification result through a formula (15)
Figure BDA00028283638600000716
Figure BDA00028283638600000717
Simulation example 1:
in the embodiment, KSC data is adopted for simulation, 176 frequency bands are arranged in the KSC data, 5211 pixel points are total after background removal, the pixel points can be divided into 13 classes and are a multi-class learning task, before simulation, tag values are sampled to generate data sets with different known tag numbers, for multi-item comparison, data sets with 5, 10, 15 and 20 known samples in each class are generated, table 1 shows that the classification performance comparison in the KSC data by adopting the method of the embodiment and the centralized method is performed, the OA parameters and the calculation time are included, and the simulation result of table 1 shows that in the KSC data, compared with the centralized method, the classification accuracy by adopting the method of the embodiment and the centralized method is completely consistent, the calculation speed by adopting the method of the embodiment is slightly higher than that by adopting the centralized method, as shown in figure 1, iterative convergence can be performed within limited times by adopting the method of the embodiment, it is demonstrated that the method of the present example produces better results and is less time consuming.
TABLE 1
Figure BDA00028283638600000718
Figure BDA0002828363860000081
Simulation example 2:
in the embodiment, IndanPines data is adopted for simulation, 200 frequency bands exist in the IndanPines data, 10249 pixel points are obtained after background removal, the pixel points can be divided into 16 classes and are also a multi-class learning task, as in simulation example 1, sampling produced datasets of only 5, 10, 15, 20 known samples for each class, and table 2 shows the comparison of classification performance in indianpins data using the method of the present example and the centralized method, which also includes OA parameters and the time consumption of calculation, the simulation results in table 2 show that, in Indian pines data, compared with the centralized method, the classification accuracy of the method is basically consistent with that of the centralized method, in the simulation example, the calculation speed of the method is far better than that of the method adopting a centralized method, and as shown in fig. 2, iterative convergence can be performed within a limited number of times by adopting the method of the example, which shows that the method of the example is more suitable for a hyperspectral image classification task under large-scale data.
TABLE 2
Number of samples known 5 10 15 20
OA value Using the method of this example 41.98% 46.69% 49.71% 51.67%
Using this example method time consuming(s) 37.09 38.59 39.1 39.83
Using centralized approach OA values 42.02% 46.7% 49.72% 51.68%
Consuming time using a centralized approach(s) 242.9 271 242.5 299.6

Claims (1)

1. A hyperspectral image classification method based on a graph model is characterized by comprising the following steps:
1) graph modeling: suppose the hyperspectral image data is X ═ X1,x2,…,xm×n]∈Rm×n×dTotal of m × n pixel points, xi∈RdIs the feature vector of the ith pixel,
Figure FDA0002828363850000011
labels on pixels in the data are all from L ═ L1,l2,...,lcAnd all the types of the C are replaced by C,
Figure FDA0002828363850000012
is a tag of a known moiety in X,
Figure FDA0002828363850000013
is an unknown tag, the values are all 0, and
Figure FDA0002828363850000014
feature vector x of different pixel points of hyperspectral image dataiThe graph G is constructed as { V, E, W }, where V is a node set of the graph, corresponding to each pixel in the data and connected by an edge, E is an edge set, the edge is used to describe similarity and adjacency relation between nodes, W is a weight matrix, and an internal element represents a similarity degree corresponding to two nodes and is defined as formula (1):
Figure FDA0002828363850000015
Figure FDA0002828363850000016
is a neighbor set of the node i;
2) the optimization problem is summarized as follows: the one-hot encoding matrix Y of the label Y is defined as shown in formula (2):
Figure FDA0002828363850000017
yifor the ith element in the label Y, the classification problem of the hyperspectral image can be summarized as shown in formula (3) by Y:
Figure FDA0002828363850000018
and (3) solving the formula (3) to obtain a one-hot coding matrix F for classification, wherein in the formula (3), alpha is a weight factor, and L isn=I-D-1/2WD-1/2Is a normalized graph Laplace matrix, I is a unit matrix, yjAnd fjThe j-th columns of Y and F respectively,
Figure FDA0002828363850000019
is fjD is a degree matrix, defined as shown in formula (4):
Figure FDA00028283638500000110
will D-1/2WD-1/2Is marked as WnThe problem in the formula (3) is composed of c optimization problems, which are independent from each other and similar in solving process, and only one of the optimization problems is analyzed, so that f is equal to fjThe method comprises the following steps:
Figure FDA00028283638500000111
in the formula (5)
Figure FDA00028283638500000112
In order to match the terms of the search,
Figure FDA00028283638500000113
is a regular term;
3) decomposition of the Hessian matrix: the problem in the formula (5) is solved by using a Newton method, and the information of the first order and the second order gradient of the problem is respectively shown in formulas (6) and (7):
Figure FDA0002828363850000021
Figure FDA0002828363850000022
thus obtaining the Hessian matrix H ═ alpha I + L of the objective functionnLet Wn=Wnd+WnuWherein W isnd=diag(Wn) From Ln=I-WnThe method comprises the following steps:
A=αI+(1+θ)(I-Wnd) (8),
B=Wnu+θ(I-Wnd) (9),
wherein theta is a parameter for controlling decomposition, and the Hessian matrix can be decomposed into a combination of A and B as shown in formula (10):
H=A-B (10);
4) approximation of newton step size: when solving by the Newton method, the Newton step length of the k step
Figure FDA0002828363850000023
From equation (10) there is:
Figure FDA0002828363850000024
let P equal Wn-αI,Q=(I-P)-1Q ≈ I + P by Taylor expansion, the Newton step size of equation (11) is approximated by equation (12):
Figure FDA0002828363850000025
5) and (3) iterative solution: equation (5) is solved according to equation (13) using the approximate newton step size in equation (12):
Figure FDA0002828363850000026
in equation (13), k in the upper right corner of the variable is the number of iterations, and f0=0,f1When 1 is equal to
Figure FDA0002828363850000027
When the result is true, terminating the iteration;
6) distributed solution: the matching terms in the formula (5) are respectively expressed as
Figure FDA0002828363850000028
The first order gradient information at the ith node is shown in equation (14):
Figure FDA0002828363850000029
wijis WnThe element in the ith row and j column corresponds to the weight value between the node i and the node j on the graph, so that for the node i, the following elements exist:
6-1) with
Figure FDA0002828363850000031
The node in (1) communicates to obtain fkIn that
Figure FDA0002828363850000032
A value of (d) above;
6-2) calculation
Figure FDA0002828363850000033
Figure FDA0002828363850000034
Is ekValue at node i, AiiIs the ith element on the diagonal of A;
6-3) with
Figure FDA0002828363850000035
The node in (1) communicates to obtain e at
Figure FDA0002828363850000036
A value of (d) above;
6-4) calculation
Figure FDA0002828363850000037
Figure FDA0002828363850000038
Is ukThe value at node i;
6-5) with
Figure FDA0002828363850000039
The node in (1) communicates to obtain u is
Figure FDA00028283638500000310
A value of (d);
6-6) calculating the step size
Figure FDA00028283638500000311
Figure FDA00028283638500000312
Is the step length skThe value at node i;
6-7) calculation
Figure FDA00028283638500000313
To fkUpdating the value at the node i;
6-8) will f on all nodesi k+1Are spliced into fk+1Judging the termination condition
Figure FDA00028283638500000314
Whether or not: if the condition is terminated
Figure FDA00028283638500000315
If it is true, fk+1Setting a corresponding column in the one-hot matrix, and starting to solve the next column of the one-hot matrix; if the condition is terminated
Figure FDA00028283638500000316
If not, f is usedk+1Performing the next iteration with the initial value;
7) solving a classification result: obtaining a one-hot coding matrix F through the step 6), and finally obtaining a classification result through a formula (15)
Figure FDA00028283638500000317
Figure FDA00028283638500000318
CN202011436246.5A 2020-12-10 2020-12-10 Hyperspectral image classification method based on graph model Active CN112417188B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN202011436246.5A CN112417188B (en) 2020-12-10 2020-12-10 Hyperspectral image classification method based on graph model

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN202011436246.5A CN112417188B (en) 2020-12-10 2020-12-10 Hyperspectral image classification method based on graph model

Publications (2)

Publication Number Publication Date
CN112417188A true CN112417188A (en) 2021-02-26
CN112417188B CN112417188B (en) 2022-05-24

Family

ID=74775010

Family Applications (1)

Application Number Title Priority Date Filing Date
CN202011436246.5A Active CN112417188B (en) 2020-12-10 2020-12-10 Hyperspectral image classification method based on graph model

Country Status (1)

Country Link
CN (1) CN112417188B (en)

Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090318815A1 (en) * 2008-05-23 2009-12-24 Michael Barnes Systems and methods for hyperspectral medical imaging
CN102521601A (en) * 2011-11-17 2012-06-27 西安电子科技大学 Method for classifying hyperspectral images based on semi-supervised conditional random field
US20130039580A1 (en) * 2011-08-10 2013-02-14 Antonio Robles-Kelly Decomposing hyperspectral or multispectral image data
CN106937531A (en) * 2014-06-14 2017-07-07 奇跃公司 Method and system for producing virtual and augmented reality
CN106934400A (en) * 2017-02-24 2017-07-07 西安电子科技大学 Based on the hyperspectral image band selection method from expression and local similarity protection
TWI624807B (en) * 2017-09-27 2018-05-21 Taichung Veterans General Hosital Iterative analysis of medical images
CN108470192A (en) * 2018-03-13 2018-08-31 广东工业大学 A kind of hyperspectral classification method and device
CN108764276A (en) * 2018-04-12 2018-11-06 西北大学 A kind of robust weights multi-characters clusterl method automatically
CN110232317A (en) * 2019-05-05 2019-09-13 五邑大学 Hyperspectral image classification method based on super-pixel segmentation and two phase classification strategy
CN110399909A (en) * 2019-07-08 2019-11-01 南京信息工程大学 A kind of hyperspectral image classification method based on label constraint elastic network(s) graph model
CN110717354A (en) * 2018-07-11 2020-01-21 哈尔滨工业大学 Superpixel classification method based on semi-supervised K-SVD and multi-scale sparse representation
CN111242867A (en) * 2020-01-14 2020-06-05 桂林电子科技大学 Graph signal distributed online reconstruction method based on truncated Taylor series approximation
CN111275201A (en) * 2020-01-21 2020-06-12 桂林电子科技大学 Sub-graph division based distributed implementation method for semi-supervised learning of graph
CN111860612A (en) * 2020-06-29 2020-10-30 西南电子技术研究所(中国电子科技集团公司第十研究所) Unsupervised hyperspectral image hidden low-rank projection learning feature extraction method

Patent Citations (14)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20090318815A1 (en) * 2008-05-23 2009-12-24 Michael Barnes Systems and methods for hyperspectral medical imaging
US20130039580A1 (en) * 2011-08-10 2013-02-14 Antonio Robles-Kelly Decomposing hyperspectral or multispectral image data
CN102521601A (en) * 2011-11-17 2012-06-27 西安电子科技大学 Method for classifying hyperspectral images based on semi-supervised conditional random field
CN106937531A (en) * 2014-06-14 2017-07-07 奇跃公司 Method and system for producing virtual and augmented reality
CN106934400A (en) * 2017-02-24 2017-07-07 西安电子科技大学 Based on the hyperspectral image band selection method from expression and local similarity protection
TWI624807B (en) * 2017-09-27 2018-05-21 Taichung Veterans General Hosital Iterative analysis of medical images
CN108470192A (en) * 2018-03-13 2018-08-31 广东工业大学 A kind of hyperspectral classification method and device
CN108764276A (en) * 2018-04-12 2018-11-06 西北大学 A kind of robust weights multi-characters clusterl method automatically
CN110717354A (en) * 2018-07-11 2020-01-21 哈尔滨工业大学 Superpixel classification method based on semi-supervised K-SVD and multi-scale sparse representation
CN110232317A (en) * 2019-05-05 2019-09-13 五邑大学 Hyperspectral image classification method based on super-pixel segmentation and two phase classification strategy
CN110399909A (en) * 2019-07-08 2019-11-01 南京信息工程大学 A kind of hyperspectral image classification method based on label constraint elastic network(s) graph model
CN111242867A (en) * 2020-01-14 2020-06-05 桂林电子科技大学 Graph signal distributed online reconstruction method based on truncated Taylor series approximation
CN111275201A (en) * 2020-01-21 2020-06-12 桂林电子科技大学 Sub-graph division based distributed implementation method for semi-supervised learning of graph
CN111860612A (en) * 2020-06-29 2020-10-30 西南电子技术研究所(中国电子科技集团公司第十研究所) Unsupervised hyperspectral image hidden low-rank projection learning feature extraction method

Non-Patent Citations (4)

* Cited by examiner, † Cited by third party
Title
JUN BAI等: "A Graph-Based Classification Method for Hyperspectral Images", 《IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING》 *
SEZER KUTLUK等: "Spectral-Spatial Classification of Hyperspectral Images Using CNNs and Approximate Sparse Multinomial Logistic Regression", 《2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO)》 *
YUANJIESHAO等: "Probabilistic class structure regularized sparse representation graph for semi-supervised hyperspectral image classification", 《PATTERN RECOGNITION》 *
王小涛: "基于低秩特征表示与分治多元逻辑回归的高光谱图像分类", 《中国博士学位论文全文数据库工程科技Ⅱ辑》 *

Also Published As

Publication number Publication date
CN112417188B (en) 2022-05-24

Similar Documents

Publication Publication Date Title
Lei et al. Intelligent fault detection of high voltage line based on the Faster R-CNN
CN110717354B (en) Super-pixel classification method based on semi-supervised K-SVD and multi-scale sparse representation
Kang et al. Dual-path network-based hyperspectral image classification
CN106599830B (en) Face key point positioning method and device
CN108009559B (en) Hyperspectral data classification method based on space-spectrum combined information
Ma et al. Robust Sparse Hyperspectral Unmixing With $\ell_ {2, 1} $ Norm
Zass et al. Probabilistic graph and hypergraph matching
Montazer et al. An improved radial basis function neural network for object image retrieval
Jia et al. ROML: A robust feature correspondence approach for matching objects in a set of images
Qin et al. Cross-domain collaborative learning via cluster canonical correlation analysis and random walker for hyperspectral image classification
CN112906756A (en) High-image classification method and system for cross-channel quantity transfer learning
CN110705636A (en) Image classification method based on multi-sample dictionary learning and local constraint coding
CN108470192B (en) Hyperspectral classification method and device
CN111340108A (en) High-dimensional data visualization method based on optimal transmission theory
Ma et al. Multifeature-based discriminative label consistent K-SVD for hyperspectral image classification
CN116643246A (en) Deep clustering radar pulse signal sorting method based on inner product distance measurement
Zhao et al. Hyperspectral target detection method based on nonlocal self-similarity and rank-1 tensor
CN113705674B (en) Non-negative matrix factorization clustering method and device and readable storage medium
CN107578445B (en) Image discriminable region extraction method based on convolution characteristic spectrum
Qin et al. Fast blind hyperspectral unmixing based on graph laplacian
CN111461229B (en) Deep neural network optimization and image classification method based on target transfer and line search
Huang et al. Hyperspectral image clustering via sparse dictionary‐based anchored regression
CN112417188B (en) Hyperspectral image classification method based on graph model
CN111062888B (en) Hyperspectral image denoising method based on multi-target low-rank sparsity and spatial-spectral total variation
CN116563649A (en) Tensor mapping network-based hyperspectral image lightweight classification method and device

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant