CN113033641B - Semi-supervised classification method for high-dimensional data - Google Patents
Semi-supervised classification method for high-dimensional data Download PDFInfo
- Publication number
- CN113033641B CN113033641B CN202110285595.XA CN202110285595A CN113033641B CN 113033641 B CN113033641 B CN 113033641B CN 202110285595 A CN202110285595 A CN 202110285595A CN 113033641 B CN113033641 B CN 113033641B
- Authority
- CN
- China
- Prior art keywords
- matrix
- subspace
- sample
- learning
- similarity
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Active
Links
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
- G06F18/2155—Generating training patterns; Bootstrap methods, e.g. bagging or boosting characterised by the incorporation of unlabelled data, e.g. multiple instance learning [MIL], semi-supervised techniques using expectation-maximisation [EM] or naïve labelling
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2132—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
- G06F18/21322—Rendering the within-class scatter matrix non-singular
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
- G06F18/2413—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches based on distances to training or reference patterns
- G06F18/24133—Distances to prototypes
- G06F18/24143—Distances to neighbourhood prototypes, e.g. restricted Coulomb energy networks [RCEN]
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/29—Graphical models, e.g. Bayesian networks
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2132—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on discrimination criteria, e.g. discriminant analysis
- G06F18/21322—Rendering the within-class scatter matrix non-singular
- G06F18/21328—Rendering the within-class scatter matrix non-singular involving subspace restrictions, e.g. nullspace techniques
Landscapes
- Engineering & Computer Science (AREA)
- Theoretical Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Physics & Mathematics (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Bioinformatics & Computational Biology (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Mathematical Physics (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
- Image Analysis (AREA)
Abstract
The invention discloses a semi-supervised classification method for high-dimensional data, and relates to the field of artificial intelligence semi-supervised learning. The method mainly overcomes the influence of data noise and redundant features on a model in high-dimensional data in the manufacturing industry, integrates subspace learning, graph construction and classifier training into a unified framework, and achieves a better classification effect. The method comprises the following steps: 1) Inputting a training data set; 2) Normalizing the data; 3) Initializing parameters and variables; 4) Subspace learning; 5) Constructing a graph; 6) Training a classifier; 7) Repeating the steps 4) -6) circularly until the algorithm is converged; 8) Classifying the test samples; 9) And obtaining the classification accuracy. The invention completes the construction of the graph from two low-dimensional spaces, namely the label space and the subspace, effectively relieves the interference of noise data and redundant characteristics on an algorithm model, ensures the quality of the graph and improves the classification effect.
Description
Technical Field
The invention relates to the technical field of artificial intelligence semi-supervised learning, in particular to a high-dimensional data semi-supervised classification method.
Background
With the advent of the intelligent era, part of the traditional manufacturing industry is gradually closing to intelligent manufacturing. Aiming at a large amount of data generated in the manufacturing industry, an intelligent decision method is applied to optimize the flows of production, sales, service and the like, and the method is one of the main problems faced by intelligent manufacturing. The manufacturing industry often accumulates large amounts of data in the process of development. However, in the general case, these large data are not all tagged. In the case of a large amount of data and a small number of labels, if we want to use a fully supervised classification algorithm to model and analyze the data to learn some patterns of the data, a satisfactory effect cannot be obtained. Then, how can the eigen-patterns of the data be learned from a large amount of data and a small number of tags? One solution is to try to label massive training data, but this is expensive and requires a lot of manpower and material resources. Obviously, a better solution is to design an algorithm model directly starting from algorithms and models, so that the algorithm model can learn a classification model with better performance and strong generalization capability from data with only a few labels. And a semi-supervised classification algorithm is just such an algorithm model. The method utilizes a small amount of labeled samples and a large amount of unlabeled samples to learn and classify the data, thereby saving the expense of manually labeling training samples. Therefore, the semi-supervised classification algorithm has important research significance, attracts the research and exploration of the majority of scientific research personnel in recent years, and has good application prospect in industry.
The semi-supervised classification algorithm based on the graph is one of the popular research directions in the semi-supervised field in recent years, and has more excellent performance. Such algorithms are based on the assumption that the data should be in manifold space and the distribution of samples should be sufficiently smooth. By smooth, it is meant that the closer the samples, i.e. the more similar the samples, the labels should be as identical as possible. In such algorithms, a graph is usually constructed to represent similarity between samples, so as to obtain a smoothness term between samples, then a loss function, a regularization term and the smoothness term are combined together to serve as an overall objective function of a model, and classifier parameters are solved by optimizing the objective function, so that the finally trained classifier not only has a small classification loss on labeled samples, but also has a sufficiently smooth classification result on all samples (including labeled samples and unlabeled samples).
However, some current graph-based semi-supervised classification algorithms are not well suited for high-dimensional data scenarios in manufacturing. For example, data in the manufacturing industry often has missing values and data noise, which may interfere with the construction of the graph and have a certain effect on the performance of the model. Another problem is that graph-based semi-supervised classification algorithms often do not perform well when processing high-dimensional data for manufacturing due to data noise and redundancy characteristics.
Disclosure of Invention
The invention aims to overcome the defects of the prior art, provides a high-dimensional data semi-supervised classification method, can effectively relieve the influence of data noise and redundancy characteristics in high-dimensional data on a model, integrates the construction process and the classifier training process of a graph into a unified framework, and obviously improves the classification effect under the semi-supervised classification scene.
In order to achieve the purpose, the technical scheme provided by the invention is as follows: a semi-supervised classification method for high-dimensional data comprises the following steps:
1) Inputting a training data set which is a high-dimensional data set;
2) Normalizing the data, eliminating the influence of different characteristic dimensions, and simultaneously improving the speed of subsequent optimization learning;
3) Initializing a regression matrixSubspace projection matrixWherein d is the number of features of the sample, c is the number of sample classes,a real matrix representing d rows and c columns; initializing a low rank decomposition matrix of WWhereinA real number matrix representing c rows and c columns; initializing a similarity matrixParameter matrixWhere n is the number of samples, and,a real number matrix representing n rows and n columns; initializing offset vectorsWhereinA real matrix representing c rows and 1 columns;
4) And (3) subspace learning: deducing an optimal solution of a low-rank decomposition matrix B, a parameter matrix C and a subspace projection matrix A according to the proposed subspace learning objective function; because the proposed objective function relates to a plurality of optimization variables, B, C and A are iteratively updated by an alternate optimization method, the optimization is carried out step by step, the subspace quality is improved, and the optimal subspace for expressing the essential characteristics of the sample is learned;
5) Comprehensively learning a sample similarity matrix from two aspects of a sample subspace and a sample label space; defining samples as nodes of the graph, defining the similarity among the samples as edges of the graph, wherein the learning process of the sample similarity matrix is the construction process of the graph;
6) Learning a semi-supervised linear regression classifier, namely a learning regression matrix W and a bias vector b, on the basis of the subspace learning in the step 4) and the similarity matrix learning in the step 5);
7) Circularly performing the step 4) to the step 6), and iteratively learning each variable until convergence; when convergence occurs, joint optimal solutions are obtained through three processes of subspace learning, graph construction and classifier learning;
8) Classifying the test samples, assuming that the input test sample is x and the sample class number is c, predicting (x) the prediction label is:
wherein (W) T x+b) i Represents a vector (W) T x + b) th element;
9) Calculating the classification accuracy: and inputting a label of the test sample, comparing the label with a prediction result, and calculating the final classification accuracy.
In step 2), the data normalization step is: obtaining the maximum value X (r) corresponding to the r row data max And minimum value X (r) min And converting the d-th row of data according to the following formula:
wherein, the first and the second end of the pipe are connected with each other,for the ith data of the r-th row,for the data after update, n is the number of samples in the dataset, d is the number of features of the samples, i ∈ {1, 2.., n }, r ∈ {1, 2.., d }.
In step 3), the initialization method is as follows: initializing a regression matrixIs an all-zero matrix; initializing a low rank decomposition matrix of WIs an all-zero matrix; initializing a similarity matrixAnd parameter matrixIs an all-zero matrix; initializing offset vectorsIs an all-zero vector; initializing a subspace projection matrix a = qf (R) as an orthogonal matrix; wherein, the first and the second end of the pipe are connected with each other,is a random matrix with each element in the interval 0,1]And qf (·) denotes QR decomposition.
In step 4), the subspace learning process is as follows:
the objective function defining the subspace learning is:
wherein tr (-) is the trace of the matrix,the F-norm of the matrix is represented,in the form of a matrix of samples,a matrix of real numbers representing d rows and n columns,in order to obtain a regression matrix,for the projection matrix of the subspace,is a low-rank decomposition matrix of W,is a parameter matrix; alpha, theta, beta are adjustable parameters;
the target function is subjected to partial derivatives of B, C and A respectively, and an updating formula of each variable can be obtained; next, each variable is updated as required:
a. according to the formula B = A T W updates B;
b. according to the formula C = (X) T AA T X+I) -1 X T AA T Updating C by X;
c. and circularly updating the subspace projection matrix A according to the following formula: a. The t+1 =qf(A t + G) until convergence;
wherein, I is an identity matrix, t represents the t-th iteration, A t Represents the value of A in the t-th iteration, A t+1 Represents the value of A in the t +1 th iteration, G represents the gradient of the objective function, and G = -2 (X (alpha L)+θ(I-C)(I-C) T )X T A-βWB T ) And qf (·) denotes QR decomposition.
In step 5), the construction process of the graph is as follows: the similarity matrix is jointly learned from two aspects of the sample label space and the sample subspace, and an objective function for defining the similarity matrix learning is as follows:
wherein tr (-) is the trace of the matrix,the F-norm of the matrix is represented,is a regression matrix of the measured values,is a sub-space projection matrix of the image,is a matrix of samples of the sample to be sampled,a matrix of real numbers representing d rows and n columns,is a matrix of the degree of similarity of the images,is a laplacian matrix and L = D-S, D is a diagonal matrix,D ii representing the elements of matrix D at row i and column i, S ij Representing the elements in the ith row and the jth column of the similarity matrix S; the parameter λ is the weight of the regularization term;
let the number of neighbors per sample be k, i.e. perThe similarity between each sample and k adjacent samples is not 0, and the similarity is 0; let x be i ,x j Respectively representing the ith and jth samples; definition e ij Is x i And x j Sum of Euclidean distance in subspace and Euclidean distance in tag space, then e ij The calculation formula of (a) is as follows:
then, according to the solution of the objective function, an update formula of the similarity matrix S can be obtained:
In step 6), the learning process of the semi-supervised linear regression classifier is as follows:
the basic objective function that defines a semi-supervised linear regression classifier is:
where tr (-) is the trace of the matrix,the F-norm of the matrix is represented,is a regression matrix of the measured values,is a matrix of samples of the sample to be sampled,a matrix of real numbers representing d rows and n columns,is a vector of the offset to be used,is the label matrix of the sample, the parameter gamma is the regularizing term weight,is a diagonal matrix if sample x i If it is a sample with a label, U ii =1, otherwise U ii =0, wherein U ii Represents the element in the ith row and ith column of the matrix U;
combining the objective function with the objective function learned by the subspace in the step 4) and the objective function learned by the similarity matrix in the step 5) to obtain a final objective function:
wherein Loss = tr ((W) T X+b1 T -Y)U(W T X+b1 T -Y) T ),Is a matrix of parameters that is,is a projection matrix of the subspace,is a similarity matrix;is a low-rank decomposition matrix of W,is a laplacian matrix and L = D-S, D is a diagonal matrix,D ii representing elements of matrix D at row i and column i, S ij Representing the elements in the ith row and the jth column of the similarity matrix S; the parameters alpha, theta and beta are weights for adjusting the importance degree of each item;
and respectively solving the partial derivatives of the final objective function to W and b to obtain the update formulas of W and b as follows:
W=[XU c X T +αXLX T +β(I-AA T )+γI] -1 XU c Y T
And then, updating W and b according to the updating formula, and finishing the learning process of the semi-supervised linear regression classifier.
Compared with the prior art, the invention has the following advantages and beneficial effects:
the method has a solid mathematical theoretical basis and has great advantages of accuracy, stability and robustness. First, constructing a graph from two low-dimensional spaces, the label space and the subspace, together, can overcome the effects of redundant features in high-dimensional data in the manufacturing industry. And moreover, the graph constructed from the two spaces has higher robustness and better adapts to the characteristic of unstable data distribution. Second, in the process of learning the subspace, the low-rank property of the regression matrix is utilized, so that the subspace can distinguish different types of samples more easily. And thirdly, integrating the three processes of subspace learning, graph construction and classifier training into a unified framework, and achieving a joint optimal solution by means of cyclic alternating optimization and mutual promotion of the three processes, thereby remarkably improving the overall learning capability of the algorithm framework.
Drawings
FIG. 1 is a logic flow diagram of the present invention.
FIG. 2 is a comparison table of the accuracy of the present invention compared with the conventional semi-supervised classification algorithm and the semi-supervised classification algorithm based on the graph, SSCNGC is the abbreviation of the method of the present invention, the number thickening is the best effect, and the data format is "accuracy + -standard deviation".
Detailed Description
The present invention will be described in further detail with reference to examples and drawings, but the embodiments of the present invention are not limited thereto.
As shown in fig. 1, the semi-supervised classification method for high-dimensional data provided by this embodiment includes the following steps:
1) And inputting a training data set which is a high-dimensional data set.
2) Normalizing the data, eliminating the influence of different characteristic dimensions, and simultaneously improving the speed of subsequent optimization learning; wherein, the data normalization step is: obtaining the maximum value X (r) corresponding to the r row data max And minimum X (r) min And converting the d-th row of data according to the following formula:
wherein the content of the first and second substances,for the ith data of the r-th row,for the data after update, n is the number of samples in the dataset, d is the number of features of the samples, i ∈ {1, 2.., n }, r ∈ {1, 2.., d }.
3) Initializing a regression matrixSubspace projection matrixWherein d is the characteristic number of the sample, and c is the sample classThe number of the first and second groups is counted,a real matrix representing d rows and c columns; initializing a low rank decomposition matrix of WWhereinA real number matrix representing c rows and c columns; initializing a similarity matrixParameter matrixWhere n is the number of samples in the sample,a real number matrix representing n rows and n columns; initializing a bias vectorWhereinA real matrix representing c rows and 1 columns;
the initialization method comprises the following steps: initializing a regression matrixIs an all-zero matrix; initializing a low rank decomposition matrix of WIs an all-zero matrix; initializing a similarity matrixAnd parameter matrixIs an all-zero matrix;initializing a bias vectorIs an all-zero vector; initializing subspace projection matrices a = qf (R) as orthogonal matrices; wherein the content of the first and second substances,is a random matrix with each element in the interval 0,1]And qf (·) denotes QR decomposition.
4) And (3) subspace learning: deriving an optimal solution of a low-rank decomposition matrix B, a parameter matrix C and a subspace projection matrix A according to the proposed subspace learning objective function; because the proposed objective function relates to a plurality of optimization variables, B, C and A are iteratively updated by an alternate optimization method, the optimization is carried out step by step, the subspace quality is improved, and the optimal subspace for expressing the essential characteristics of the sample is learned;
the subspace learning process is as follows:
the objective function defining the subspace learning is:
wherein tr (-) is the trace of the matrix,the F-norm of the matrix is represented,is a sample matrix; alpha, theta, beta are adjustable parameters;
the target function is subjected to partial derivatives of B, C and A respectively, and an updating formula of each variable can be obtained; then, each variable is updated as required:
a. according to the formula B = A T W updates B;
b. according to the formula C = (X) T AA T X+I) -1 X T AA T Updating C by X;
c. cyclically updating the subspaces according to the following formulaInter-projection matrix a: a. The t+1 =qf(A t + G) until convergence;
wherein, I is an identity matrix, t represents the t-th iteration, A t Represents the value of A in the t-th iteration, A t+1 Represents the value of A in the t +1 th iteration, G represents the gradient of the objective function, G = -2 (X (alpha L + theta (I-C) T )X T A-βWB T ) And qf (·) denotes QR decomposition.
5) Comprehensively learning a sample similarity matrix from two aspects of a sample subspace and a sample label space; defining samples as nodes of the graph, defining the similarity among the samples as edges of the graph, wherein the learning process of the sample similarity matrix is the construction process of the graph;
the construction process of the graph is as follows: the similarity matrix is jointly learned from two aspects of the sample label space and the sample subspace, and an objective function for defining the similarity matrix learning is as follows:
wherein the parameter λ is a weight of the regularization term;
setting the number of neighbors of each sample as k, namely, the similarity between each sample and the k neighbor samples is not 0, and the similarity is 0; let x i ,x j Respectively representing the ith and jth samples; definition e ij Is x i And x j Sum of Euclidean distance in subspace and Euclidean distance in tag space, then e ij The calculation formula of (a) is as follows:
then, according to the solution of the objective function, an update formula of the similarity matrix S can be obtained:
6) Learning a semi-supervised linear regression classifier, namely learning a regression matrix W and a bias vector b, on the basis of the subspace learning in the step 4) and the similarity matrix learning in the step 5);
the learning process of the semi-supervised linear regression classifier is as follows:
the basic objective function that defines a semi-supervised linear regression classifier is:
wherein the content of the first and second substances,is a diagonal matrix if sample x i Is a labeled sample, then U ii =1, otherwise U ii =0, wherein U ii Represents the element in the ith row and ith column of the matrix U;
combining the objective function with the objective function learned by the subspace in the step 4) and the objective function learned by the similarity matrix in the step 5) to obtain a final objective function:
wherein, loss = tr ((W) T X+b1 T -Y)U(W T X+b1 T -Y) T ) (ii) a The parameters alpha, theta and beta are weights for adjusting the importance degree of each item;
and respectively solving the partial derivatives of the final objective function to W and b to obtain the update formulas of W and b as follows:
W=[XU c X T +αXLX T +β(I-AA T )+γI] -1 XU c Y T
And then, updating W and b according to the updating formula, and finishing the learning process of the semi-supervised linear regression classifier.
7) Circularly performing the step 4) to the step 6), and iteratively learning each variable until convergence; when convergence occurs, the joint optimal solution is obtained through the three processes of subspace learning, graph construction and classifier learning.
8) Classifying the test samples, and assuming that the input test sample is x, predicting (x) the prediction label as:
wherein (W) T x+b) i Represents a vector (W) T x + b) th element.
9) Calculating the classification accuracy: and inputting a label of the test sample, comparing the label with a prediction result, and calculating the final classification accuracy.
Fig. 2 is an accuracy comparison table between the present invention and the conventional semi-supervised classification algorithm and the semi-supervised classification algorithm based on the graph, the SSCNGC is the abbreviation of the method of the present invention, the figure thickening is the best effect, and the data format is "accuracy ± standard deviation". As can be seen from the figure, in the experiment of 16 high-dimensional data sets, the invention achieves the highest accuracy on 15 data sets and achieves the improvement of more than 5% on 9 data sets, which shows that the invention has stronger superiority compared with the traditional semi-supervised algorithm.
The above embodiments are preferred embodiments of the present invention, but the present invention is not limited to the above embodiments, and any other changes, modifications, substitutions, combinations, and simplifications which do not depart from the spirit and principle of the present invention should be construed as equivalents thereof, and all such changes, modifications, substitutions, combinations, and simplifications are intended to be included in the scope of the present invention.
Claims (5)
1. A semi-supervised classification method for high-dimensional data is characterized by comprising the following steps:
1) Inputting a training data set which is a high-dimensional data set;
2) Normalizing the data, eliminating the influence of different characteristic dimensions, and simultaneously improving the speed of subsequent optimization learning;
3) Initializing a regression matrixSubspace projection matrixWhere d is the number of features of the sample, c is the number of sample classes,a real matrix representing d rows and c columns; initializing a low rank decomposition matrix of WWhereinA real number matrix representing c rows and c columns; initializing a similarity matrixParameter matrixWhere n is the number of samples in the sample,a real number matrix representing n rows and n columns; initialization bias directionMeasurement ofWhereinA real matrix representing c rows and 1 columns;
4) Subspace learning: deriving an optimal solution of a low-rank decomposition matrix B, a parameter matrix C and a subspace projection matrix A according to the proposed subspace learning objective function; because the proposed objective function relates to a plurality of optimization variables, B, C and A are iteratively updated by an alternate optimization method, the optimization is carried out step by step, the subspace quality is improved, and the optimal subspace for expressing the essential characteristics of the sample is learned; the subspace learning process is as follows:
the objective function defining the subspace learning is:
wherein tr (-) is the trace of the matrix,the F-norm of the matrix is represented,in the form of a matrix of samples,a matrix of real numbers representing d rows and n columns,in the form of a regression matrix,for the projection matrix of the subspace,is a low-rank decomposition matrix of W,is a parameter matrix; alpha, theta, beta are adjustable parameters;
the target function is subjected to partial derivatives of B, C and A respectively, and an updating formula of each variable can be obtained; next, each variable is updated as required:
a. according to the formula B = A T W updates B;
b. according to the formula C = (X) T AA T X+I) -1 X T AA T Updating C by X;
c. circularly updating the subspace projection matrix A according to the following formula: a. The t+1 =qf(A t + G) until convergence;
wherein, I is an identity matrix, t represents the t-th iteration, A t Represents the value of A in the t-th iteration, A t+1 Denotes the value of A in the t +1 th iteration, G denotes the gradient of the objective function, G = -2 (X (. Alpha.L + theta (I-C)) T )X T A-βWB T ) Qf (·) denotes QR decomposition;
5) Comprehensively learning a sample similarity matrix from two aspects of a sample subspace and a sample label space; defining samples as nodes of the graph, defining the similarity among the samples as edges of the graph, wherein the learning process of the sample similarity matrix is the construction process of the graph;
6) Learning a semi-supervised linear regression classifier, namely learning a regression matrix W and a bias vector b, on the basis of the subspace learning in the step 4) and the similarity matrix learning in the step 5);
7) Circularly performing the step 4) to the step 6), and iteratively learning each variable until convergence; when convergence occurs, joint optimal solutions are obtained through three processes of subspace learning, graph construction and classifier learning;
8) Classifying the test samples, assuming that the input test sample is x and the number of sample classes is c, predicting (x) of the prediction label is as follows:
wherein (W) T x+b) i Represents a vector (W) T x + b) th element;
9) Calculating the classification accuracy: and inputting labels of the test samples, comparing the labels with the prediction result, and calculating the final classification accuracy.
2. The semi-supervised classification method for high-dimensional data as recited in claim 1, wherein: in step 2), the data normalization step is: obtaining the maximum value X (r) corresponding to the r row data max And minimum X (r) min And converting the d-th row of data according to the following formula:
3. The semi-supervised classification method for high-dimensional data as recited in claim 1, wherein: in step 3), the initialization method is as follows: initializing a regression matrixIs an all-zero matrix; initializing a low rank decomposition matrix of WIs an all-zero matrix; initializing a similarity matrixAnd parameter matrixIs an all-zero matrix; initializing a bias vectorIs an all-zero vector; initializing subspace projection matrices a = qf (R) as orthogonal matrices; wherein the content of the first and second substances,is a random matrix with each element in the interval 0,1]And qf (·) denotes QR decomposition.
4. The semi-supervised classification method for high-dimensional data as recited in claim 1, wherein: in step 5), the construction process of the graph is as follows: the similarity matrix is jointly learned from two aspects of the sample label space and the sample subspace, and an objective function for defining the similarity matrix learning is as follows:
wherein tr (-) is the trace of the matrix,the F-norm of the matrix is represented,is a regression matrix, and the regression matrix is,is a subspace projectionThe shadow matrix is a matrix of the image,is a matrix of samples of the sample to be sampled,a matrix of real numbers representing d rows and n columns,is a matrix of the degree of similarity (or similarity matrix),is a laplacian matrix and L = D-S, D is a diagonal matrix,D ii representing the elements of matrix D at row i and column i, S ij Representing the elements in the ith row and the jth column of the similarity matrix S; the parameter λ is the weight of the regularization term;
setting the number of neighbors of each sample as k, namely, the similarity between each sample and the k neighbor samples is not 0, and the similarity is 0; let x i ,x j Respectively representing the ith and the j th samples; definition e ij Is x i And x j Sum of Euclidean distance in subspace and Euclidean distance in tag space, then e ij The calculation formula of (c) is as follows:
then, according to the solution of the objective function, an update formula of the similarity matrix S can be obtained:
5. The semi-supervised classification method for high-dimensional data as recited in claim 1, wherein: in step 6), the learning process of the semi-supervised linear regression classifier is as follows:
the basic objective function that defines a semi-supervised linear regression classifier is:
where tr (-) is the trace of the matrix,the F-norm of the matrix is represented,is a regression matrix, and the regression matrix is,is a matrix of samples of the sample to be sampled,a matrix of real numbers representing d rows and n columns,is a vector of the offset to the offset,is the label matrix of the sample, the parameter gamma is the regularization term weight,is a diagonal matrix if sample x i If it is a sample with a label, U ii =1, otherwise U ii =0, wherein U ii Represents the matrix U at the ith row and ith columnThe element (b);
combining the objective function with the objective function learned by the subspace in the step 4) and the objective function learned by the similarity matrix in the step 5) to obtain a final objective function:
wherein Loss = tr ((W) T X+b1 T -Y)U(W T X+b1 T -Y) T ),Is a matrix of parameters that is a function of,is a projection matrix of the subspace,is a similarity matrix;is a low-rank decomposition matrix of W,is a laplacian matrix and L = D-S, D is a diagonal matrix,D ii representing the elements of matrix D at row i and column i, S ij Representing the elements in the ith row and the jth column of the similarity matrix S; the parameters alpha, theta and beta are weights for adjusting the importance degree of each item;
and respectively solving the partial derivatives of the final objective function to W and b to obtain the update formulas of W and b as follows:
W=[XU c X T +αXLX T +β(I-AA T )+γI] -1 XU c Y T
And then, updating W and b according to the updating formula, and finishing the learning process of the semi-supervised linear regression classifier.
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110285595.XA CN113033641B (en) | 2021-03-17 | 2021-03-17 | Semi-supervised classification method for high-dimensional data |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202110285595.XA CN113033641B (en) | 2021-03-17 | 2021-03-17 | Semi-supervised classification method for high-dimensional data |
Publications (2)
Publication Number | Publication Date |
---|---|
CN113033641A CN113033641A (en) | 2021-06-25 |
CN113033641B true CN113033641B (en) | 2022-12-16 |
Family
ID=76471055
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202110285595.XA Active CN113033641B (en) | 2021-03-17 | 2021-03-17 | Semi-supervised classification method for high-dimensional data |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN113033641B (en) |
Families Citing this family (1)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN114841214B (en) * | 2022-05-18 | 2023-06-02 | 杭州电子科技大学 | Pulse data classification method and device based on semi-supervised discrimination projection |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015167526A1 (en) * | 2014-04-30 | 2015-11-05 | Hewlett-Packard Development Company, L.P | Facilitating interpretation of high-dimensional data clusters |
CN106778832A (en) * | 2016-11-28 | 2017-05-31 | 华南理工大学 | The semi-supervised Ensemble classifier method of high dimensional data based on multiple-objection optimization |
CN111027582A (en) * | 2019-09-20 | 2020-04-17 | 哈尔滨理工大学 | Semi-supervised feature subspace learning method and device based on low-rank graph learning |
CN112232438A (en) * | 2020-11-05 | 2021-01-15 | 华东理工大学 | High-dimensional image representation-oriented multi-kernel subspace learning framework |
Family Cites Families (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN102968639A (en) * | 2012-09-28 | 2013-03-13 | 武汉科技大学 | Semi-supervised image clustering subspace learning algorithm based on local linear regression |
CN110717354B (en) * | 2018-07-11 | 2023-05-12 | 哈尔滨工业大学 | Super-pixel classification method based on semi-supervised K-SVD and multi-scale sparse representation |
CN109784392B (en) * | 2019-01-07 | 2020-12-22 | 华南理工大学 | Hyperspectral image semi-supervised classification method based on comprehensive confidence |
-
2021
- 2021-03-17 CN CN202110285595.XA patent/CN113033641B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
WO2015167526A1 (en) * | 2014-04-30 | 2015-11-05 | Hewlett-Packard Development Company, L.P | Facilitating interpretation of high-dimensional data clusters |
CN106778832A (en) * | 2016-11-28 | 2017-05-31 | 华南理工大学 | The semi-supervised Ensemble classifier method of high dimensional data based on multiple-objection optimization |
CN111027582A (en) * | 2019-09-20 | 2020-04-17 | 哈尔滨理工大学 | Semi-supervised feature subspace learning method and device based on low-rank graph learning |
CN112232438A (en) * | 2020-11-05 | 2021-01-15 | 华东理工大学 | High-dimensional image representation-oriented multi-kernel subspace learning framework |
Non-Patent Citations (1)
Title |
---|
自适应半监督集成分类算法在高维数据上的研究;张乙东;《中国优秀硕士学位论文全文数据库信息科技辑》;20200131;全文 * |
Also Published As
Publication number | Publication date |
---|---|
CN113033641A (en) | 2021-06-25 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
CN111583263B (en) | Point cloud segmentation method based on joint dynamic graph convolution | |
WO2021003951A1 (en) | Hyperspectral image classification method based on label-constrained elastic network graph model | |
Liu et al. | Composite neighbor-aware convolutional metric networks for hyperspectral image classification | |
Liu et al. | Group collaborative representation for image set classification | |
CN110263855B (en) | Method for classifying images by utilizing common-basis capsule projection | |
CN115410088B (en) | Hyperspectral image field self-adaption method based on virtual classifier | |
CN111027636B (en) | Unsupervised feature selection method and system based on multi-label learning | |
Bahrami et al. | Joint auto-weighted graph fusion and scalable semi-supervised learning | |
Bi et al. | Beyond mahalanobis metric: cayley-klein metric learning | |
Gong et al. | Deep manifold embedding for hyperspectral image classification | |
Tao et al. | RDEC: integrating regularization into deep embedded clustering for imbalanced datasets | |
CN113033641B (en) | Semi-supervised classification method for high-dimensional data | |
CN116645579A (en) | Feature fusion method based on heterogeneous graph attention mechanism | |
CN111259938B (en) | Manifold learning and gradient lifting model-based image multi-label classification method | |
You et al. | Robust structure low-rank representation in latent space | |
Chen et al. | Semi-supervised convolutional neural networks with label propagation for image classification | |
Wang et al. | Domain adaptation network based on hypergraph regularized denoising autoencoder | |
Liu et al. | Inconsistency distillation for consistency: Enhancing multi-view clustering via mutual contrastive teacher-student leaning | |
CN114419382A (en) | Method and system for embedding picture of unsupervised multi-view image | |
Guo et al. | Double manifolds regularized non-negative matrix factorization for data representation | |
Nikolaou et al. | Margin maximization as lossless maximal compression | |
CN111897988A (en) | Hyperspectral remote sensing image classification method and system | |
Du et al. | Robust spectral clustering via matrix aggregation | |
CN112766354A (en) | Knowledge graph-based small sample picture identification method and system | |
Zhao et al. | Robust graph convolutional clustering with adaptive graph learning |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |