CN112257739A - Sparse representation classification method based on disturbance compressed sensing - Google Patents
Sparse representation classification method based on disturbance compressed sensing Download PDFInfo
- Publication number
- CN112257739A CN112257739A CN202010890870.6A CN202010890870A CN112257739A CN 112257739 A CN112257739 A CN 112257739A CN 202010890870 A CN202010890870 A CN 202010890870A CN 112257739 A CN112257739 A CN 112257739A
- Authority
- CN
- China
- Prior art keywords
- class
- obtaining
- sparse representation
- disturbance
- sparse
- Prior art date
- Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
- Granted
Links
- 238000000034 method Methods 0.000 title claims abstract description 22
- 238000012549 training Methods 0.000 claims abstract description 68
- 238000012360 testing method Methods 0.000 claims abstract description 48
- 239000011159 matrix material Substances 0.000 claims abstract description 40
- 239000013598 vector Substances 0.000 claims abstract description 26
- 238000005457 optimization Methods 0.000 claims abstract description 13
- 238000004364 calculation method Methods 0.000 claims description 15
- 238000012937 correction Methods 0.000 claims description 13
- 239000000126 substance Substances 0.000 claims description 5
- 238000004422 calculation algorithm Methods 0.000 description 36
- 230000007246 mechanism Effects 0.000 description 15
- 238000005516 engineering process Methods 0.000 description 7
- 238000013461 design Methods 0.000 description 6
- 238000005259 measurement Methods 0.000 description 5
- 238000010586 diagram Methods 0.000 description 3
- 238000002474 experimental method Methods 0.000 description 3
- 230000006835 compression Effects 0.000 description 2
- 238000007906 compression Methods 0.000 description 2
- 230000001419 dependent effect Effects 0.000 description 2
- 238000012986 modification Methods 0.000 description 2
- 230000004048 modification Effects 0.000 description 2
- 230000008447 perception Effects 0.000 description 2
- 241000197727 Euscorpius alpha Species 0.000 description 1
- 230000009286 beneficial effect Effects 0.000 description 1
- 230000007547 defect Effects 0.000 description 1
- 238000009795 derivation Methods 0.000 description 1
- 239000000428 dust Substances 0.000 description 1
- 238000010801 machine learning Methods 0.000 description 1
- 238000013178 mathematical model Methods 0.000 description 1
- 230000008569 process Effects 0.000 description 1
- 238000012545 processing Methods 0.000 description 1
- 230000009467 reduction Effects 0.000 description 1
- 238000007670 refining Methods 0.000 description 1
- 230000000717 retained effect Effects 0.000 description 1
- 230000009466 transformation Effects 0.000 description 1
Images
Classifications
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/24—Classification techniques
- G06F18/241—Classification techniques relating to the classification model, e.g. parametric or non-parametric approaches
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/213—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods
- G06F18/2136—Feature extraction, e.g. by transforming the feature space; Summarisation; Mappings, e.g. subspace methods based on sparsity criteria, e.g. with an overcomplete basis
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/21—Design or setup of recognition systems or techniques; Extraction of features in feature space; Blind source separation
- G06F18/214—Generating training patterns; Bootstrap methods, e.g. bagging or boosting
-
- G—PHYSICS
- G06—COMPUTING; CALCULATING OR COUNTING
- G06F—ELECTRIC DIGITAL DATA PROCESSING
- G06F18/00—Pattern recognition
- G06F18/20—Analysing
- G06F18/28—Determining representative reference patterns, e.g. by averaging or distorting; Generating dictionaries
Landscapes
- Engineering & Computer Science (AREA)
- Data Mining & Analysis (AREA)
- Theoretical Computer Science (AREA)
- Computer Vision & Pattern Recognition (AREA)
- Artificial Intelligence (AREA)
- Bioinformatics & Computational Biology (AREA)
- Bioinformatics & Cheminformatics (AREA)
- Evolutionary Biology (AREA)
- Evolutionary Computation (AREA)
- Physics & Mathematics (AREA)
- General Engineering & Computer Science (AREA)
- General Physics & Mathematics (AREA)
- Life Sciences & Earth Sciences (AREA)
- Information Retrieval, Db Structures And Fs Structures Therefor (AREA)
Abstract
The invention discloses a sparse representation classification method based on disturbance compressed sensing, which comprises the following steps: obtaining a sparse representation model, obtaining a disturbance dictionary corresponding to each category of test samples, solving an optimization problem to obtain sparse vectors, obtaining residual errors corresponding to each category of test samples, and obtaining a category with the minimum residual error. The existing sparse representation method mostly assumes that a training sample can linearly represent a test sample, but the assumption is often not true in a practical system. Aiming at the problem, the sparse representation is classified and modeled into a disturbance compressed sensing model, and a better dictionary matrix and a sparse coefficient vector are obtained by solving a disturbance reconstruction problem, so that the classification accuracy is improved.
Description
Technical Field
The invention relates to the technical field of perception, in particular to a sparse representation classification method based on perturbation compressed perception.
Background
The basic idea of Sparse Representation Classification (SRC) is: test samples are represented using training samples and representation sparsity is constrained using an L0 sparse norm or an L1 convex norm. At present, sparse representation classification schemes are widely applied to scenes such as handwriting recognition and face recognition, and are expanded to machine learning technologies such as semi-supervised classification and dimension reduction. The mathematical model of the sparse representation classification method is
y=Dα
Wherein y ∈ RMFor the test specimens, D ∈ RM×NFor the matrix formed by the training samples, x ∈ iNα∈RMThe vector is sparsely represented, i.e. alpha has at most only K non-zero values.
In an actual sparse representation classification system, training samples and test samples used by the system are inevitably affected by various non-ideal factors, such as gaussian noise necessarily contained in the samples, light, dust and occlusion in image samples, and even slight distortion of the samples caused by hardware problems. If the influence of the factors is not considered in the design of the sparse representation classification system, the accuracy of the sparse representation classification system in practical application is reduced.
Disclosure of Invention
In order to solve the limitations and defects in the prior art, the invention provides a sparse representation classification method based on perturbation compressed sensing, which comprises the following steps:
obtaining a sparse representation model according to a test sample input by a system, wherein a calculation formula of the sparse representation model is as follows:
y=(D+E)α
wherein E is a disturbance matrix representing the known matrix D and the ideal dictionary matrixThe error between;
obtaining a disturbance dictionary corresponding to the test sample of each category, wherein the calculation formula is as follows:
Hl:yl=(D+El)αl
wherein, y1For class 1 test specimens, E1Disturbance dictionary, α, for class 1 test samples1Sparse vectors corresponding to the type 1 test samples;
solving the following optimization problem to obtain sparse vectors:
obtaining the residual error corresponding to each category of test samples, wherein the calculation formula is as follows:
obtain class min r with minimum residualj(y)||2。
Specifically, the corresponding solving step includes:
Step 102: obtaining training sample with highest correlationAdding the training sample into a candidate support set T ═ Tubui;
step 103: for each candidateClass j, selecting all training samples T belonging to class j in the supporting setjAnd corresponding coefficientCalculating residual error r corresponding to j-th classj=y-D[Tj]xj;
Step 105: obtaining a correction parameter corresponding to the category with the minimum residual error, wherein the calculation formula is as follows:
step 106: obtaining an optimized dictionary of classes with minimum residuals, the calculation formula is as follows:
Step 108: if K is K, go to step 109, otherwise, reset K to K +1 and return to step 102;
step 109: obtaining coefficients of support set positionThe rest positions are set to zero x [ T ]c]=0。
Optionally, the step 102 includes:
obtain the most relevantHigh set of S training samplesAnd adding the training sample into a candidate support set T ═ Turi.
Optionally, the method further includes:
Step 202: obtaining K training sample sets with highest correlationAdding the training sample set into a candidate support set
Step 203: calculating coefficientsSelecting the training samples with the first K maximum coefficients to form a candidate support set
Step 204: for each candidate category j, selecting all training samples T belonging to the jth category in the support setjAnd training sample T of class jjCorresponding coefficientComputing class j training samples TjCorresponding residual error rj=y-D[Tj]xj;
Step 206: computing class j training samples TjCorresponding correctionParameters, the calculation formula is as follows:
step 207: computing class j training samples TjThe calculation formula of the optimized dictionary is as follows:
Step 209: if rk||2≥||rk-1||2Executing step 2010, otherwise, returning to step 202;
step 2010: calculating coefficients for support set positionsThe rest positions are set to zero x [ T ]c]=0。
The invention has the following beneficial effects:
the sparse representation classification method based on the perturbation compressed sensing provided by the embodiment comprises the following steps: obtaining a sparse representation model, obtaining a disturbance dictionary corresponding to the test sample of each category, solving an optimization problem to obtain a sparse vector, obtaining a residual error corresponding to the test sample of each category, and obtaining the category with the minimum residual error.
In the embodiment, a better dictionary matrix and a better sparse coefficient vector are obtained by solving the disturbance reconstruction problem, so that the classification accuracy is improved. Even if there is disturbance in the training sample used, the technical scheme provided by the embodiment can obtain excellent performance, good robustness and wide application range.
Drawings
Fig. 1 is a schematic diagram illustrating a relationship between classification accuracy of CDPGOMPs and sample length according to an embodiment of the present invention.
Fig. 2 is a classification accuracy chart of the cdsps algorithm compared with the OMP algorithm and the SP algorithm according to an embodiment of the present invention.
Detailed Description
In order to make those skilled in the art better understand the technical solution of the present invention, the following describes in detail a sparse representation classification method based on perturbation compressive sensing provided by the present invention with reference to the accompanying drawings.
Example one
Since the technical solution provided by the present embodiment is designed based on the perturbation compressive sensing technology, the related background of perturbation compressive sensing needs to be introduced. Compressive sensing is a milestone theory applied to signal processing that allows compression to be done while the signal is being acquired. Specifically, the compressed sensing technology obtains compressed low-dimensional data by projecting a signal by using a random measurement matrix, and ensures that an original signal can be correctly reconstructed from the low-dimensional compressed data.
In a sense, the compressed sensing theory is to sample information rather than data, and the dimension of a compressed sample is determined by the structure and content of the information in the signal on the premise of ensuring that the original signal can be reconstructed. The compressed sensing technology faces the following problems in practical application. Since the compressed sensing technology uses special hardware to directly obtain compressed data, and some unknown non-ideal factors are often included in a hardware system, the signal process experienced by the compressed data often has a slight difference from the original design of the system.
That is, there is often inconsistency between the measurement matrix preset by the system and the unknown measurement matrix actually used, which is a problem of disturbance of the measurement matrix. And the compressed sensing problem under the disturbance condition is also called as the disturbance compressed sensing problem.
The theoretical basis of the classification scheme based on sparse representation isThere is a high degree of similarity between samples belonging to the same class. Therefore, the test sample should be linearly representable by training samples belonging to the same class. Suppose D ═ D in a given training set0,D1,...,Dj,...,DL-1]Has L classes of data, and a training data set of each classIncludes Q training samples corresponding to the category
Conventional sparse representation classification schemes attempt to linearly (sparsely) represent test samples by selecting a combination of K training samples. And the K training samples come from a training dictionary D consisting of all training samples belonging to different classes. The class label of the test sample y is given by the set of samples D that yields the lowest reconstruction error (residual)jThe category to which it belongs.
In an ideal case, if the class of a known test sample y is j, y can be approximately represented as
Wherein alpha isjIs a jth class training sample DjThe coefficient of (a). In the actual scenario, the correct label to which the test sample y belongs is unknown. Thus, for a test sample y of an unknown class, the traditional sparse representation classification design idea is to use all training samples in the dictionary D for representation, i.e.
y≈Dα (1)
Wherein the content of the first and second substances,αja coefficient vector corresponding to the jth class sample. Ideally, the coefficient vector α should be sparse, i.e., divided by the position corresponding to the training samples belonging to the same class as the test sample toIn addition, all other terms are zero. Thus, the sparse solution α in the formula can be obtained by solving the following optimization problem:
where the L0 norm is defined as the number of non-zero elements in the vector α.
However, in practical scenarios, the sparse vector α obtained by the system often contains non-zero elements belonging to a plurality of different classes of training samples. Therefore, the sparse representation classification system constructs a corresponding sparse vector containing only the non-zero elements of the class training sample for each class. That is, for the j-th class, only the elements of H ≦ Q training samples belonging to the class are retained, and the corresponding sparse coefficient vector form is
The mark alphaj={αj,1,αj,2,...,αj,HIs the coefficient vector corresponding to the jth class sample, and the corresponding training sample is Dj=[d1,d2,...,dH]Then the residual error corresponding to the jth class is defined as
rj(y)=||y-Djαj||2,j=0,1,2,...,L-1
Finally, the class of the test sample y is then determined to be the class with the smallest residual error, i.e. the class with the smallest residual error
label(y)=arg minj{||rj(y)||2}
It can be seen that the core of the sparse representation classification scheme lies in the sparse representation model formed by the formula (1) and the corresponding optimization problem (formula (2)) and the solving algorithm. At present, various methods are available to find an approximate solution to the problem, for example, reconstruction algorithms based on an orthogonal matching pursuit algorithm, a generalized orthogonal matching pursuit algorithm, a subspace pursuit algorithm, and the like.
Conventional sparse representation classification schemes assume that test samples are located in a subspace formed by training samples belonging to the same class, i.e., any test sample can be linearly represented by a training sample in the same class. However, in practical applications, due to the influence of non-ideal factors such as noise in samples, most of test samples cannot be completely represented linearly by training samples, which causes the decrease of the correlation between the sparse coefficient vector α and the test sample y, and further causes the poor classification performance of the sparse representation classification scheme.
The present embodiment is based on the problem that a more optimal dictionary matrix is designed and used for representing the test sample. Dictionary matrix used in the embodimentThe method is obtained by optimizing a traditional dictionary based on a disturbance compression sensing technology, and can better represent a test sample. Thus, a high correlation is maintained between the resulting sparse coefficient vector and the test sample. Therefore, the embodiment improves the sparse representation model in the sparse representation classification system and the corresponding optimization problem thereof, and provides different specific derivation algorithms.
To more accurately represent the test sample, the present embodiment will use an optimized dictionary matrixInstead of the dictionary matrix used in conventional sparse representation classification. The dictionary matrixOn the basis of a matrix D formed by training samples, an unknown disturbance matrix E for correction is introduced to obtain a more accurate representation model, so that the classification accuracy is improved. Therefore, the key technical problem of the embodiment is to solve an optimized dictionary matrix based on the perturbation compressive sensing technology(i.e., unknown disturbance matrix E) and its corresponding sparse coefficientsAnd (5) vector quantity.
The embodiment provides a sparse representation classification method based on perturbation compressed sensing, which comprises the following steps: obtaining a sparse representation model, obtaining a disturbance dictionary corresponding to the test sample of each category, solving an optimization problem to obtain a sparse vector, obtaining a residual error corresponding to the test sample of each category, and obtaining the category with the minimum residual error.
The existing sparse representation method mostly assumes that a training sample can linearly represent a test sample, but the assumption is often not true in a practical system. For the problem, the sparse representation classification is modeled into a disturbance compressed sensing model, and a better dictionary matrix and a sparse coefficient vector are obtained by solving a disturbance reconstruction problem, so that the classification accuracy is improved.
Since most of the test samples cannot be completely represented linearly by the training samples in practical applications, the more accurate model should be formula (1) compared to formula (1)
y=Dα+y⊥
Wherein, y⊥Representing the error between the test sample y and its linear representation D α.
Since the matrix D cannot completely linearly represent the test samples, this problem necessarily affects the classification. In this regard, the present embodiment uses a more optimal dictionary matrixTo linearly represent the test specimen y. Wherein the matrix D is a known matrix formed by all training samples, and the unknown matrix E represents the known matrix D and an ideal dictionary matrixThe error between. If the embodiment can construct a more ideal dictionary matrixThis embodiment enables a more accurate linearity to be constructed between the test sample and the training sample. In general y⊥Also dependent on the sparse vector alphaThen, y is equal to D α + y⊥Transformation of representation model into
y=(D+E)α (3)
The sparse representation vector alpha is solved based on the model, and the classification accuracy can be effectively improved. It can be found that the model in the above equation is completely consistent with the model of perturbation compressed sensing. Therefore, the present embodiment will design a classification framework and a corresponding sparse solution algorithm based on the perturbation compressive sensing theory. In order to be consistent with the compressed sensing of disturbances, the matrix E will be referred to as the disturbance matrix hereinafter.
DictionaryConsists of two parts, of which only the matrix D consisting of training samples is known and the other part of the matrix E is unknown. In addition, since different classes of training data tend to be generated independently of each other, it can be assumed that the corresponding perturbations in each class are different from each other. Thus, the present embodiment designs a different perturbation dictionary for each class of test sample, i.e.
Hl:yl=(D+El)αl (4)
From the above point of view, to obtain the matrix ElAnd a sparse coefficient vector alphalThis embodiment will solve the following optimization problem:
wherein the content of the first and second substances,λ is the regularization parameter. In the embodiment, a better dictionary matrix and a better sparse coefficient vector are obtained by solving the disturbance reconstruction problem, so that the classification accuracy is improved. Thus, the present embodiment provides a framework of sparse representation classification method based on perturbation compressed sensing as shown in table 1 below.
Table 1: sparse representation classification framework based on disturbance compressed sensing
For the above classification framework, this embodiment proposes a variety of solving algorithms. First, the present embodiment proposes a Class-based perturbation Orthogonal Matching Pursuit (CDPOMP) algorithm according to a Perturbed Orthogonal Matching Pursuit (POMP) algorithm, which finds a suitable candidate support set in each loop and optimizes the dictionary matrix of each Class independently, as shown in table 2. Next, the present embodiment will explain the dictionary optimization mechanism in the CDPOMP algorithm in detail.
According to the model given in equation (4), the corresponding perturbations for each category should be different. To achieve this, the present embodiment designs a dictionary optimization mechanism in a Perturbation Orthogonal Matching Pursuit (POMP) algorithm based on a dictionary correction mechanism in the algorithm.
The original mechanism uses a complete support set consisting of training samples of multiple classes. In the mechanism improved by the present embodiment, the correction parameter in each class is calculated based only on the training samples belonging to the corresponding class in the support set. Since the training samples of each class are different from the training samples of other classes, the optimized dictionaries computed for the different classes are also different. The improved dictionary correction mechanism requires separate computation for each possible class to generate a different correction dictionary for each class in each iteration.
In order to reduce the computational complexity, the present embodiment further optimizes the mechanism. In each iteration, the corresponding correction parameters are generated only for the training samples and residuals corresponding to the class to which the test sample is most likely to belong (i.e. steps 3-6 in table 2 below). Therefore, the algorithm provided by the embodiment can ensure that the optimized dictionaries in each category are different from each other, and improve the classification performance with lower complexity.
Table 2: disturbance orthogonal matching tracking algorithm based on category
Further, in the present embodiment, based on the framework of table 1, the Generic Orthogonal Matching Pursuit (GOMP) algorithm proposes the Class-based perturbation Generalized Orthogonal Matching Pursuit (cdpgold) with better performance, as shown in table 3. In the algorithm, the support set selection mechanism is optimized, the speed of selecting the correct non-zero position is increased, and the classification performance is further improved. The performance of the CDPOMP algorithm described in table 2 is limited by the error in the dictionary optimization mechanism calculations. Only one non-zero position is added to the support set in each cycle of the CDPOMP algorithm. Thus, all training samples in the support set may all belong to the wrong category, which results in the dictionary correction mechanism inevitably working in the wrong way. In the CDPGOMP algorithm as an improved algorithm, the speed of expanding the support set is significantly increased, and the CDPGOMP algorithm contains more training samples. Therefore, there is a higher probability that training samples belonging to the correct class exist in the support set, and there is also a higher probability that the dictionary correction mechanism works based on the correct class. That is, the present embodiment effectively avoids the occurrence of a situation in which the dictionary correction mechanism operates in an erroneous manner, thereby reducing the rate of erroneous classification.
Table 3: disturbance generalized orthogonal matching tracking algorithm based on category
Finally, in the present embodiment, a Class-dependent Perturbed Subspace Pursuit (cdsps) algorithm is proposed based on a Subspace Pursuit (SP) algorithm according to the framework of table 1, as shown in table 4. The cdsps algorithm also selects and adds a number of training samples with the largest amplitude correlation into the support set in one cycle (step 2 in table 4). After that, a support set refining mechanism is introduced, and the support set is trimmed by the corresponding projection coefficients (step 3 in table 4). The mechanism can effectively eliminate a part of training samples belonging to wrong categories in the support set, thereby further reducing the possibility that the dictionary correction mechanism is incorrect in operation. Meanwhile, in order to reduce complexity, the cdsps algorithm optimizes the residual calculation modes (step 4 in table 4) of each category, and multiplexes the projection coefficients in step 3, thereby avoiding performing pseudo-inverse operation for many times and reducing the complexity of the cdsps algorithm.
Table 4: disturbance subspace tracking algorithm based on category
The sparse representation classification scheme provided by the embodiment is evaluated by taking a traditional sparse representation classification method (using an OMP/SP algorithm to solve an optimization problem) as a reference. Fig. 1 is a schematic diagram illustrating a relationship between classification accuracy of a CDPGOMP and a sample length according to a first embodiment of the present invention, and fig. 2 is a classification accuracy diagram of a cdpgsp algorithm according to a first embodiment of the present invention, compared with an OMP algorithm and an SP algorithm. As shown in fig. 1 and 2, the sample dataset of the experiment was an MNIST handwritten digital dataset comprising 70000 handwritten digital images of 28 × 28 pixels. In the experiment, the training set contained 60000 samples and the test set contained 10000 samples. The experimental results are averaged 1000 times, each time, the raw data are compressed into compressed samples with the length M through a measurement matrix randomly generated according to a standard, and 500 training samples and 100 testing samples are randomly selected from each category.
Experiments show that the performance of the CDPSP scheme provided by the embodiment is far better than that of the OMP scheme and the SP scheme, so that the sparse representation classification method based on disturbance compressed sensing provided by the embodiment obtains a better dictionary matrix and a better sparse coefficient vector by solving a disturbance reconstruction problem, and the classification accuracy is improved.
It will be understood that the above embodiments are merely exemplary embodiments taken to illustrate the principles of the present invention, which is not limited thereto. It will be apparent to those skilled in the art that various modifications and improvements can be made without departing from the spirit and substance of the invention, and these modifications and improvements are also considered to be within the scope of the invention.
Claims (4)
1. A sparse representation classification method based on perturbation compressed sensing is characterized by comprising the following steps:
obtaining a sparse representation model according to a test sample input by a system, wherein a calculation formula of the sparse representation model is as follows:
y=(D+E)α
wherein E is a disturbance matrix representing the known matrix D and the ideal dictionary matrixThe error between;
obtaining a disturbance dictionary corresponding to the test sample of each category, wherein the calculation formula is as follows:
Hl:yl=(D+El)αl
wherein, ylFor class I test specimens, ElDisturbance dictionary, α, for class I test sampleslSparse vectors corresponding to the class I test samples;
solving the following optimization problem to obtain sparse vectors:
obtaining the residual error corresponding to each category of test samples, wherein the calculation formula is as follows:
obtain class min r with minimum residualj(y)||2。
2. The sparse representation classification method based on perturbation compressed sensing of claim 1, further comprising:
Step 102: obtaining training sample with highest correlationAdding the training sample into a candidate support set T ═ Tubui;
step 103: for each candidate class j, selecting all training samples T belonging to the jth class in the support setjAnd corresponding coefficientCalculating residual error r corresponding to j-th classj=y-D[Tj]xj;
Step 105: obtaining a correction parameter corresponding to the category with the minimum residual error, wherein the calculation formula is as follows:
step 106: obtaining an optimized dictionary of classes with minimum residuals, the calculation formula is as follows:
Step 108: if K is K, go to step 109, otherwise, reset K to K +1 and return to step 102;
4. The sparse representation classification method based on perturbation compressed sensing of claim 1, further comprising:
Step 202: obtaining K training sample sets with highest correlationAdding the training sample set into a candidate support set
Step 203: calculating coefficientsSelecting the training samples with the first K maximum coefficients to form a candidate support set
Step 204: for each candidate category j, selecting all training samples T belonging to the jth category in the support setjAnd training sample T of class jjCorresponding coefficientComputing class j training samples TjCorresponding residual error rj=y-D[Tj]xj;
Step 206: computing class j training samples TjThe corresponding correction parameters are calculated according to the following formula:
step 207: computing class j training samples TjThe calculation formula of the optimized dictionary is as follows:
Step 209: if | rk‖2≥‖rk-1‖2Executing step 2010, otherwise, returning to step 202;
Priority Applications (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010890870.6A CN112257739B (en) | 2020-08-29 | 2020-08-29 | Sparse representation classification method based on disturbance compressed sensing |
Applications Claiming Priority (1)
Application Number | Priority Date | Filing Date | Title |
---|---|---|---|
CN202010890870.6A CN112257739B (en) | 2020-08-29 | 2020-08-29 | Sparse representation classification method based on disturbance compressed sensing |
Publications (2)
Publication Number | Publication Date |
---|---|
CN112257739A true CN112257739A (en) | 2021-01-22 |
CN112257739B CN112257739B (en) | 2023-12-22 |
Family
ID=74223828
Family Applications (1)
Application Number | Title | Priority Date | Filing Date |
---|---|---|---|
CN202010890870.6A Active CN112257739B (en) | 2020-08-29 | 2020-08-29 | Sparse representation classification method based on disturbance compressed sensing |
Country Status (1)
Country | Link |
---|---|
CN (1) | CN112257739B (en) |
Cited By (2)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115790815A (en) * | 2023-01-17 | 2023-03-14 | 常熟理工学院 | Method and system for rapidly identifying disturbance of distributed optical fiber sensing system |
CN116582132A (en) * | 2023-07-06 | 2023-08-11 | 广东工业大学 | Compressed sensing reconstruction method and system based on improved structured disturbance model |
Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345621A (en) * | 2013-07-09 | 2013-10-09 | 东南大学 | Face classification method based on sparse concentration index |
WO2015048232A1 (en) * | 2013-09-26 | 2015-04-02 | Tokitae Llc | Systems, devices, and methods for classification and sensor identification using enhanced sparsity |
CN104951787A (en) * | 2015-06-17 | 2015-09-30 | 江苏大学 | Power quality disturbance identification method for distinguishing dictionary learning under SRC framework |
CN105574475A (en) * | 2014-11-05 | 2016-05-11 | 华东师范大学 | Common vector dictionary based sparse representation classification method |
-
2020
- 2020-08-29 CN CN202010890870.6A patent/CN112257739B/en active Active
Patent Citations (4)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN103345621A (en) * | 2013-07-09 | 2013-10-09 | 东南大学 | Face classification method based on sparse concentration index |
WO2015048232A1 (en) * | 2013-09-26 | 2015-04-02 | Tokitae Llc | Systems, devices, and methods for classification and sensor identification using enhanced sparsity |
CN105574475A (en) * | 2014-11-05 | 2016-05-11 | 华东师范大学 | Common vector dictionary based sparse representation classification method |
CN104951787A (en) * | 2015-06-17 | 2015-09-30 | 江苏大学 | Power quality disturbance identification method for distinguishing dictionary learning under SRC framework |
Non-Patent Citations (1)
Title |
---|
沈跃;张瀚文;刘国海;刘慧;陈兆岭;: "基于判别字典学习的电能质量扰动识别方法", 仪器仪表学报, no. 10 * |
Cited By (3)
Publication number | Priority date | Publication date | Assignee | Title |
---|---|---|---|---|
CN115790815A (en) * | 2023-01-17 | 2023-03-14 | 常熟理工学院 | Method and system for rapidly identifying disturbance of distributed optical fiber sensing system |
CN116582132A (en) * | 2023-07-06 | 2023-08-11 | 广东工业大学 | Compressed sensing reconstruction method and system based on improved structured disturbance model |
CN116582132B (en) * | 2023-07-06 | 2023-10-13 | 广东工业大学 | Compressed sensing reconstruction method and system based on improved structured disturbance model |
Also Published As
Publication number | Publication date |
---|---|
CN112257739B (en) | 2023-12-22 |
Similar Documents
Publication | Publication Date | Title |
---|---|---|
Wang et al. | Balanced distribution adaptation for transfer learning | |
Aljundi et al. | Landmarks-based kernelized subspace alignment for unsupervised domain adaptation | |
Yang et al. | Subspace clustering via good neighbors | |
Zhang | A survey of unsupervised domain adaptation for visual recognition | |
Cai et al. | Rapid robust principal component analysis: CUR accelerated inexact low rank estimation | |
Ebadi et al. | Foreground segmentation with tree-structured sparse RPCA | |
CN110647912A (en) | Fine-grained image recognition method and device, computer equipment and storage medium | |
Yang et al. | Efficient and robust multiview clustering with anchor graph regularization | |
CN105528620B (en) | method and system for combined robust principal component feature learning and visual classification | |
CN110008844B (en) | KCF long-term gesture tracking method fused with SLIC algorithm | |
CN112257739B (en) | Sparse representation classification method based on disturbance compressed sensing | |
CN109117860B (en) | Image classification method based on subspace projection and dictionary learning | |
Agarwal et al. | Least squares revisited: Scalable approaches for multi-class prediction | |
Zou et al. | Quaternion block sparse representation for signal recovery and classification | |
CN109657693B (en) | Classification method based on correlation entropy and transfer learning | |
CN109815440B (en) | Dimension reduction method combining graph optimization and projection learning | |
Guo et al. | Accelerating patch-based low-rank image restoration using kd-forest and Lanczos approximation | |
CN110826554B (en) | Infrared target detection method | |
CN105389560B (en) | Figure optimization Dimensionality Reduction method based on local restriction | |
CN108492283B (en) | Hyperspectral image anomaly detection method based on band-constrained sparse representation | |
Wang et al. | Robust variable screening for regression using factor profiling | |
US11756319B2 (en) | Shift invariant loss for deep learning based image segmentation | |
Lv et al. | A robust mixed error coding method based on nonconvex sparse representation | |
Zhang et al. | Nonlinear dictionary learning based deep neural networks | |
Wu et al. | Nonconvex approach for sparse and low-rank constrained models with dual momentum |
Legal Events
Date | Code | Title | Description |
---|---|---|---|
PB01 | Publication | ||
PB01 | Publication | ||
SE01 | Entry into force of request for substantive examination | ||
SE01 | Entry into force of request for substantive examination | ||
GR01 | Patent grant | ||
GR01 | Patent grant |