CN105469101A - Mixed two-dimensional probabilistic principal component analysis method - Google Patents

Mixed two-dimensional probabilistic principal component analysis method Download PDF

Info

Publication number
CN105469101A
CN105469101A CN201511022718.1A CN201511022718A CN105469101A CN 105469101 A CN105469101 A CN 105469101A CN 201511022718 A CN201511022718 A CN 201511022718A CN 105469101 A CN105469101 A CN 105469101A
Authority
CN
China
Prior art keywords
sigma
formula
gamma
rsqb
lsqb
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Pending
Application number
CN201511022718.1A
Other languages
Chinese (zh)
Inventor
孙艳丰
刘思萌
句福娇
胡永利
尹宝才
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Beijing University of Technology
Original Assignee
Beijing University of Technology
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Beijing University of Technology filed Critical Beijing University of Technology
Priority to CN201511022718.1A priority Critical patent/CN105469101A/en
Publication of CN105469101A publication Critical patent/CN105469101A/en
Pending legal-status Critical Current

Links

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06VIMAGE OR VIDEO RECOGNITION OR UNDERSTANDING
    • G06V10/00Arrangements for image or video recognition or understanding
    • G06V10/40Extraction of image or video features
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F18/00Pattern recognition
    • G06F18/20Analysing
    • G06F18/24Classification techniques

Landscapes

  • Engineering & Computer Science (AREA)
  • Theoretical Computer Science (AREA)
  • Physics & Mathematics (AREA)
  • Data Mining & Analysis (AREA)
  • General Physics & Mathematics (AREA)
  • Artificial Intelligence (AREA)
  • Computer Vision & Pattern Recognition (AREA)
  • Evolutionary Biology (AREA)
  • Evolutionary Computation (AREA)
  • Bioinformatics & Computational Biology (AREA)
  • General Engineering & Computer Science (AREA)
  • Bioinformatics & Cheminformatics (AREA)
  • Life Sciences & Earth Sciences (AREA)
  • Multimedia (AREA)
  • Image Processing (AREA)

Abstract

The invention discloses a mixed two-dimensional probabilistic principal component analysis method, the method can perform the dimensionality reduction for the two-dimensional data in row and column directions, and the reconstruction effect is better than the same in the prior art. The method comprises the steps: (1) a probabilistic second order principal component analysis model 2DPCA is built according to a formula (1); (2) a mixed 2DPC is built according to a formula (2); (3) the parameter in formula (2) is estimated through the maximum likelihood function of a formula (4); (4) when the formula (4) is solved, the parameter is optimized through a variation EM(Expectation Maximization) algorithm, in E step of the EM algorithm, the posteriori distribution of a hidden variable is solved, M step relates to parameter updating in formula (4), the E step and the M step are performed in an iterative way, so that the likelihood function value is increased and tends to be stable.

Description

A kind of mixing Two-dimensional Probabilistic principal component analytical method
Technical field
The invention belongs to the technical field of feature extraction and Data Dimensionality Reduction, relate to a kind of mixing Two-dimensional Probabilistic principal component analytical method particularly.
Background technology
Higher-dimension and multi-modal data are seen everywhere in the research of modern computer vision.The high-dimensional of data not only increases the complicacy of algorithm and the expense of storage, and also reduces algorithm popularity in actual applications.But high dimensional data is evenly distributed on a lower dimensional space or popular world often.So, find a kind of mapping relations of higher-dimension observation data in lower dimensional space to become a challenging problem of machine learning research.In the past few decades, the algorithm about Data Dimensionality Reduction obtains remarkable progress.
Principal component analysis (PCA) (PrincipalComponentAnalysis, PCA) is a kind of dimension reduction method being widely used in pattern-recognition and machine learning.PCA has a variety of explanation, and wherein a kind of is suppose that the high dimensional data observed is a kind of linear mapping of data in lower dimensional space.The data obtained in reality often have inner dependency structure, such as, and image.In order to apply PCA on 2-D data, the most direct way is by 2-D data vectorization.But the data after vectorization are all generally high dimensional datas, so not only can cause the problem of dimension disaster but also have ignored the spatial structural form of 2-D data.So the PCA of data vector is not a kind of feature extracting method of optimum.Therefore a kind of PCA (2DPCA) of 2-D data is proposed.Compared with traditional PCA, 2DPCA directly carries out dimensionality reduction computing to 2D data matrix, therefore remains the structural relation between data, and achieves the experimental result being better than PCA.
The PCA of these non-probability only relies on raw data, does not suppose any parameter, also the priori of observation data is not applied in predicting the outcome.Not enough in order to overcome this, TippingandBishop proposes a kind of pca model (PPCA) of probability.This model is that tables of data is shown as one-dimensional vector equally, and supposes that noise obedience average is 0, and oblique variance is the Gaussian distribution of unit matrix.Compared with traditional PCA, PPCA is that the theory of applied probability reaches Data Dimensionality Reduction, and the parameter of model draws by Maximum-likelihood estimation (MLE).And then Zhaoetal. proposes the probability P CA (2DPPCA) based on 2-D data, this is a breakthrough of conventional P PCA to 2D model.2DPPCA define only the overall situation projection of sample at image space, and in order to represent the local message of sample better, Wangetal. proposes mixing probability P CA (MP2DPCA) based on human face data.
The probability distribution of one group of 2-D data is quite complicated, and in most cases it can not be distributed by a certain particular probability and represent, so need to use Gaussian mixtures to be similar to.According to Bishop, by using abundant Gaussian distribution, and regulate their average and the coefficient of variance and linear combination, nearly all continuous probability density can both be similar to arbitrary precision.The linear combination of multiple Gaussian distribution is called that mixed Gaussian (mixtureofGaussians) distributes.Rasmussen proposes gauss hybrid models (GMM), and GMM is applied to picture background and extracts by Zivkovic, and Lietal. uses mixed Gaussian regretional analysis to carry out subspace clustering.The probability distribution of second order data uses the hybrid analog-digital simulation of multiple Gaussian distribution, namely be by Second Order High Dimensional data dimensionality reduction in multiple principal direction, calculative parameter is more, therefore variation EM algorithm (VariationalExpectationMaximization, VariationalEM) can be utilized to separate this model.
But above method all can not carry out the dimensionality reduction in row and column both direction to 2-D data.
Summary of the invention
Technology of the present invention is dealt with problems and is: overcome the deficiencies in the prior art, and provide a kind of mixing Two-dimensional Probabilistic principal component analytical method, it can carry out dimensionality reduction to 2-D data in row and column both direction, and quality reconstruction is better.
Technical solution of the present invention is: this mixing Two-dimensional Probabilistic principal component analytical method, and the Gaussian mixtures of matrix variables obeyed by sample, and the method comprises the following steps:
(1) the second order principal component model 2DPCA of probability is built according to formula (1)
X n=LB nR T+M+E n;(1)
Wherein L (p × r) and R (q × c) is the dimensionality reduction matrix in row and column direction respectively; B n(r × c) is sample X nhidden variable core, be called matrix of coefficients; R≤p, c≤q is the number of lines and columns after dimensionality reduction; M (p × q) is Mean Matrix, E nthe noise meeting matrix Gaussian distribution, its each component meets
(2) 2DPCA of mixing is built according to formula (2)
The average item M be wherein separated kthe average of kth class sample set, L kand R kthe kth class dimensionality reduction that sample set obtains, π kblending ratio, it is the variance of a kth Gaussian distribution;
(3) by the maximum likelihood function of formula (4), the parameter in formula (2) is estimated
Z nkvalue is 1 or 0, represents the n-th sample and whether belongs to a kth Gaussian distribution.
(4) when solution formula (4), variation greatest hope EM algorithm is utilized to carry out Optimal Parameters in the E step of EM algorithm, to hidden variable solve its Posterior distrbutionp: wherein average, be the covariance matrix in row and column direction respectively, calculate formula (9)-(11):
T n ( k ) = cσ k 2 [ t r ( R k T R k S n ( k ) ) L k T L k + σ k 2 t r ( S n ( k ) ) I r ] - 1 - - - ( 9 )
S n ( k ) = rσ k 2 [ t r ( L k T L k T n ( k ) ) R k T R k + σ k 2 t r ( T n ( k ) ) I c ] - 1 - - - ( 10 )
v e c ( Q n ( k ) ) = [ R k T R k ⊗ L k T L k + σ k I ⊗ σ k I ] - 1 y - - - ( 11 )
Wherein y = v e c ( L k T ( X n - M k ) R k ) ;
M walks the parameter in more new formula (4), obtains formula (12)-(14):
L k = [ Σ n = 1 N γ n k ( X n - M k ) R k ( Q n ( k ) ) T ] [ Σ n = 1 N γ n k Q n ( k ) R k T R k ( Q n ( k ) ) T + t r ( R k T R k S n ( k ) ) T n ( k ) ] - 1 - - - ( 12 )
R k = [ Σ n = 1 N γ n k ( X n - M k ) T L k Q n ( k ) ] [ Σ n = 1 N γ n k ( Q n ( k ) ) T L k T L k Q n ( k ) + t r ( L k T L k T n ( k ) ) S n ( k ) ] - 1 - - - ( 13 )
&sigma; k 2 = 1 pqN k { &Sigma; n = 1 N &gamma; n k t r ( X n - M k ) T ( X n - M k ) - 2 &Sigma; n = 1 N &gamma; n k t r ( R k < B n ( k ) > T L k T ( X n - M k ) ) + &Sigma; n = 1 N &gamma; n k t r ( < B n ( k ) T L k T L k B n ( k ) > R k T R k ) } - - - ( 14 )
Wherein γ n, krepresent that the n-th sample belongs to the posterior probability of a kth Gaussian distribution;
Iteration carries out E step and M step, makes likelihood function value increase and tend towards stability.
The present invention is based on mixed Gauss model and dimensionality reduction is carried out to 2-D data, by introducing hidden variable, use variation EM algorithm solving model parameter, and the matrix of coefficients after dimensionality reduction, reach the effect that 2-D data is compressed, use dimensionality reduction matrix and matrix of coefficients reconstructed image, compare with original image and obtain losing less image, matrix of coefficients after dimensionality reduction is regarded as the feature of sample, coefficient of performance matrix is classified to sample, so can carry out dimensionality reduction to 2-D data in row and column both direction, quality reconstruction is better.
Accompanying drawing explanation
Fig. 1 shows the reconstructed error of algorithms of different on MNIST database, K=10 in K=5, Fig. 1 c in K=2, Fig. 1 b in Fig. 1 a.
Fig. 2 a is the reconstructed error curve map on Yale database, and Fig. 2 b is the reconstructed error curve map on AR database, K=5 in two figure.
Embodiment
This mixing Two-dimensional Probabilistic principal component analytical method, the Gaussian mixtures of matrix variables obeyed by sample, and the method comprises the following steps:
(1) the second order principal component model 2DPCA of probability is built according to formula (1)
X n=LB nR T+M+E n;(1)
Wherein L (p × r) and R (q × c) is the dimensionality reduction matrix in row and column direction respectively; B n(r × c) is sample X nhidden variable core, be called matrix of coefficients; R≤p, c≤q is the number of lines and columns after dimensionality reduction; M (p × q) is Mean Matrix, E nthe noise meeting matrix Gaussian distribution, its each component meets
(2) 2DPCA of mixing is built according to formula (2)
The average item M be wherein separated kthe average of kth class sample set, L kand R kthe kth class dimensionality reduction that sample set obtains, π kblending ratio, it is the variance of a kth Gaussian distribution;
(3) by the maximum likelihood function of formula (4), the parameter in formula (2) is estimated
Z nkvalue is 1 or 0, represents the n-th sample and whether belongs to a kth Gaussian distribution.
(4) when solution formula (4), variation greatest hope EM algorithm is utilized to carry out Optimal Parameters in the E step of EM algorithm, to hidden variable solve its Posterior distrbutionp: wherein average, be the covariance matrix in row and column direction respectively, calculate formula (9)-(11):
T n ( k ) = c&sigma; k 2 &lsqb; t r ( R k T R k S n ( k ) ) L k T L k + &sigma; k 2 t r ( S n ( k ) ) I r &rsqb; - 1 - - - ( 9 )
S n ( k ) = r&sigma; k 2 &lsqb; t r ( L k T L k T n ( k ) ) R k T R k + &sigma; k 2 t r ( T n ( k ) ) I c &rsqb; - 1 - - - ( 10 )
v e c ( Q n ( k ) ) = &lsqb; R k T R k &CircleTimes; L k T L k + &sigma; k I &CircleTimes; &sigma; k I &rsqb; - 1 y - - - ( 11 )
Wherein y = v e c ( L k T ( X n - M k ) R k ) ;
M walks the parameter in more new formula (4), obtains formula (12)-(14):
L k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) R k ( Q n ( k ) ) T &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k Q n ( k ) R k T R k ( Q n ( k ) ) T + t r ( R k T R k S n ( k ) ) T n ( k ) &rsqb; - 1 - - - ( 12 )
R k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) T L k Q n ( k ) &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k ( Q n ( k ) ) T L k T L k Q n ( k ) + t r ( L k T L k T n ( k ) ) S n ( k ) &rsqb; - 1 - - - ( 13 )
&sigma; k 2 = 1 pqN k { &Sigma; n = 1 N &gamma; n k t r ( X n - M k ) T ( X n - M k ) - 2 &Sigma; n = 1 N &gamma; n k t r ( R k < B n ( k ) > T L k T ( X n - M k ) ) + &Sigma; n = 1 N &gamma; n k t r ( < B n ( k ) T L k T L k B n ( k ) > R k T R k ) } - - - ( 14 )
Wherein γ n, krepresent that the n-th sample belongs to the posterior probability of a kth Gaussian distribution;
Iteration carries out E step and M step, makes likelihood function value increase and tend towards stability.
The present invention is based on mixed Gauss model and dimensionality reduction is carried out to 2-D data, by introducing hidden variable, use variation EM algorithm solving model parameter, and the matrix of coefficients after dimensionality reduction, reach the effect that 2-D data is compressed, use dimensionality reduction matrix and matrix of coefficients reconstructed image, compare with original image and obtain losing less image, matrix of coefficients after dimensionality reduction is regarded as the feature of sample, coefficient of performance matrix is classified to sample, so can carry out dimensionality reduction to 2-D data in row and column both direction, quality reconstruction is better.
Illustrate in greater detail the present invention now.
In order to solution formula (2), utilize the density function of the Posterior distrbutionp of all variablees in variation approximate data solving model.
The second order PCA (PSOPCA) of 1 probability
Make χ={ X 1, X 2..., X nn number of sample of one group of I.i.d. random variables, wherein each size is 2D principal component analysis (PCA) can be expressed as following form:
X n=LB nR T+M+E n;(1)
Wherein L (p × r) and R (q × c) is the dimensionality reduction matrix in row and column direction respectively, B n(r × c) is sample X nhidden variable core, i.e. matrix of coefficients.R≤p, c≤q is the number of lines and columns after dimensionality reduction.M (p × q) is Mean Matrix, E nthe noise meeting matrix Gaussian distribution, that is, namely its each component meets then model (1) is exactly the probability 2DPCA of standard.
2MixB2DPPCA model
For more complicated data set, only the original sample set of matching is difficult to a principal component model, because this principal component analysis (PCA) is a kind of overall dimensionality reduction model, for the sample of some Data distribution8 complexity, a principal direction is only found obviously to be irrational.Therefore, propose a kind of local 2DPCA model, utilize the mixed form of multiple 2DPCA to find one group of dimensionality reduction direction, thus better can represent the major component of raw data.
In this project, consider the 2DPCA of mixing, object is to the two dimensional sample data of complexity after obtaining classification, and solves the dimensionality reduction matrix of each class.Suppose sample X nobey the gauss hybrid models (GMM) be made up of K Gaussian distribution, namely
It should be noted that an average item M be separated kbe associated with K blending constituent, in fact it is the average of kth class sample set, L kand R kit is the kth class dimensionality reduction that sample set obtains.π kblending ratio, &pi; k > 0 , &Sigma; k = 1 K &pi; k = 1.
Introduce a K and tie up binary random variables z, in K element, only have an element z kequal 1, other elements are 0. namely z k∈ 0,1} and so p (z k=1)=π k, then the Posterior distrbutionp of z is defined as:
p ( z ) = &Pi; k = 1 K &pi; k z k
Therefore the conditional probability of a sample Xn obedience kth Gaussian distribution is
Then the conditional probability of sample Xn can be write as:
Suppose hidden variable matrix prior probability obey average be 0, row, column covariance is the Gaussian distribution of unit matrix:
The variation greatest hope of 3MixB2DPPCA solves
By maximum likelihood function, the parameter in formula (2) is estimated.By utilizing the thought of mixed Gauss model, formula (2) can reach classifies to the data in sample set, and finds different classes of dimensionality reduction matrix L kand R k(k=1 ..., K).The likelihood function of this model is:
When solving this model, variation greatest hope (EM) algorithm optimization Model Parameter can be utilized L k , R k , &pi; k , &sigma; k 2 , ( k = 1 , ... , K ) .
(1) variable γ is introduced n, k, it represents that namely the n-th sample belongs to the posterior probability of a kth Gaussian distribution,
&gamma; n k = &pi; k p ( X n | k ) p ( X n ) = &pi; k p ( v e c ( X n ) | k ) p ( v e c ( X n ) ) - - - ( 5 )
Wherein p (X n| k) be sample X nmarginal probability.Due in given hidden variable rear X nconditional probability be:
Therefore vec (X n) marginal distribution be: wherein C = &sigma; 2 I + W k W k T , W k = R k &CircleTimes; L k
Thus, the weight coefficient of each Gaussian distribution can be obtained:
&pi; k = 1 N &Sigma; n = 1 N &gamma; n k - - - ( 7 )
And average:
M k = &Sigma; n = 1 N &gamma; n k X n &Sigma; n = 1 N &gamma; n k - - - ( 8 )
(2) parameter of first initialization mixture model (1) in the E step of EM algorithm, to hidden variable solve its Posterior distrbutionp: wherein the covariance matrix in row and column direction respectively.Can draw as calculated:
T n ( k ) = c&sigma; k 2 &lsqb; t r ( R k T R k S n ( k ) ) L k T L k + &sigma; k 2 t r ( S n ( k ) ) I r &rsqb; - 1 - - - ( 9 )
S n ( k ) = r&sigma; k 2 &lsqb; t r ( L k T L k T n ( k ) ) R k T R k + &sigma; k 2 t r ( T n ( k ) ) I c &rsqb; - 1 - - - ( 10 )
v e c ( Q n ( k ) ) = &lsqb; R k T R k &CircleTimes; L k T L k + &sigma; k I &CircleTimes; &sigma; k I &rsqb; - 1 y - - - ( 11 )
Wherein y = v e c ( L k T ( X n - M k ) R k ) .
(3), after E step obtains the Posterior distrbutionp of hidden variable, M walks the parameter in Renewal model, i.e. dimensionality reduction matrix.By maximization likelihood function, can obtain:
L k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) R k ( Q n ( k ) ) T &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k Q n ( k ) R k T R k ( Q n ( k ) ) T + t r ( R k T R k S n ( k ) ) T n ( k ) &rsqb; - 1 - - - ( 12 )
R k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) T L k Q n ( k ) &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k ( Q n ( k ) ) T L k T L k Q n ( k ) + t r ( L k T L k T n ( k ) ) S n ( k ) &rsqb; - 1 - - - ( 13 )
Iteration carries out E step and M step, makes likelihood function value increase and tend towards stability.
The present invention has done corresponding experiment on four databases, handwritten form database, Yale database, AR database and FERET database.These experiments are to prove that method in this paper can be lost smaller to sample dimensionality reduction, and matrix of coefficients after dimensionality reduction is as the feature of original image, can effectively to Images Classification.The algorithm related to has: GLRAM (GeneralizedLowRankApproximationsofMatrices), mixPPCA (MixtureProbabilisticPCA), mixB2DPPCA (MixtureofBilateral-ProjectionTwo-dimensionalProbabilisti cPCA).
The preparation of 1 data and experiment parameter setting
Following four databases are applied in experiment:
◆ a subset of MINIST database ( http:// yann.lecun.com/exdb/mnist)
◆ Yale database
( http://vision.ucsd.edu/content/yale-face-database)
◆ AR database
( http://rvl1.ecn.purdue.edu/~aleix/aleix_face_DB.html)
◆ FERET database
( http://www.itl.nist.gov/iad/humanid/feret/feret_master.html)
From MNIST database, select 1000 pictures, namely each digital random selects 100.All images are gray level image and size is 28 × 28.
Yale database comprises 15 people, everyone 11 width images.These images obtain in different illumination conditions and expression.Training stage we select everyone 6 width images (90) as training, remaining conduct test, all image sizes are 64 × 64.
AR database comprises the 4000 width images of 126 people, everyone 26 width face direct pictures, and this 26 width image comprises expression shape change, illumination variation and blocks.Everyone 26 width images are divided into two parts (shooting time is separated by two weeks), and two parts are 13.In this experiment, select the image of 30 people to test (male 15 people, female 15 people), and only select everyone unscreened 14 width images.Wherein everyone front 7 width are used for training, and rear 7 width are tested, and every pictures down-sampling is 64 × 64.
FERET database comprises the 1400 width images of 200 people, and everyone has 7 width images, and comprise angle, expression, light and shade change in this 7 width image, the size of every pictures is 80 × 80.The image selecting 50 people is at random tested, and everyone random choose 5 is trained, and tests in addition 2.
π in this experiment kinitial value is 1/K, T kand S kinitial value is unit matrix, σ kinitial value is 1, L kand R kinitial value is given at random.
2 reconstructed errors
The mainly reconstructed error of more different dimension reduction method on MNIST and Yale database.The method related to has GLRAM, mixB2DPPCA.GLRAM and mixB2DPPCA that mixPPCA and the present invention propose is the feature extracting method of two dimension, at row, column both direction dimensionality reduction; MixPPCA and mixB2DPPCA utilizes gauss hybrid models to Data Dimensionality Reduction.
Fig. 1 is the reconstructed error of algorithms of different on MNIST database.As can be seen from the figure, when K is identical, the reconstructed error of method in this paper is used to be better than mixPPCA; When K changes, the reconstructed error of GLRAM is constant, this is because GLRAM method uses non-mixed model, has nothing to do with the size of K; When K increases, the reconstructed error of mixPPCA and mixB2DPPCA all decreases, and mixB2DPPCA is always better than mixPPCA and GLRAM.
Fig. 2 is the reconstructed error curve map on Yale database and AR database, wherein K=5.As can be seen from the figure the algorithm using the present invention to propose can obtain the less reconstructed image of reconstructed error, i.e. the loss reduction when dimensionality reduction.
3 discriminations
3rd the main robustness verifying mixB2DPPCA algorithm at Yale, AR and FERET database in discrimination of experiment.Utilize arest neighbors (1-NN) algorithm as sorter in experiment.Table 1 is the discriminations of GLRAM, mixPPCA and mixB2DPCA tri-kinds of methods on Yale storehouse.Wherein the dimension (r, c) of dimensionality reduction is respectively (2,2), (4,4), (6,6) and (8,8).The K of gauss hybrid models respectively value is 4,6,8.
Table 1
Table 2 is the discriminations on AR database.Dimensionality reduction dimension (r, c) is respectively (4,4), (6,6) and (8,8).The K of gauss hybrid models respectively value is 6,8,10.
Table 2
Table 3 is the discriminations on FERET database.Dimensionality reduction dimension (r, c) is respectively (4,4), (6,6), (8,8) and (10,10).The K of gauss hybrid models respectively value is 6,8,10.
Table 3
As can be seen from the table above, the feature using mixB2DPPCA of the present invention to extract has clear superiority in identification.
The above; it is only preferred embodiment of the present invention; not any pro forma restriction is done to the present invention, every above embodiment is done according to technical spirit of the present invention any simple modification, equivalent variations and modification, all still belong to the protection domain of technical solution of the present invention.

Claims (1)

1. mix a Two-dimensional Probabilistic principal component analytical method, it is characterized in that, the Gaussian mixtures of matrix variables obeyed by sample, and the method comprises the following steps:
(1) the second order principal component model 2DPCA of probability is built according to formula (1)
X n=LB nR T+M+E n;(1)
Wherein L (p × r) and R (q × c) is the dimensionality reduction matrix in row and column direction respectively; B n(r × c) is sample X nhidden variable core, be called matrix of coefficients; R≤p, c≤q is the number of lines and columns after dimensionality reduction; M (p × q) is Mean Matrix, E nthe noise meeting matrix Gaussian distribution, its each component meets
(2) 2DPCA of mixing is built according to formula (2)
The average item M be wherein separated kthe average of kth class sample set, L kand R kthe kth class dimensionality reduction that sample set obtains, π kblending ratio, π k> 0, it is the variance of a kth Gaussian distribution;
(3) by the maximum likelihood function of formula (4), the parameter in formula (2) is estimated
Z nkvalue is 1 or 0, represents the n-th sample and whether belongs to a kth Gaussian distribution.
(4) when solution formula (4), variation greatest hope EM algorithm is utilized to carry out Optimal Parameters in the E step of EM algorithm, to hidden variable solve its Posterior distrbutionp: wherein average, be the covariance matrix in row and column direction respectively, calculate formula (9)-(11):
T n ( k ) = c&sigma; k 2 &lsqb; t r ( R k T R k S n ( k ) ) L k T L k + &sigma; k 2 t r ( S n ( k ) ) I r &rsqb; - 1 - - - ( 9 )
S n ( k ) = r&sigma; k 2 &lsqb; t r ( L k T L k T n ( k ) ) R k T R k + &sigma; k 2 t r ( T n ( k ) ) I c &rsqb; - 1 - - - ( 10 )
v e c ( Q n ( k ) ) = &lsqb; R k T R k &CircleTimes; L k T L k + &sigma; k I &CircleTimes; &sigma; k I &rsqb; - 1 y - - - ( 11 )
Wherein y = v e c ( L k T ( X n - M k ) R k ) ;
M walks the parameter in more new formula (4), obtains formula (12)-(14):
L k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) R k ( Q n ( k ) ) T &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k Q n ( k ) R k T R k ( Q n ( k ) ) T + t r ( R k T R k S n ( k ) ) T n ( k ) &rsqb; - 1 - - - ( 12 )
R k = &lsqb; &Sigma; n = 1 N &gamma; n k ( X n - M k ) T L k Q n ( k ) &rsqb; &lsqb; &Sigma; n = 1 N &gamma; n k ( Q n ( k ) ) T L k T L k Q n ( k ) + t r ( L k T L k T n ( k ) ) S n ( k ) &rsqb; - 1 - - - ( 13 )
&sigma; k 2 = 1 pqN k { &Sigma; n = 1 N &gamma; n k t r ( X n - M k ) T ( X n - M k ) - 2 &Sigma; n = 1 N &gamma; n k t r ( R k < B n ( k ) > T L k T ( X n - M k ) ) + &Sigma; n = 1 N &gamma; n k t r ( < B n ( k ) T L k T L k B n ( k ) > R k T R k ) } - - - ( 14 )
Wherein γ n, krepresent that the n-th sample belongs to the posterior probability of a kth Gaussian distribution;
Iteration carries out E step and M step, makes likelihood function value increase and tend towards stability.
CN201511022718.1A 2015-12-31 2015-12-31 Mixed two-dimensional probabilistic principal component analysis method Pending CN105469101A (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201511022718.1A CN105469101A (en) 2015-12-31 2015-12-31 Mixed two-dimensional probabilistic principal component analysis method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201511022718.1A CN105469101A (en) 2015-12-31 2015-12-31 Mixed two-dimensional probabilistic principal component analysis method

Publications (1)

Publication Number Publication Date
CN105469101A true CN105469101A (en) 2016-04-06

Family

ID=55606772

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201511022718.1A Pending CN105469101A (en) 2015-12-31 2015-12-31 Mixed two-dimensional probabilistic principal component analysis method

Country Status (1)

Country Link
CN (1) CN105469101A (en)

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107038456A (en) * 2017-03-23 2017-08-11 北京工业大学 A kind of image classification method of the probability linear discriminant analysis based on L1 norms
CN108549789A (en) * 2018-04-19 2018-09-18 中南大学 A kind of integrated modelling approach of tertiary cathode material feed proportioning system
CN109917777A (en) * 2019-04-16 2019-06-21 浙江科技学院 Fault detection method based on mixing multi-sampling rate Probabilistic Principal Component Analysis model

Cited By (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107038456A (en) * 2017-03-23 2017-08-11 北京工业大学 A kind of image classification method of the probability linear discriminant analysis based on L1 norms
CN108549789A (en) * 2018-04-19 2018-09-18 中南大学 A kind of integrated modelling approach of tertiary cathode material feed proportioning system
CN108549789B (en) * 2018-04-19 2021-08-03 中南大学 Integrated modeling method for ternary cathode material batching system
CN109917777A (en) * 2019-04-16 2019-06-21 浙江科技学院 Fault detection method based on mixing multi-sampling rate Probabilistic Principal Component Analysis model
CN109917777B (en) * 2019-04-16 2020-08-25 浙江科技学院 Fault detection method based on mixed multi-sampling rate probability principal component analysis model

Similar Documents

Publication Publication Date Title
Cherian et al. Riemannian dictionary learning and sparse coding for positive definite matrices
Zhang et al. Jointly learning structured analysis discriminative dictionary and analysis multiclass classifier
Gosselin et al. Revisiting the fisher vector for fine-grained classification
Yang et al. Robust kernel representation with statistical local features for face recognition
CN104268593B (en) The face identification method of many rarefaction representations under a kind of Small Sample Size
Shen et al. Face image classification by pooling raw features
Xie et al. Orientational pyramid matching for recognizing indoor scenes
Bongsoo Choy et al. Enriching object detection with 2d-3d registration and continuous viewpoint estimation
Huang et al. Globality-locality preserving projections for biometric data dimensionality reduction
CN103295032B (en) Based on the image classification method of spatial Fisher vector
Zhu et al. Deep learning multi-view representation for face recognition
CN103514443A (en) Single-sample face recognition transfer learning method based on LPP (Low Power Point) feature extraction
CN106529586A (en) Image classification method based on supplemented text characteristic
CN105844261A (en) 3D palmprint sparse representation recognition method based on optimization feature projection matrix
Liu et al. Texture analysis with shape co-occurrence patterns
CN107358172A (en) A kind of human face characteristic point initial method based on facial orientation classification
Xie et al. Incorporating visual adjectives for image classification
CN105469101A (en) Mixed two-dimensional probabilistic principal component analysis method
Tripathi et al. Real time object detection using CNN
CN104376312B (en) Face identification method based on bag of words compressed sensing feature extraction
CN106803105B (en) Image classification method based on sparse representation dictionary learning
Wang et al. Differential geometry boosts convolutional neural networks for object detection
Zheng et al. Extracting non-negative basis images using pixel dispersion penalty
Zeng et al. Learning-based multiple pooling fusion in multi-view convolutional neural network for 3D model classification and retrieval
CN105046286A (en) Supervision multi-view feature selection method based on automatic generation of view and unit with l1 and l2 norm minimization

Legal Events

Date Code Title Description
C06 Publication
PB01 Publication
C10 Entry into substantive examination
SE01 Entry into force of request for substantive examination
RJ01 Rejection of invention patent application after publication

Application publication date: 20160406

RJ01 Rejection of invention patent application after publication